import os from typing import Dict bib = """ @inproceedings{dimosthenis-etal-2022-twitter, title = "{T}witter {T}opic {C}lassification", author = "Antypas, Dimosthenis and Ushio, Asahi and Camacho-Collados, Jose and Neves, Leonardo and Silva, Vitor and Barbieri, Francesco", booktitle = "Proceedings of the 29th International Conference on Computational Linguistics", month = oct, year = "2022", address = "Gyeongju, Republic of Korea", publisher = "International Committee on Computational Linguistics" } """ def get_readme(model_name: str, metric: Dict, language_model, extra_desc: str = ''): return f"""--- datasets: - cardiffnlp/tweet_topic_multi metrics: - f1 - accuracy model-index: - name: {model_name} results: - task: type: text-classification name: Text Classification dataset: name: cardiffnlp/tweet_topic_multi type: cardiffnlp/tweet_topic_multi args: cardiffnlp/tweet_topic_multi split: test_2021 metrics: - name: F1 type: f1 value: {metric['test/eval_f1']} - name: F1 (macro) type: f1_macro value: {metric['test/eval_f1_macro']} - name: Accuracy type: accuracy value: {metric['test/eval_accuracy']} pipeline_tag: text-classification widget: - text: "I'm sure the {"{@Tampa Bay Lightning@}"} would’ve rather faced the Flyers but man does their experience versus the Blue Jackets this year and last help them a lot versus this Islanders team. Another meat grinder upcoming for the good guys" example_title: "Example 1" - text: "Love to take night time bike rides at the jersey shore. Seaside Heights boardwalk. Beautiful weather. Wishing everyone a safe Labor Day weekend in the US." example_title: "Example 2" --- # {model_name} This model is a fine-tuned version of [{language_model}](https://huggingface.co/{language_model}) on the [tweet_topic_multi](https://huggingface.co/datasets/cardiffnlp/tweet_topic_multi). {extra_desc} Fine-tuning script can be found [here](https://huggingface.co/datasets/cardiffnlp/tweet_topic_multi/blob/main/lm_finetuning.py). It achieves the following results on the test_2021 set: - F1 (micro): {metric['test/eval_f1']} - F1 (macro): {metric['test/eval_f1_macro']} - Accuracy: {metric['test/eval_accuracy']} ### Usage ```python pipe = pipeline("text-classification", "cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-multi-2020", problem_type="multi_label_classification") ``` ### Reference If you use any resource from T-NER, please consider to cite our [paper](https://aclanthology.org/2021.eacl-demos.7/). ``` {bib} ``` """