Edit model card

Introduction

This is a fine-tuned LM in our papers below and the related GitHub repo is here.

Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model (Cao et al., WWW 2024 Workshop - SocialNLP)

Pay Attention to the Robustness of Chinese Minority Language Models! Syllable-level Textual Adversarial Attack on Tibetan Script (Cao et al., ACL 2023 Workshop - TrustNLP)

Citation

If you think our work useful, please kindly cite our paper.

@inproceedings{10.1145/3589335.3652503,
    author = {Cao, Xi and Qun, Nuo and Gesang, Quzong and Zhu, Yulei and Nyima, Trashi},
    title = {Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model},
    year = {2024},
    isbn = {9798400701726},
    publisher = {Association for Computing Machinery},
    address = {New York, NY, USA},
    url = {https://doi.org/10.1145/3589335.3652503},
    doi = {10.1145/3589335.3652503},
    booktitle = {Companion Proceedings of the ACM on Web Conference 2024},
    pages = {1672–1680},
    numpages = {9},
    keywords = {language model, robustness, textual adversarial attack, tibetan},
    location = {Singapore, Singapore},
    series = {WWW '24}
}
@inproceedings{cao-etal-2023-pay-attention,
    title = "Pay Attention to the Robustness of {C}hinese Minority Language Models! Syllable-level Textual Adversarial Attack on {T}ibetan Script",
    author = "Cao, Xi  and
      Dawa, Dolma  and
      Qun, Nuo  and
      Nyima, Trashi",
    booktitle = "Proceedings of the 3rd Workshop on Trustworthy Natural Language Processing (TrustNLP 2023)",
    month = jul,
    year = "2023",
    address = "Toronto, Canada",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.trustnlp-1.4",
    pages = "35--46"
}
Downloads last month
11
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Collection including UTibetNLP/cino-large-v2_TUSA