go-inoue commited on
Commit
d383f6d
1 Parent(s): f1f2f2e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -10,6 +10,7 @@ widget:
10
  **CAMeLBERT MSA SA Model** is a Sentiment Analysis (SA) model that was built by fine-tuning the [CAMeLBERT Modern Standard Arabic (MSA)](https://huggingface.co/CAMeL-Lab/bert-base-arabic-camelbert-msa/) model.
11
  For the fine-tuning, we used the [ASTD](https://aclanthology.org/D15-1299.pdf), [ArSAS](http://lrec-conf.org/workshops/lrec2018/W30/pdf/22_W30.pdf), and [SemEval](https://aclanthology.org/S17-2088.pdf) datasets.
12
  Our fine-tuning procedure and the hyperparameters we used can be found in our paper *"[The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models](https://arxiv.org/abs/2103.06678)."* Our fine-tuning code can be found [here](https://github.com/CAMeL-Lab/CAMeLBERT).
 
13
  ## Intended uses
14
  You can use the CAMeLBERT MSA SA model directly as part of our [CAMeL Tools](https://github.com/CAMeL-Lab/camel_tools) SA component (*recommended*) or as part of the transformers pipeline.
15
  #### How to use
@@ -30,7 +31,9 @@ You can also use the SA model directly with a transformers pipeline:
30
  [{'label': 'positive', 'score': 0.9616648554801941},
31
  {'label': 'negative', 'score': 0.9779177904129028}]
32
  ```
33
- *Note*: to download our models, you would need `transformers>=3.5.0`. Otherwise, you could download the models
 
 
34
  ## Citation
35
  ```bibtex
36
  @inproceedings{inoue-etal-2021-interplay,
10
  **CAMeLBERT MSA SA Model** is a Sentiment Analysis (SA) model that was built by fine-tuning the [CAMeLBERT Modern Standard Arabic (MSA)](https://huggingface.co/CAMeL-Lab/bert-base-arabic-camelbert-msa/) model.
11
  For the fine-tuning, we used the [ASTD](https://aclanthology.org/D15-1299.pdf), [ArSAS](http://lrec-conf.org/workshops/lrec2018/W30/pdf/22_W30.pdf), and [SemEval](https://aclanthology.org/S17-2088.pdf) datasets.
12
  Our fine-tuning procedure and the hyperparameters we used can be found in our paper *"[The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models](https://arxiv.org/abs/2103.06678)."* Our fine-tuning code can be found [here](https://github.com/CAMeL-Lab/CAMeLBERT).
13
+
14
  ## Intended uses
15
  You can use the CAMeLBERT MSA SA model directly as part of our [CAMeL Tools](https://github.com/CAMeL-Lab/camel_tools) SA component (*recommended*) or as part of the transformers pipeline.
16
  #### How to use
31
  [{'label': 'positive', 'score': 0.9616648554801941},
32
  {'label': 'negative', 'score': 0.9779177904129028}]
33
  ```
34
+ *Note*: to download our models, you would need `transformers>=3.5.0`.
35
+ Otherwise, you could download the models manually.
36
+
37
  ## Citation
38
  ```bibtex
39
  @inproceedings{inoue-etal-2021-interplay,