DCU-NLP commited on
Commit
3b2eb69
1 Parent(s): 63f70ee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -1
README.md CHANGED
@@ -11,7 +11,7 @@ widget:
11
 
12
  # bert-base-irish-cased-v1
13
 
14
- [gaBERT](https://arxiv.org/abs/2107.12930) is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.
15
 
16
  ## Model description
17
 
@@ -36,3 +36,27 @@ The following hyperparameters were used during training:
36
  - TensorFlow 2.9.1
37
  - Datasets 2.3.2
38
  - Tokenizers 0.12.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  # bert-base-irish-cased-v1
13
 
14
+ [gaBERT](https://aclanthology.org/2022.lrec-1.511/) is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.
15
 
16
  ## Model description
17
 
 
36
  - TensorFlow 2.9.1
37
  - Datasets 2.3.2
38
  - Tokenizers 0.12.1
39
+
40
+ ### Citation
41
+ If you use this model in your research, please consider citing the following paper:
42
+ ```
43
+ @inproceedings{barry-etal-2022-gabert,
44
+ title = "ga{BERT} {---} an {I}rish Language Model",
45
+ author = "Barry, James and
46
+ Wagner, Joachim and
47
+ Cassidy, Lauren and
48
+ Cowap, Alan and
49
+ Lynn, Teresa and
50
+ Walsh, Abigail and
51
+ {\'O} Meachair, M{\'\i}che{\'a}l J. and
52
+ Foster, Jennifer",
53
+ booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
54
+ month = jun,
55
+ year = "2022",
56
+ address = "Marseille, France",
57
+ publisher = "European Language Resources Association",
58
+ url = "https://aclanthology.org/2022.lrec-1.511",
59
+ pages = "4774--4788",
60
+ abstract = "The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many NLP tasks. We introduce gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and the monolingual Irish WikiBERT, and we show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We compare the results of fine-tuning a gaBERT model with an mBERT model for the task of identifying verbal multiword expressions, and show that the fine-tuned gaBERT model also performs better at this task. We release gaBERT and related code to the community.",
61
+ }
62
+ ```