readme: update base model and backbone model (#3)
Browse files- readme: update base model and backbone model (b19a21e8d863072101f5f4c07c56062f5ff72fde)
README.md
CHANGED
@@ -5,7 +5,7 @@ tags:
|
|
5 |
- flair
|
6 |
- token-classification
|
7 |
- sequence-tagger-model
|
8 |
-
base_model:
|
9 |
inference: false
|
10 |
widget:
|
11 |
- text: — 469 . Πεδία . Les tribraques formés par un seul mot sont rares chez les
|
@@ -17,7 +17,7 @@ widget:
|
|
17 |
|
18 |
This Flair model was fine-tuned on the
|
19 |
[AjMC French](https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-ajmc.md)
|
20 |
-
NER Dataset using
|
21 |
|
22 |
The AjMC dataset consists of NE-annotated historical commentaries in the field of Classics,
|
23 |
and was created in the context of the [Ajax MultiCommentary](https://mromanello.github.io/ajax-multi-commentary/)
|
@@ -36,7 +36,7 @@ Thus, the inference widget is not working with hmByT5 at the moment on the Model
|
|
36 |
This should be fixed in future, when ByT5 fine-tuning is supported in Flair directly.
|
37 |
|
38 |
[1]: https://github.com/stefan-it/hmBench/blob/main/byt5_embeddings.py
|
39 |
-
|
40 |
# Results
|
41 |
|
42 |
We performed a hyper-parameter search over the following parameters with 5 different seeds per configuration:
|
|
|
5 |
- flair
|
6 |
- token-classification
|
7 |
- sequence-tagger-model
|
8 |
+
base_model: hmbyt5-preliminary/byt5-small-historic-multilingual-span20-flax
|
9 |
inference: false
|
10 |
widget:
|
11 |
- text: — 469 . Πεδία . Les tribraques formés par un seul mot sont rares chez les
|
|
|
17 |
|
18 |
This Flair model was fine-tuned on the
|
19 |
[AjMC French](https://github.com/hipe-eval/HIPE-2022-data/blob/main/documentation/README-ajmc.md)
|
20 |
+
NER Dataset using hmByT5 as backbone LM.
|
21 |
|
22 |
The AjMC dataset consists of NE-annotated historical commentaries in the field of Classics,
|
23 |
and was created in the context of the [Ajax MultiCommentary](https://mromanello.github.io/ajax-multi-commentary/)
|
|
|
36 |
This should be fixed in future, when ByT5 fine-tuning is supported in Flair directly.
|
37 |
|
38 |
[1]: https://github.com/stefan-it/hmBench/blob/main/byt5_embeddings.py
|
39 |
+
|
40 |
# Results
|
41 |
|
42 |
We performed a hyper-parameter search over the following parameters with 5 different seeds per configuration:
|