pipeline_tag
stringclasses 48
values | library_name
stringclasses 198
values | text
stringlengths 1
900k
| metadata
stringlengths 2
438k
| id
stringlengths 5
122
| last_modified
null | tags
sequencelengths 1
1.84k
| sha
null | created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
1
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2175
- Accuracy: 0.9215
- F1: 0.9216
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.7814 | 1.0 | 250 | 0.3105 | 0.907 | 0.9046 |
| 0.2401 | 2.0 | 500 | 0.2175 | 0.9215 | 0.9216 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.9215, "name": "Accuracy"}, {"type": "f1", "value": 0.9215538311282218, "name": "F1"}]}]}]} | Crives/distilbert-base-uncased-finetuned-emotion | null | [
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-emotion
=========================================
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2175
* Accuracy: 0.9215
* F1: 0.9216
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.10.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
56,
101,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
text-generation | transformers | #rick DialoGPT Model | {"tags": ["conversational"]} | Cryptikdw/DialoGPT-small-rick | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| #rick DialoGPT Model | [] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
39
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
text-generation | transformers |
# Paladin Danse DialoGPT Model | {"tags": ["conversational"]} | Cthyllax/DialoGPT-medium-PaladinDanse | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Paladin Danse DialoGPT Model | [
"# Paladin Danse DialoGPT Model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Paladin Danse DialoGPT Model"
] | [
39,
10
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Paladin Danse DialoGPT Model"
] |
token-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# IceBERT-finetuned-ner
This model is a fine-tuned version of [vesteinn/IceBERT](https://huggingface.co/vesteinn/IceBERT) on the mim_gold_ner dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0807
- Precision: 0.8927
- Recall: 0.8632
- F1: 0.8777
- Accuracy: 0.9850
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.0544 | 1.0 | 2904 | 0.0774 | 0.8859 | 0.8490 | 0.8670 | 0.9833 |
| 0.0284 | 2.0 | 5808 | 0.0781 | 0.8709 | 0.8590 | 0.8649 | 0.9840 |
| 0.0166 | 3.0 | 8712 | 0.0807 | 0.8927 | 0.8632 | 0.8777 | 0.9850 |
### Framework versions
- Transformers 4.11.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
| {"license": "gpl-3.0", "tags": ["generated_from_trainer"], "datasets": ["mim_gold_ner"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "IceBERT-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "mim_gold_ner", "type": "mim_gold_ner", "args": "mim-gold-ner"}, "metrics": [{"type": "precision", "value": 0.8927335640138409, "name": "Precision"}, {"type": "recall", "value": 0.8631855657784682, "name": "Recall"}, {"type": "f1", "value": 0.8777109531620194, "name": "F1"}, {"type": "accuracy", "value": 0.9849836396073506, "name": "Accuracy"}]}]}]} | Culmenus/IceBERT-finetuned-ner | null | [
"transformers",
"pytorch",
"tensorboard",
"roberta",
"token-classification",
"generated_from_trainer",
"dataset:mim_gold_ner",
"license:gpl-3.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| IceBERT-finetuned-ner
=====================
This model is a fine-tuned version of vesteinn/IceBERT on the mim\_gold\_ner dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0807
* Precision: 0.8927
* Recall: 0.8632
* F1: 0.8777
* Accuracy: 0.9850
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.11.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] | [
61,
101,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
token-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# XLMR-ENIS-finetuned-ner
This model is a fine-tuned version of [vesteinn/XLMR-ENIS](https://huggingface.co/vesteinn/XLMR-ENIS) on the mim_gold_ner dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0891
- Precision: 0.8804
- Recall: 0.8517
- F1: 0.8658
- Accuracy: 0.9837
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.0573 | 1.0 | 2904 | 0.1024 | 0.8608 | 0.8003 | 0.8295 | 0.9799 |
| 0.0307 | 2.0 | 5808 | 0.0899 | 0.8707 | 0.8380 | 0.8540 | 0.9825 |
| 0.0198 | 3.0 | 8712 | 0.0891 | 0.8804 | 0.8517 | 0.8658 | 0.9837 |
### Framework versions
- Transformers 4.11.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
| {"license": "agpl-3.0", "tags": ["generated_from_trainer"], "datasets": ["mim_gold_ner"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "XLMR-ENIS-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "mim_gold_ner", "type": "mim_gold_ner", "args": "mim-gold-ner"}, "metrics": [{"type": "precision", "value": 0.8803619696791632, "name": "Precision"}, {"type": "recall", "value": 0.8517339397384878, "name": "Recall"}, {"type": "f1", "value": 0.8658113730929264, "name": "F1"}, {"type": "accuracy", "value": 0.9837103244207861, "name": "Accuracy"}]}]}]} | Culmenus/XLMR-ENIS-finetuned-ner | null | [
"transformers",
"pytorch",
"tensorboard",
"xlm-roberta",
"token-classification",
"generated_from_trainer",
"dataset:mim_gold_ner",
"license:agpl-3.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #xlm-roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-agpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| XLMR-ENIS-finetuned-ner
=======================
This model is a fine-tuned version of vesteinn/XLMR-ENIS on the mim\_gold\_ner dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0891
* Precision: 0.8804
* Recall: 0.8517
* F1: 0.8658
* Accuracy: 0.9837
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.11.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-agpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] | [
64,
101,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #token-classification #generated_from_trainer #dataset-mim_gold_ner #license-agpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
automatic-speech-recognition | transformers |
# Wav2Vec2-Large-XLSR-53-Vietnamese
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Vietnamese using the [Common Voice](https://huggingface.co/datasets/common_voice), [Infore_25h dataset](https://files.huylenguyen.com/25hours.zip) (Password: BroughtToYouByInfoRe)
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
```python
import torch
import torchaudio
from datasets import load_dataset
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
test_dataset = load_dataset("common_voice", "vi", split="test[:2%]")
processor = Wav2Vec2Processor.from_pretrained("CuongLD/wav2vec2-large-xlsr-vietnamese")
model = Wav2Vec2ForCTC.from_pretrained("CuongLD/wav2vec2-large-xlsr-vietnamese")
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits
predicted_ids = torch.argmax(logits, dim=-1)
print("Prediction:", processor.batch_decode(predicted_ids))
print("Reference:", test_dataset["sentence"][:2])
```
## Evaluation
The model can be evaluated as follows on the Vietnamese test data of Common Voice.
```python
import torch
import torchaudio
from datasets import load_dataset, load_metric
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
import re
test_dataset = load_dataset("common_voice", "vi", split="test")
wer = load_metric("wer")
processor = Wav2Vec2Processor.from_pretrained("CuongLD/wav2vec2-large-xlsr-vietnamese")
model = Wav2Vec2ForCTC.from_pretrained("CuongLD/wav2vec2-large-xlsr-vietnamese")
model.to("cuda")
chars_to_ignore_regex = '[\,\?\.\!\-\;\:\"\“]'
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower()
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def evaluate(batch):
inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits
pred_ids = torch.argmax(logits, dim=-1)
batch["pred_strings"] = processor.batch_decode(pred_ids)
return batch
result = test_dataset.map(evaluate, batched=True, batch_size=8)
print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"])))
```
**Test Result**: 58.63 %
## Training
The Common Voice `train`, `validation`, and `Infore_25h` datasets were used for training
The script used for training can be found [here](https://drive.google.com/file/d/1AW9R8IlsapiSGh9n3aECf23t-zhk3wUh/view?usp=sharing)
=======================To here===============================>
Your model in then available under *huggingface.co/CuongLD/wav2vec2-large-xlsr-vietnamese* for everybody to use 🎉.
## How to evaluate my trained checkpoint
Having uploaded your model, you should now evaluate your model in a final step. This should be as simple as
copying the evaluation code of your model card into a python script and running it. Make sure to note
the final result on the model card **both** under the YAML tags at the very top **and** below your evaluation code under "Test Results".
## Rules of training and evaluation
In this section, we will quickly go over what data is allowed to be used as training
data, what kind of data preprocessing is allowed be used, and how the model should be evaluated.
To make it very simple regarding the first point: **All data except the official common voice `test` data set can be used as training data**. For models trained in a language that is not included in Common Voice, the author of the model is responsible to
leave a reasonable amount of data for evaluation.
Second, the rules regarding the preprocessing are not that as straight-forward. It is allowed (and recommended) to
normalize the data to only have lower-case characters. It is also allowed (and recommended) to remove typographical
symbols and punctuation marks. A list of such symbols can *e.g.* be fonud [here](https://en.wikipedia.org/wiki/List_of_typographical_symbols_and_punctuation_marks) - however here we already must be careful. We should **not** remove a symbol that
would change the meaning of the words, *e.g.* in English, we should not remove the single quotation mark `'` since it
would change the meaning of the word `"it's"` to `"its"` which would then be incorrect. So the golden rule here is to
not remove any characters that could change the meaning of a word into another word. This is not always obvious and should
be given some consideration. As another example, it is fine to remove the "Hypen-minus" sign "`-`" since it doesn't change the
meaninng of a word to another one. *E.g.* "`fine-tuning`" would be changed to "`finetuning`" which has still the same meaning.
Since those choices are not always obvious when in doubt feel free to ask on Slack or even better post on the forum, as was
done, *e.g.* [here](https://discuss.huggingface.co/t/spanish-asr-fine-tuning-wav2vec2/4586).
## Tips and tricks
This section summarizes a couple of tips and tricks across various topics. It will continously be updated during the week.
### How to combine multiple datasets into one
Check out [this](https://discuss.huggingface.co/t/how-to-combine-local-data-files-with-an-official-dataset/4685) post.
### How to effectively preprocess the data
### How to do efficiently load datasets with limited ram and hard drive space
Check out [this](https://discuss.huggingface.co/t/german-asr-fine-tuning-wav2vec2/4558/8?u=patrickvonplaten) post.
### How to do hyperparameter tuning
### How to preprocess and evaluate character based languages
## Further reading material
It is recommended that take some time to read up on how Wav2vec2 works in theory.
Getting a better understanding of the theory and the inner mechanisms of the model often helps when fine-tuning the model.
**However**, if you don't like reading blog posts/papers, don't worry - it is by no means necessary to go through the theory to fine-tune Wav2Vec2 on your language of choice.
If you are interested in learning more about the model though, here are a couple of resources that are important to better understand Wav2Vec2:
- [Facebook's Wav2Vec2 blog post](https://ai.facebook.com/blog/wav2vec-state-of-the-art-speech-recognition-through-self-supervision/)
- [Official Wav2Vec2 paper](https://arxiv.org/abs/2006.11477)
- [Official XLSR Wav2vec2 paper](https://arxiv.org/pdf/2006.13979.pdf)
- [Hugging Face Blog](https://huggingface.co/blog/fine-tune-xlsr-wav2vec2)
- [How does CTC (Connectionist Temporal Classification) work](https://distill.pub/2017/ctc/)
It helps to have a good understanding of the following points:
- How was XLSR-Wav2Vec2 pretrained? -> Feature vectors were masked and had to be predicted by the model; very similar in spirit to masked language model of BERT.
- What parts of XLSR-Wav2Vec2 are responsible for what? What is the feature extractor part used for? -> extract feature vectors from the 1D raw audio waveform; What is the transformer part doing? -> mapping feature vectors to contextualized feature vectors; ...
- What part of the model needs to be fine-tuned? -> The pretrained model **does not** include a language head to classify the contextualized features to letters. This is randomly initialized when loading the pretrained checkpoint and has to be fine-tuned. Also, note that the authors recommend to **not** further fine-tune the feature extractor.
- What data was used to XLSR-Wav2Vec2? The checkpoint we will use for further fine-tuning was pretrained on **53** languages.
- What languages are considered to be similar by XLSR-Wav2Vec2? In the official [XLSR Wav2Vec2 paper](https://arxiv.org/pdf/2006.13979.pdf), the authors show nicely which languages share a common contextualized latent space. It might be useful for you to extend your training data with data of other languages that are considered to be very similar by the model (or you).
## FAQ
- Can a participant fine-tune models for more than one language?
Yes! A participant can fine-tune models in as many languages she/he likes
- Can a participant use extra data (apart from the common voice data)?
Yes! All data except the official common voice `test data` can be used for training.
If a participant wants to train a model on a language that is not part of Common Voice (which
is very much encouraged!), the participant should make sure that some test data is held out to
make sure the model is not overfitting.
- Can we fine-tune for high-resource languages?
Yes! While we do not really recommend people to fine-tune models in English since there are
already so many fine-tuned speech recognition models in English. However, it is very much
appreciated if participants want to fine-tune models in other "high-resource" languages, such
as French, Spanish, or German. For such cases, one probably needs to train locally and apply
might have to apply tricks such as lazy data loading (check the ["Lazy data loading"](#how-to-do-lazy-data-loading) section for more details).
| {"language": "vi", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice, infore_25h"], "metrics": ["wer"], "model-index": [{"name": "Cuong-Cong XLSR Wav2Vec2 Large 53", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice vi", "type": "common_voice", "args": "vi"}, "metrics": [{"type": "wer", "value": 58.63, "name": "Test WER"}]}]}]} | CuongLD/wav2vec2-large-xlsr-vietnamese | null | [
"transformers",
"pytorch",
"jax",
"wav2vec2",
"automatic-speech-recognition",
"audio",
"speech",
"xlsr-fine-tuning-week",
"vi",
"arxiv:2006.11477",
"arxiv:2006.13979",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2006.11477",
"2006.13979"
] | [
"vi"
] | TAGS
#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #vi #arxiv-2006.11477 #arxiv-2006.13979 #license-apache-2.0 #model-index #endpoints_compatible #region-us
|
# Wav2Vec2-Large-XLSR-53-Vietnamese
Fine-tuned facebook/wav2vec2-large-xlsr-53 on Vietnamese using the Common Voice, Infore_25h dataset (Password: BroughtToYouByInfoRe)
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
## Evaluation
The model can be evaluated as follows on the Vietnamese test data of Common Voice.
Test Result: 58.63 %
## Training
The Common Voice 'train', 'validation', and 'Infore_25h' datasets were used for training
The script used for training can be found here
=======================To here===============================>
Your model in then available under *URL for everybody to use .
## How to evaluate my trained checkpoint
Having uploaded your model, you should now evaluate your model in a final step. This should be as simple as
copying the evaluation code of your model card into a python script and running it. Make sure to note
the final result on the model card both under the YAML tags at the very top and below your evaluation code under "Test Results".
## Rules of training and evaluation
In this section, we will quickly go over what data is allowed to be used as training
data, what kind of data preprocessing is allowed be used, and how the model should be evaluated.
To make it very simple regarding the first point: All data except the official common voice 'test' data set can be used as training data. For models trained in a language that is not included in Common Voice, the author of the model is responsible to
leave a reasonable amount of data for evaluation.
Second, the rules regarding the preprocessing are not that as straight-forward. It is allowed (and recommended) to
normalize the data to only have lower-case characters. It is also allowed (and recommended) to remove typographical
symbols and punctuation marks. A list of such symbols can *e.g.* be fonud here - however here we already must be careful. We should not remove a symbol that
would change the meaning of the words, *e.g.* in English, we should not remove the single quotation mark ''' since it
would change the meaning of the word '"it's"' to '"its"' which would then be incorrect. So the golden rule here is to
not remove any characters that could change the meaning of a word into another word. This is not always obvious and should
be given some consideration. As another example, it is fine to remove the "Hypen-minus" sign "'-'" since it doesn't change the
meaninng of a word to another one. *E.g.* "'fine-tuning'" would be changed to "'finetuning'" which has still the same meaning.
Since those choices are not always obvious when in doubt feel free to ask on Slack or even better post on the forum, as was
done, *e.g.* here.
## Tips and tricks
This section summarizes a couple of tips and tricks across various topics. It will continously be updated during the week.
### How to combine multiple datasets into one
Check out this post.
### How to effectively preprocess the data
### How to do efficiently load datasets with limited ram and hard drive space
Check out this post.
### How to do hyperparameter tuning
### How to preprocess and evaluate character based languages
## Further reading material
It is recommended that take some time to read up on how Wav2vec2 works in theory.
Getting a better understanding of the theory and the inner mechanisms of the model often helps when fine-tuning the model.
However, if you don't like reading blog posts/papers, don't worry - it is by no means necessary to go through the theory to fine-tune Wav2Vec2 on your language of choice.
If you are interested in learning more about the model though, here are a couple of resources that are important to better understand Wav2Vec2:
- Facebook's Wav2Vec2 blog post
- Official Wav2Vec2 paper
- Official XLSR Wav2vec2 paper
- Hugging Face Blog
- How does CTC (Connectionist Temporal Classification) work
It helps to have a good understanding of the following points:
- How was XLSR-Wav2Vec2 pretrained? -> Feature vectors were masked and had to be predicted by the model; very similar in spirit to masked language model of BERT.
- What parts of XLSR-Wav2Vec2 are responsible for what? What is the feature extractor part used for? -> extract feature vectors from the 1D raw audio waveform; What is the transformer part doing? -> mapping feature vectors to contextualized feature vectors; ...
- What part of the model needs to be fine-tuned? -> The pretrained model does not include a language head to classify the contextualized features to letters. This is randomly initialized when loading the pretrained checkpoint and has to be fine-tuned. Also, note that the authors recommend to not further fine-tune the feature extractor.
- What data was used to XLSR-Wav2Vec2? The checkpoint we will use for further fine-tuning was pretrained on 53 languages.
- What languages are considered to be similar by XLSR-Wav2Vec2? In the official XLSR Wav2Vec2 paper, the authors show nicely which languages share a common contextualized latent space. It might be useful for you to extend your training data with data of other languages that are considered to be very similar by the model (or you).
## FAQ
- Can a participant fine-tune models for more than one language?
Yes! A participant can fine-tune models in as many languages she/he likes
- Can a participant use extra data (apart from the common voice data)?
Yes! All data except the official common voice 'test data' can be used for training.
If a participant wants to train a model on a language that is not part of Common Voice (which
is very much encouraged!), the participant should make sure that some test data is held out to
make sure the model is not overfitting.
- Can we fine-tune for high-resource languages?
Yes! While we do not really recommend people to fine-tune models in English since there are
already so many fine-tuned speech recognition models in English. However, it is very much
appreciated if participants want to fine-tune models in other "high-resource" languages, such
as French, Spanish, or German. For such cases, one probably needs to train locally and apply
might have to apply tricks such as lazy data loading (check the "Lazy data loading" section for more details).
| [
"# Wav2Vec2-Large-XLSR-53-Vietnamese \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Vietnamese using the Common Voice, Infore_25h dataset (Password: BroughtToYouByInfoRe)\n \nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Vietnamese test data of Common Voice. \n\n\n\nTest Result: 58.63 %",
"## Training\n\nThe Common Voice 'train', 'validation', and 'Infore_25h' datasets were used for training \n\nThe script used for training can be found here \n\n=======================To here===============================>\n\nYour model in then available under *URL for everybody to use .",
"## How to evaluate my trained checkpoint\n\nHaving uploaded your model, you should now evaluate your model in a final step. This should be as simple as \ncopying the evaluation code of your model card into a python script and running it. Make sure to note \nthe final result on the model card both under the YAML tags at the very top and below your evaluation code under \"Test Results\".",
"## Rules of training and evaluation\n\nIn this section, we will quickly go over what data is allowed to be used as training \ndata, what kind of data preprocessing is allowed be used, and how the model should be evaluated.\n\nTo make it very simple regarding the first point: All data except the official common voice 'test' data set can be used as training data. For models trained in a language that is not included in Common Voice, the author of the model is responsible to \nleave a reasonable amount of data for evaluation.\n\nSecond, the rules regarding the preprocessing are not that as straight-forward. It is allowed (and recommended) to \nnormalize the data to only have lower-case characters. It is also allowed (and recommended) to remove typographical \nsymbols and punctuation marks. A list of such symbols can *e.g.* be fonud here - however here we already must be careful. We should not remove a symbol that \nwould change the meaning of the words, *e.g.* in English, we should not remove the single quotation mark ''' since it \nwould change the meaning of the word '\"it's\"' to '\"its\"' which would then be incorrect. So the golden rule here is to \nnot remove any characters that could change the meaning of a word into another word. This is not always obvious and should \nbe given some consideration. As another example, it is fine to remove the \"Hypen-minus\" sign \"'-'\" since it doesn't change the \nmeaninng of a word to another one. *E.g.* \"'fine-tuning'\" would be changed to \"'finetuning'\" which has still the same meaning.\n\nSince those choices are not always obvious when in doubt feel free to ask on Slack or even better post on the forum, as was \ndone, *e.g.* here.",
"## Tips and tricks\n\nThis section summarizes a couple of tips and tricks across various topics. It will continously be updated during the week.",
"### How to combine multiple datasets into one\n\nCheck out this post.",
"### How to effectively preprocess the data",
"### How to do efficiently load datasets with limited ram and hard drive space\n\nCheck out this post.",
"### How to do hyperparameter tuning",
"### How to preprocess and evaluate character based languages",
"## Further reading material\n\nIt is recommended that take some time to read up on how Wav2vec2 works in theory. \nGetting a better understanding of the theory and the inner mechanisms of the model often helps when fine-tuning the model. \n\nHowever, if you don't like reading blog posts/papers, don't worry - it is by no means necessary to go through the theory to fine-tune Wav2Vec2 on your language of choice.\n\nIf you are interested in learning more about the model though, here are a couple of resources that are important to better understand Wav2Vec2:\n\n- Facebook's Wav2Vec2 blog post\n- Official Wav2Vec2 paper\n- Official XLSR Wav2vec2 paper\n- Hugging Face Blog\n- How does CTC (Connectionist Temporal Classification) work\n\nIt helps to have a good understanding of the following points:\n\n- How was XLSR-Wav2Vec2 pretrained? -> Feature vectors were masked and had to be predicted by the model; very similar in spirit to masked language model of BERT.\n\n- What parts of XLSR-Wav2Vec2 are responsible for what? What is the feature extractor part used for? -> extract feature vectors from the 1D raw audio waveform; What is the transformer part doing? -> mapping feature vectors to contextualized feature vectors; ...\n\n- What part of the model needs to be fine-tuned? -> The pretrained model does not include a language head to classify the contextualized features to letters. This is randomly initialized when loading the pretrained checkpoint and has to be fine-tuned. Also, note that the authors recommend to not further fine-tune the feature extractor.\n\n- What data was used to XLSR-Wav2Vec2? The checkpoint we will use for further fine-tuning was pretrained on 53 languages. \n\n- What languages are considered to be similar by XLSR-Wav2Vec2? In the official XLSR Wav2Vec2 paper, the authors show nicely which languages share a common contextualized latent space. It might be useful for you to extend your training data with data of other languages that are considered to be very similar by the model (or you).",
"## FAQ\n\n- Can a participant fine-tune models for more than one language? \nYes! A participant can fine-tune models in as many languages she/he likes\n- Can a participant use extra data (apart from the common voice data)?\nYes! All data except the official common voice 'test data' can be used for training.\nIf a participant wants to train a model on a language that is not part of Common Voice (which \nis very much encouraged!), the participant should make sure that some test data is held out to \nmake sure the model is not overfitting.\n- Can we fine-tune for high-resource languages? \nYes! While we do not really recommend people to fine-tune models in English since there are\nalready so many fine-tuned speech recognition models in English. However, it is very much \nappreciated if participants want to fine-tune models in other \"high-resource\" languages, such \nas French, Spanish, or German. For such cases, one probably needs to train locally and apply \nmight have to apply tricks such as lazy data loading (check the \"Lazy data loading\" section for more details)."
] | [
"TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #vi #arxiv-2006.11477 #arxiv-2006.13979 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"# Wav2Vec2-Large-XLSR-53-Vietnamese \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Vietnamese using the Common Voice, Infore_25h dataset (Password: BroughtToYouByInfoRe)\n \nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Vietnamese test data of Common Voice. \n\n\n\nTest Result: 58.63 %",
"## Training\n\nThe Common Voice 'train', 'validation', and 'Infore_25h' datasets were used for training \n\nThe script used for training can be found here \n\n=======================To here===============================>\n\nYour model in then available under *URL for everybody to use .",
"## How to evaluate my trained checkpoint\n\nHaving uploaded your model, you should now evaluate your model in a final step. This should be as simple as \ncopying the evaluation code of your model card into a python script and running it. Make sure to note \nthe final result on the model card both under the YAML tags at the very top and below your evaluation code under \"Test Results\".",
"## Rules of training and evaluation\n\nIn this section, we will quickly go over what data is allowed to be used as training \ndata, what kind of data preprocessing is allowed be used, and how the model should be evaluated.\n\nTo make it very simple regarding the first point: All data except the official common voice 'test' data set can be used as training data. For models trained in a language that is not included in Common Voice, the author of the model is responsible to \nleave a reasonable amount of data for evaluation.\n\nSecond, the rules regarding the preprocessing are not that as straight-forward. It is allowed (and recommended) to \nnormalize the data to only have lower-case characters. It is also allowed (and recommended) to remove typographical \nsymbols and punctuation marks. A list of such symbols can *e.g.* be fonud here - however here we already must be careful. We should not remove a symbol that \nwould change the meaning of the words, *e.g.* in English, we should not remove the single quotation mark ''' since it \nwould change the meaning of the word '\"it's\"' to '\"its\"' which would then be incorrect. So the golden rule here is to \nnot remove any characters that could change the meaning of a word into another word. This is not always obvious and should \nbe given some consideration. As another example, it is fine to remove the \"Hypen-minus\" sign \"'-'\" since it doesn't change the \nmeaninng of a word to another one. *E.g.* \"'fine-tuning'\" would be changed to \"'finetuning'\" which has still the same meaning.\n\nSince those choices are not always obvious when in doubt feel free to ask on Slack or even better post on the forum, as was \ndone, *e.g.* here.",
"## Tips and tricks\n\nThis section summarizes a couple of tips and tricks across various topics. It will continously be updated during the week.",
"### How to combine multiple datasets into one\n\nCheck out this post.",
"### How to effectively preprocess the data",
"### How to do efficiently load datasets with limited ram and hard drive space\n\nCheck out this post.",
"### How to do hyperparameter tuning",
"### How to preprocess and evaluate character based languages",
"## Further reading material\n\nIt is recommended that take some time to read up on how Wav2vec2 works in theory. \nGetting a better understanding of the theory and the inner mechanisms of the model often helps when fine-tuning the model. \n\nHowever, if you don't like reading blog posts/papers, don't worry - it is by no means necessary to go through the theory to fine-tune Wav2Vec2 on your language of choice.\n\nIf you are interested in learning more about the model though, here are a couple of resources that are important to better understand Wav2Vec2:\n\n- Facebook's Wav2Vec2 blog post\n- Official Wav2Vec2 paper\n- Official XLSR Wav2vec2 paper\n- Hugging Face Blog\n- How does CTC (Connectionist Temporal Classification) work\n\nIt helps to have a good understanding of the following points:\n\n- How was XLSR-Wav2Vec2 pretrained? -> Feature vectors were masked and had to be predicted by the model; very similar in spirit to masked language model of BERT.\n\n- What parts of XLSR-Wav2Vec2 are responsible for what? What is the feature extractor part used for? -> extract feature vectors from the 1D raw audio waveform; What is the transformer part doing? -> mapping feature vectors to contextualized feature vectors; ...\n\n- What part of the model needs to be fine-tuned? -> The pretrained model does not include a language head to classify the contextualized features to letters. This is randomly initialized when loading the pretrained checkpoint and has to be fine-tuned. Also, note that the authors recommend to not further fine-tune the feature extractor.\n\n- What data was used to XLSR-Wav2Vec2? The checkpoint we will use for further fine-tuning was pretrained on 53 languages. \n\n- What languages are considered to be similar by XLSR-Wav2Vec2? In the official XLSR Wav2Vec2 paper, the authors show nicely which languages share a common contextualized latent space. It might be useful for you to extend your training data with data of other languages that are considered to be very similar by the model (or you).",
"## FAQ\n\n- Can a participant fine-tune models for more than one language? \nYes! A participant can fine-tune models in as many languages she/he likes\n- Can a participant use extra data (apart from the common voice data)?\nYes! All data except the official common voice 'test data' can be used for training.\nIf a participant wants to train a model on a language that is not part of Common Voice (which \nis very much encouraged!), the participant should make sure that some test data is held out to \nmake sure the model is not overfitting.\n- Can we fine-tune for high-resource languages? \nYes! While we do not really recommend people to fine-tune models in English since there are\nalready so many fine-tuned speech recognition models in English. However, it is very much \nappreciated if participants want to fine-tune models in other \"high-resource\" languages, such \nas French, Spanish, or German. For such cases, one probably needs to train locally and apply \nmight have to apply tricks such as lazy data loading (check the \"Lazy data loading\" section for more details)."
] | [
79,
77,
18,
26,
109,
78,
399,
31,
17,
12,
23,
10,
14,
481,
232
] | [
"TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #vi #arxiv-2006.11477 #arxiv-2006.13979 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Vietnamese \n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Vietnamese using the Common Voice, Infore_25h dataset (Password: BroughtToYouByInfoRe)\n \nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Vietnamese test data of Common Voice. \n\n\n\nTest Result: 58.63 %## Training\n\nThe Common Voice 'train', 'validation', and 'Infore_25h' datasets were used for training \n\nThe script used for training can be found here \n\n=======================To here===============================>\n\nYour model in then available under *URL for everybody to use .## How to evaluate my trained checkpoint\n\nHaving uploaded your model, you should now evaluate your model in a final step. This should be as simple as \ncopying the evaluation code of your model card into a python script and running it. Make sure to note \nthe final result on the model card both under the YAML tags at the very top and below your evaluation code under \"Test Results\".## Rules of training and evaluation\n\nIn this section, we will quickly go over what data is allowed to be used as training \ndata, what kind of data preprocessing is allowed be used, and how the model should be evaluated.\n\nTo make it very simple regarding the first point: All data except the official common voice 'test' data set can be used as training data. For models trained in a language that is not included in Common Voice, the author of the model is responsible to \nleave a reasonable amount of data for evaluation.\n\nSecond, the rules regarding the preprocessing are not that as straight-forward. It is allowed (and recommended) to \nnormalize the data to only have lower-case characters. It is also allowed (and recommended) to remove typographical \nsymbols and punctuation marks. A list of such symbols can *e.g.* be fonud here - however here we already must be careful. We should not remove a symbol that \nwould change the meaning of the words, *e.g.* in English, we should not remove the single quotation mark ''' since it \nwould change the meaning of the word '\"it's\"' to '\"its\"' which would then be incorrect. So the golden rule here is to \nnot remove any characters that could change the meaning of a word into another word. This is not always obvious and should \nbe given some consideration. As another example, it is fine to remove the \"Hypen-minus\" sign \"'-'\" since it doesn't change the \nmeaninng of a word to another one. *E.g.* \"'fine-tuning'\" would be changed to \"'finetuning'\" which has still the same meaning.\n\nSince those choices are not always obvious when in doubt feel free to ask on Slack or even better post on the forum, as was \ndone, *e.g.* here.## Tips and tricks\n\nThis section summarizes a couple of tips and tricks across various topics. It will continously be updated during the week.### How to combine multiple datasets into one\n\nCheck out this post.### How to effectively preprocess the data### How to do efficiently load datasets with limited ram and hard drive space\n\nCheck out this post.### How to do hyperparameter tuning### How to preprocess and evaluate character based languages## Further reading material\n\nIt is recommended that take some time to read up on how Wav2vec2 works in theory. \nGetting a better understanding of the theory and the inner mechanisms of the model often helps when fine-tuning the model. \n\nHowever, if you don't like reading blog posts/papers, don't worry - it is by no means necessary to go through the theory to fine-tune Wav2Vec2 on your language of choice.\n\nIf you are interested in learning more about the model though, here are a couple of resources that are important to better understand Wav2Vec2:\n\n- Facebook's Wav2Vec2 blog post\n- Official Wav2Vec2 paper\n- Official XLSR Wav2vec2 paper\n- Hugging Face Blog\n- How does CTC (Connectionist Temporal Classification) work\n\nIt helps to have a good understanding of the following points:\n\n- How was XLSR-Wav2Vec2 pretrained? -> Feature vectors were masked and had to be predicted by the model; very similar in spirit to masked language model of BERT.\n\n- What parts of XLSR-Wav2Vec2 are responsible for what? What is the feature extractor part used for? -> extract feature vectors from the 1D raw audio waveform; What is the transformer part doing? -> mapping feature vectors to contextualized feature vectors; ...\n\n- What part of the model needs to be fine-tuned? -> The pretrained model does not include a language head to classify the contextualized features to letters. This is randomly initialized when loading the pretrained checkpoint and has to be fine-tuned. Also, note that the authors recommend to not further fine-tune the feature extractor.\n\n- What data was used to XLSR-Wav2Vec2? The checkpoint we will use for further fine-tuning was pretrained on 53 languages. \n\n- What languages are considered to be similar by XLSR-Wav2Vec2? In the official XLSR Wav2Vec2 paper, the authors show nicely which languages share a common contextualized latent space. It might be useful for you to extend your training data with data of other languages that are considered to be very similar by the model (or you).## FAQ\n\n- Can a participant fine-tune models for more than one language? \nYes! A participant can fine-tune models in as many languages she/he likes\n- Can a participant use extra data (apart from the common voice data)?\nYes! All data except the official common voice 'test data' can be used for training.\nIf a participant wants to train a model on a language that is not part of Common Voice (which \nis very much encouraged!), the participant should make sure that some test data is held out to \nmake sure the model is not overfitting.\n- Can we fine-tune for high-resource languages? \nYes! While we do not really recommend people to fine-tune models in English since there are\nalready so many fine-tuned speech recognition models in English. However, it is very much \nappreciated if participants want to fine-tune models in other \"high-resource\" languages, such \nas French, Spanish, or German. For such cases, one probably needs to train locally and apply \nmight have to apply tricks such as lazy data loading (check the \"Lazy data loading\" section for more details)."
] |
text-generation | null |
# Sora DialoGPT Model | {"tags": ["conversational"]} | CurtisBowser/DialoGPT-medium-sora-two | null | [
"pytorch",
"conversational",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#pytorch #conversational #region-us
|
# Sora DialoGPT Model | [
"# Sora DialoGPT Model"
] | [
"TAGS\n#pytorch #conversational #region-us \n",
"# Sora DialoGPT Model"
] | [
13,
7
] | [
"TAGS\n#pytorch #conversational #region-us \n# Sora DialoGPT Model"
] |
text-generation | transformers |
# Sora DialoGPT Model
| {"tags": ["conversational"]} | CurtisBowser/DialoGPT-medium-sora | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us",
"has_space"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #has_space
|
# Sora DialoGPT Model
| [
"# Sora DialoGPT Model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #has_space \n",
"# Sora DialoGPT Model"
] | [
43,
7
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us #has_space \n# Sora DialoGPT Model"
] |
text-generation | transformers |
# Sora DialoGPT Model | {"tags": ["conversational"]} | CurtisBowser/DialoGPT-small-sora | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Sora DialoGPT Model | [
"# Sora DialoGPT Model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Sora DialoGPT Model"
] | [
39,
7
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Sora DialoGPT Model"
] |
text-generation | transformers |
# Chandler Bot DialoGPT model | {"tags": ["conversational"]} | CyberMuffin/DialoGPT-small-ChandlerBot | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Chandler Bot DialoGPT model | [
"# Chandler Bot DialoGPT model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Chandler Bot DialoGPT model"
] | [
39,
7
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Chandler Bot DialoGPT model"
] |
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# electra-base-discriminator-finetuned-cola
This model is a fine-tuned version of [google/electra-base-discriminator](https://huggingface.co/google/electra-base-discriminator) on the glue dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6367
- Matthews Correlation: 0.6824
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Matthews Correlation |
|:-------------:|:-----:|:----:|:---------------:|:--------------------:|
| 0.4139 | 1.0 | 535 | 0.4137 | 0.6381 |
| 0.2452 | 2.0 | 1070 | 0.4887 | 0.6504 |
| 0.17 | 3.0 | 1605 | 0.5335 | 0.6757 |
| 0.1135 | 4.0 | 2140 | 0.6367 | 0.6824 |
| 0.0817 | 5.0 | 2675 | 0.6742 | 0.6755 |
### Framework versions
- Transformers 4.15.0
- Pytorch 1.10.0+cu111
- Datasets 1.17.0
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "electra-base-discriminator-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.6824089073723449, "name": "Matthews Correlation"}]}]}]} | D3xter1922/electra-base-discriminator-finetuned-cola | null | [
"transformers",
"pytorch",
"tensorboard",
"electra",
"text-classification",
"generated_from_trainer",
"dataset:glue",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #electra #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| electra-base-discriminator-finetuned-cola
=========================================
This model is a fine-tuned version of google/electra-base-discriminator on the glue dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6367
* Matthews Correlation: 0.6824
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.15.0
* Pytorch 1.10.0+cu111
* Datasets 1.17.0
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #electra #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] | [
55,
101,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #electra #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
text-generation | transformers |
# Anakin Skywalker DialoGPT Model | {"tags": ["conversational"]} | DARKVIP3R/DialoGPT-medium-Anakin | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Anakin Skywalker DialoGPT Model | [
"# Anakin Skywalker DialoGPT Model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Anakin Skywalker DialoGPT Model"
] | [
39,
9
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Anakin Skywalker DialoGPT Model"
] |
fill-mask | transformers |
# bert-base-irish-cased-v1
[gaBERT](https://aclanthology.org/2022.lrec-1.511/) is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.
## Model description
Encoder-based Transformer to be used to obtain features for finetuning for downstream tasks in Irish.
## Intended uses & limitations
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: None
- training_precision: float32
### Framework versions
- Transformers 4.20.1
- TensorFlow 2.9.1
- Datasets 2.3.2
- Tokenizers 0.12.1
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
```
@inproceedings{barry-etal-2022-gabert,
title = "ga{BERT} {---} an {I}rish Language Model",
author = "Barry, James and
Wagner, Joachim and
Cassidy, Lauren and
Cowap, Alan and
Lynn, Teresa and
Walsh, Abigail and
{\'O} Meachair, M{\'\i}che{\'a}l J. and
Foster, Jennifer",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.511",
pages = "4774--4788",
abstract = "The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many NLP tasks. We introduce gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and the monolingual Irish WikiBERT, and we show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We compare the results of fine-tuning a gaBERT model with an mBERT model for the task of identifying verbal multiword expressions, and show that the fine-tuned gaBERT model also performs better at this task. We release gaBERT and related code to the community.",
}
```
| {"tags": ["generated_from_keras_callback"], "widget": [{"text": "Ceolt\u00f3ir [MASK] ab ea Johnny Cash."}], "model-index": [{"name": "bert-base-irish-cased-v1", "results": []}]} | DCU-NLP/bert-base-irish-cased-v1 | null | [
"transformers",
"pytorch",
"tf",
"bert",
"fill-mask",
"generated_from_keras_callback",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #bert #fill-mask #generated_from_keras_callback #autotrain_compatible #endpoints_compatible #region-us
|
# bert-base-irish-cased-v1
gaBERT is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.
## Model description
Encoder-based Transformer to be used to obtain features for finetuning for downstream tasks in Irish.
## Intended uses & limitations
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: None
- training_precision: float32
### Framework versions
- Transformers 4.20.1
- TensorFlow 2.9.1
- Datasets 2.3.2
- Tokenizers 0.12.1
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
| [
"# bert-base-irish-cased-v1\n\ngaBERT is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.",
"## Model description\n\nEncoder-based Transformer to be used to obtain features for finetuning for downstream tasks in Irish.",
"## Intended uses & limitations\n\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32",
"### Framework versions\n\n- Transformers 4.20.1\n- TensorFlow 2.9.1\n- Datasets 2.3.2\n- Tokenizers 0.12.1",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
"TAGS\n#transformers #pytorch #tf #bert #fill-mask #generated_from_keras_callback #autotrain_compatible #endpoints_compatible #region-us \n",
"# bert-base-irish-cased-v1\n\ngaBERT is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.",
"## Model description\n\nEncoder-based Transformer to be used to obtain features for finetuning for downstream tasks in Irish.",
"## Intended uses & limitations\n\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32",
"### Framework versions\n\n- Transformers 4.20.1\n- TensorFlow 2.9.1\n- Datasets 2.3.2\n- Tokenizers 0.12.1",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
41,
52,
27,
57,
32,
38,
25
] | [
"TAGS\n#transformers #pytorch #tf #bert #fill-mask #generated_from_keras_callback #autotrain_compatible #endpoints_compatible #region-us \n# bert-base-irish-cased-v1\n\ngaBERT is a BERT-base model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper.## Model description\n\nEncoder-based Transformer to be used to obtain features for finetuning for downstream tasks in Irish.## Intended uses & limitations\n\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32### Framework versions\n\n- Transformers 4.20.1\n- TensorFlow 2.9.1\n- Datasets 2.3.2\n- Tokenizers 0.12.1### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] |
null | transformers |
# gaELECTRA
[gaELECTRA](https://aclanthology.org/2022.lrec-1.511/) is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.
### Limitations and bias
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
```
@inproceedings{barry-etal-2022-gabert,
title = "ga{BERT} {---} an {I}rish Language Model",
author = "Barry, James and
Wagner, Joachim and
Cassidy, Lauren and
Cowap, Alan and
Lynn, Teresa and
Walsh, Abigail and
{\'O} Meachair, M{\'\i}che{\'a}l J. and
Foster, Jennifer",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.511",
pages = "4774--4788",
abstract = "The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many NLP tasks. We introduce gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and the monolingual Irish WikiBERT, and we show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We compare the results of fine-tuning a gaBERT model with an mBERT model for the task of identifying verbal multiword expressions, and show that the fine-tuned gaBERT model also performs better at this task. We release gaBERT and related code to the community.",
}
``` | {"language": ["ga"], "license": "apache-2.0", "tags": ["irish", "electra"], "widget": [{"text": "Ceolt\u00f3ir [MASK] ab ea Johnny Cash."}]} | DCU-NLP/electra-base-irish-cased-discriminator-v1 | null | [
"transformers",
"pytorch",
"electra",
"pretraining",
"irish",
"ga",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"ga"
] | TAGS
#transformers #pytorch #electra #pretraining #irish #ga #license-apache-2.0 #endpoints_compatible #region-us
|
# gaELECTRA
gaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.
### Limitations and bias
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
| [
"# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.",
"### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
"TAGS\n#transformers #pytorch #electra #pretraining #irish #ga #license-apache-2.0 #endpoints_compatible #region-us \n",
"# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.",
"### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
36,
71,
57,
25
] | [
"TAGS\n#transformers #pytorch #electra #pretraining #irish #ga #license-apache-2.0 #endpoints_compatible #region-us \n# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] |
fill-mask | transformers |
# gaELECTRA
[gaELECTRA](https://aclanthology.org/2022.lrec-1.511/) is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.
### Limitations and bias
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
```
@inproceedings{barry-etal-2022-gabert,
title = "ga{BERT} {---} an {I}rish Language Model",
author = "Barry, James and
Wagner, Joachim and
Cassidy, Lauren and
Cowap, Alan and
Lynn, Teresa and
Walsh, Abigail and
{\'O} Meachair, M{\'\i}che{\'a}l J. and
Foster, Jennifer",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://aclanthology.org/2022.lrec-1.511",
pages = "4774--4788",
abstract = "The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many NLP tasks. We introduce gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and the monolingual Irish WikiBERT, and we show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We compare the results of fine-tuning a gaBERT model with an mBERT model for the task of identifying verbal multiword expressions, and show that the fine-tuned gaBERT model also performs better at this task. We release gaBERT and related code to the community.",
}
``` | {"language": ["ga"], "license": "apache-2.0", "tags": ["irish", "electra"], "widget": [{"text": "Ceolt\u00f3ir [MASK] ab ea Johnny Cash."}]} | DCU-NLP/electra-base-irish-cased-generator-v1 | null | [
"transformers",
"pytorch",
"electra",
"fill-mask",
"irish",
"ga",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"ga"
] | TAGS
#transformers #pytorch #electra #fill-mask #irish #ga #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# gaELECTRA
gaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.
### Limitations and bias
Some data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.
### BibTeX entry and citation info
If you use this model in your research, please consider citing our paper:
| [
"# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.",
"### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
"TAGS\n#transformers #pytorch #electra #fill-mask #irish #ga #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.",
"### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.",
"### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] | [
45,
71,
57,
25
] | [
"TAGS\n#transformers #pytorch #electra #fill-mask #irish #ga #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# gaELECTRA\ngaELECTRA is an ELECTRA model trained on 7.9M Irish sentences. For more details, including the hyperparameters and pretraining corpora used please refer to our paper. For fine-tuning this model on a token classification task, e.g. Named Entity Recognition, use the discriminator model.### Limitations and bias\nSome data used to pretrain gaBERT was scraped from the web which potentially contains ethically problematic text (bias, hate, adult content, etc.). Consequently, downstream tasks/applications using gaBERT should be thoroughly tested with respect to ethical considerations.### BibTeX entry and citation info\nIf you use this model in your research, please consider citing our paper:"
] |
fill-mask | transformers |
# Danish BERT (uncased) model
[BotXO.ai](https://www.botxo.ai/) developed this model. For data and training details see their [GitHub repository](https://github.com/botxo/nordic_bert).
The original model was trained in TensorFlow then I converted it to Pytorch using [transformers-cli](https://huggingface.co/transformers/converting_tensorflow_models.html?highlight=cli).
For TensorFlow version download here: https://www.dropbox.com/s/19cjaoqvv2jicq9/danish_bert_uncased_v2.zip?dl=1
## Architecture
```python
from transformers import AutoModelForPreTraining
model = AutoModelForPreTraining.from_pretrained("DJSammy/bert-base-danish-uncased_BotXO,ai")
params = list(model.named_parameters())
print('danish_bert_uncased_v2 has {:} different named parameters.\n'.format(len(params)))
print('==== Embedding Layer ====\n')
for p in params[0:5]:
print("{:<55} {:>12}".format(p[0], str(tuple(p[1].size()))))
print('\n==== First Transformer ====\n')
for p in params[5:21]:
print("{:<55} {:>12}".format(p[0], str(tuple(p[1].size()))))
print('\n==== Last Transformer ====\n')
for p in params[181:197]:
print("{:<55} {:>12}".format(p[0], str(tuple(p[1].size()))))
print('\n==== Output Layer ====\n')
for p in params[197:]:
print("{:<55} {:>12}".format(p[0], str(tuple(p[1].size()))))
# danish_bert_uncased_v2 has 206 different named parameters.
# ==== Embedding Layer ====
# bert.embeddings.word_embeddings.weight (32000, 768)
# bert.embeddings.position_embeddings.weight (512, 768)
# bert.embeddings.token_type_embeddings.weight (2, 768)
# bert.embeddings.LayerNorm.weight (768,)
# bert.embeddings.LayerNorm.bias (768,)
# ==== First Transformer ====
# bert.encoder.layer.0.attention.self.query.weight (768, 768)
# bert.encoder.layer.0.attention.self.query.bias (768,)
# bert.encoder.layer.0.attention.self.key.weight (768, 768)
# bert.encoder.layer.0.attention.self.key.bias (768,)
# bert.encoder.layer.0.attention.self.value.weight (768, 768)
# bert.encoder.layer.0.attention.self.value.bias (768,)
# bert.encoder.layer.0.attention.output.dense.weight (768, 768)
# bert.encoder.layer.0.attention.output.dense.bias (768,)
# bert.encoder.layer.0.attention.output.LayerNorm.weight (768,)
# bert.encoder.layer.0.attention.output.LayerNorm.bias (768,)
# bert.encoder.layer.0.intermediate.dense.weight (3072, 768)
# bert.encoder.layer.0.intermediate.dense.bias (3072,)
# bert.encoder.layer.0.output.dense.weight (768, 3072)
# bert.encoder.layer.0.output.dense.bias (768,)
# bert.encoder.layer.0.output.LayerNorm.weight (768,)
# bert.encoder.layer.0.output.LayerNorm.bias (768,)
# ==== Last Transformer ====
# bert.encoder.layer.11.attention.self.query.weight (768, 768)
# bert.encoder.layer.11.attention.self.query.bias (768,)
# bert.encoder.layer.11.attention.self.key.weight (768, 768)
# bert.encoder.layer.11.attention.self.key.bias (768,)
# bert.encoder.layer.11.attention.self.value.weight (768, 768)
# bert.encoder.layer.11.attention.self.value.bias (768,)
# bert.encoder.layer.11.attention.output.dense.weight (768, 768)
# bert.encoder.layer.11.attention.output.dense.bias (768,)
# bert.encoder.layer.11.attention.output.LayerNorm.weight (768,)
# bert.encoder.layer.11.attention.output.LayerNorm.bias (768,)
# bert.encoder.layer.11.intermediate.dense.weight (3072, 768)
# bert.encoder.layer.11.intermediate.dense.bias (3072,)
# bert.encoder.layer.11.output.dense.weight (768, 3072)
# bert.encoder.layer.11.output.dense.bias (768,)
# bert.encoder.layer.11.output.LayerNorm.weight (768,)
# bert.encoder.layer.11.output.LayerNorm.bias (768,)
# ==== Output Layer ====
# bert.pooler.dense.weight (768, 768)
# bert.pooler.dense.bias (768,)
# cls.predictions.bias (32000,)
# cls.predictions.transform.dense.weight (768, 768)
# cls.predictions.transform.dense.bias (768,)
# cls.predictions.transform.LayerNorm.weight (768,)
# cls.predictions.transform.LayerNorm.bias (768,)
# cls.seq_relationship.weight (2, 768)
# cls.seq_relationship.bias (2,)
```
## Example Pipeline
```python
from transformers import pipeline
unmasker = pipeline('fill-mask', model='DJSammy/bert-base-danish-uncased_BotXO,ai')
unmasker('København er [MASK] i Danmark.')
# Copenhagen is the [MASK] of Denmark.
# =>
# [{'score': 0.788068950176239,
# 'sequence': '[CLS] københavn er hovedstad i danmark. [SEP]',
# 'token': 12610,
# 'token_str': 'hovedstad'},
# {'score': 0.07606703042984009,
# 'sequence': '[CLS] københavn er hovedstaden i danmark. [SEP]',
# 'token': 8108,
# 'token_str': 'hovedstaden'},
# {'score': 0.04299738258123398,
# 'sequence': '[CLS] københavn er metropol i danmark. [SEP]',
# 'token': 23305,
# 'token_str': 'metropol'},
# {'score': 0.008163209073245525,
# 'sequence': '[CLS] københavn er ikke i danmark. [SEP]',
# 'token': 89,
# 'token_str': 'ikke'},
# {'score': 0.006238455418497324,
# 'sequence': '[CLS] københavn er ogsa i danmark. [SEP]',
# 'token': 25253,
# 'token_str': 'ogsa'}]
```
| {"language": "da", "license": "cc-by-4.0", "tags": ["bert", "masked-lm"], "datasets": ["common_crawl", "wikipedia"], "pipeline_tag": "fill-mask", "widget": [{"text": "K\u00f8benhavn er [MASK] i Danmark."}]} | DJSammy/bert-base-danish-uncased_BotXO-ai | null | [
"transformers",
"pytorch",
"jax",
"bert",
"masked-lm",
"fill-mask",
"da",
"dataset:common_crawl",
"dataset:wikipedia",
"license:cc-by-4.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #jax #bert #masked-lm #fill-mask #da #dataset-common_crawl #dataset-wikipedia #license-cc-by-4.0 #endpoints_compatible #region-us
|
# Danish BERT (uncased) model
URL developed this model. For data and training details see their GitHub repository.
The original model was trained in TensorFlow then I converted it to Pytorch using transformers-cli.
For TensorFlow version download here: URL
## Architecture
## Example Pipeline
| [
"# Danish BERT (uncased) model \n\nURL developed this model. For data and training details see their GitHub repository. \n\nThe original model was trained in TensorFlow then I converted it to Pytorch using transformers-cli.\n\nFor TensorFlow version download here: URL",
"## Architecture",
"## Example Pipeline"
] | [
"TAGS\n#transformers #pytorch #jax #bert #masked-lm #fill-mask #da #dataset-common_crawl #dataset-wikipedia #license-cc-by-4.0 #endpoints_compatible #region-us \n",
"# Danish BERT (uncased) model \n\nURL developed this model. For data and training details see their GitHub repository. \n\nThe original model was trained in TensorFlow then I converted it to Pytorch using transformers-cli.\n\nFor TensorFlow version download here: URL",
"## Architecture",
"## Example Pipeline"
] | [
54,
58,
3,
4
] | [
"TAGS\n#transformers #pytorch #jax #bert #masked-lm #fill-mask #da #dataset-common_crawl #dataset-wikipedia #license-cc-by-4.0 #endpoints_compatible #region-us \n# Danish BERT (uncased) model \n\nURL developed this model. For data and training details see their GitHub repository. \n\nThe original model was trained in TensorFlow then I converted it to Pytorch using transformers-cli.\n\nFor TensorFlow version download here: URL## Architecture## Example Pipeline"
] |
text-classification | transformers | ** Human-Directed Sentiment Analysis in Arabic
A supervised training procedure to classify human-directed-sentiment in a text. We define the human-directed-sentiment as the polarity of one user towards a second person who is involved with him in a discussion. | {} | DSI/human-directed-sentiment | null | [
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
| Human-Directed Sentiment Analysis in Arabic
A supervised training procedure to classify human-directed-sentiment in a text. We define the human-directed-sentiment as the polarity of one user towards a second person who is involved with him in a discussion. | [] | [
"TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
28
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
text-classification | transformers |
# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT
[Blog post »](https://people.cs.kuleuven.be/~pieter.delobelle/attitudes-towards-covid-19-measures/?utm_source=huggingface&utm_medium=social&utm_campaign=corona_tweets) · [paper »](http://arxiv.org/abs/2104.09947)
This model can be used to determine if a tweet expresses support or not for a curfew. The model was trained on manually labeled tweets from Belgium in Dutch, French and English.
We categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).
![chart.png](https://github.com/iPieter/bert-corona-tweets/raw/master/chart.png)
Models used in this paper are on HuggingFace:
- https://huggingface.co/DTAI-KULeuven/mbert-corona-tweets-belgium-curfew-support
- https://huggingface.co/DTAI-KULeuven/mbert-corona-tweets-belgium-topics
| {"language": ["multilingual", "nl", "fr", "en"], "tags": ["Tweets", "Sentiment analysis"], "widget": [{"text": "I really wish I could leave my house after midnight, this makes no sense!"}]} | DTAI-KULeuven/mbert-corona-tweets-belgium-curfew-support | null | [
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"Tweets",
"Sentiment analysis",
"multilingual",
"nl",
"fr",
"en",
"arxiv:2104.09947",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2104.09947"
] | [
"multilingual",
"nl",
"fr",
"en"
] | TAGS
#transformers #pytorch #jax #bert #text-classification #Tweets #Sentiment analysis #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us
|
# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT
Blog post » · paper »
This model can be used to determine if a tweet expresses support or not for a curfew. The model was trained on manually labeled tweets from Belgium in Dutch, French and English.
We categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).
!URL
Models used in this paper are on HuggingFace:
- URL
- URL
| [
"# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nThis model can be used to determine if a tweet expresses support or not for a curfew. The model was trained on manually labeled tweets from Belgium in Dutch, French and English. \n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #Tweets #Sentiment analysis #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nThis model can be used to determine if a tweet expresses support or not for a curfew. The model was trained on manually labeled tweets from Belgium in Dutch, French and English. \n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] | [
58,
176
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #Tweets #Sentiment analysis #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us \n# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nThis model can be used to determine if a tweet expresses support or not for a curfew. The model was trained on manually labeled tweets from Belgium in Dutch, French and English. \n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] |
text-classification | transformers |
# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT
[Blog post »](https://people.cs.kuleuven.be/~pieter.delobelle/attitudes-towards-covid-19-measures/?utm_source=huggingface&utm_medium=social&utm_campaign=corona_tweets) · [paper »](http://arxiv.org/abs/2104.09947)
We categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).
![chart.png](https://github.com/iPieter/bert-corona-tweets/raw/master/chart.png)
Models used in this paper are on HuggingFace:
- https://huggingface.co/DTAI-KULeuven/mbert-corona-tweets-belgium-curfew-support
- https://huggingface.co/DTAI-KULeuven/mbert-corona-tweets-belgium-topics
| {"language": ["multilingual", "nl", "fr", "en"], "tags": ["Dutch", "French", "English", "Tweets", "Topic classification"], "widget": [{"text": "I really can't wait for this lockdown to be over and go back to waking up early."}]} | DTAI-KULeuven/mbert-corona-tweets-belgium-topics | null | [
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"Dutch",
"French",
"English",
"Tweets",
"Topic classification",
"multilingual",
"nl",
"fr",
"en",
"arxiv:2104.09947",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2104.09947"
] | [
"multilingual",
"nl",
"fr",
"en"
] | TAGS
#transformers #pytorch #jax #bert #text-classification #Dutch #French #English #Tweets #Topic classification #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us
|
# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT
Blog post » · paper »
We categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).
!URL
Models used in this paper are on HuggingFace:
- URL
- URL
| [
"# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #Dutch #French #English #Tweets #Topic classification #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] | [
64,
135
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #Dutch #French #English #Tweets #Topic classification #multilingual #nl #fr #en #arxiv-2104.09947 #autotrain_compatible #endpoints_compatible #region-us \n# Measuring Shifts in Attitudes Towards COVID-19 Measures in Belgium Using Multilingual BERT\nBlog post » · paper »\n\nWe categorized several months worth of these Tweets by topic (government COVID measure) and opinion expressed. Below is a timeline of the relative number of Tweets on the curfew topic (middle) and the fraction of those Tweets that find the curfew too strict, too loose, or a suitable measure (bottom), with the number of daily cases in Belgium to give context on the pandemic situation (top).\n\n!URL\n\n\nModels used in this paper are on HuggingFace: \n- URL \n- URL"
] |
fill-mask | transformers |
<p align="center">
<img src="https://github.com/iPieter/robbertje/raw/master/images/robbertje_logo_with_name.png" alt="RobBERTje: A collection of distilled Dutch BERT-based models" width="75%">
</p>
# About RobBERTje
RobBERTje is a collection of distilled models based on [RobBERT](http://github.com/iPieter/robbert). There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch [the repository](http://github.com/iPieter/robbertje) for updates.
# News
- **February 21, 2022**: Our paper about RobBERTje has been published in [volume 11 of CLIN journal](https://www.clinjournal.org/clinj/article/view/131)!
- **July 2, 2021**: Publicly released 4 RobBERTje models.
- **May 12, 2021**: RobBERTje was accepted at [CLIN31](https://www.clin31.ugent.be) for an oral presentation!
# The models
| Model | Description | Parameters | Training size | Huggingface id |
|--------------|-------------|------------------|-------------------|------------------------------------------------------------------------------------|
| Non-shuffled | Trained on the non-shuffled variant of the oscar corpus, without any operations to preserve this order during training and distillation. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-non-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-non-shuffled) |
| Shuffled | Trained on the publicly available and shuffled OSCAR corpus. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-shuffled) |
| Merged (p=0.5) | Same as the non-shuffled variant, but sequential sentences of the same document are merged with a probability of 50%. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-merged](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-merged) |
| BORT | A smaller version with 8 attention heads instead of 12 and 4 layers instead of 6 (and 12 for RobBERT). | 46 M | 1 GB | this model |
# Results
## Intrinsic results
We calculated the _pseudo perplexity_ (PPPL) from [cite](), which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
| Model | PPPL |
|-------------------|-----------|
| RobBERT (teacher) | 7.76 |
| Non-shuffled | 12.95 |
| Shuffled | 18.74 |
| Merged (p=0.5) | 17.10 |
| BORT | 26.44 |
## Extrinsic results
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a [Dutch NLI task named SICK-NL](https://arxiv.org/abs/2101.05716) was also released and we evaluated our models with it as well.
| Model | DBRD | DIE-DAT | NER | POS |SICK-NL |
|------------------|-----------|-----------|-----------|-----------|----------|
| RobBERT (teacher)|94.4 | 99.2 |89.1 |96.4 | 84.2 |
| Non-shuffled |90.2 | 98.4 |82.9 |95.5 | 83.4 |
| Shuffled |92.5 | 98.2 |82.7 |95.6 | 83.4 |
| Merged (p=0.5) |92.9 | 96.5 |81.8 |95.2 | 82.8 |
| BORT |89.6 | 92.2 |79.7 |94.3 | 81.0 |
| {"language": "nl", "license": "mit", "tags": ["Dutch", "Flemish", "RoBERTa", "RobBERT", "RobBERTje"], "datasets": ["oscar", "dbrd", "lassy-ud", "europarl-mono", "conll2002"], "thumbnail": "https://github.com/iPieter/RobBERT/raw/master/res/robbert_logo.png", "widget": [{"text": "Hallo, ik ben RobBERTje, een gedistilleerd <mask> taalmodel van de KU Leuven."}]} | DTAI-KULeuven/robbertje-1-gb-bort | null | [
"transformers",
"pytorch",
"roberta",
"fill-mask",
"Dutch",
"Flemish",
"RoBERTa",
"RobBERT",
"RobBERTje",
"nl",
"dataset:oscar",
"dataset:dbrd",
"dataset:lassy-ud",
"dataset:europarl-mono",
"dataset:conll2002",
"arxiv:2101.05716",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2101.05716"
] | [
"nl"
] | TAGS
#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
![](URL alt=)
About RobBERTje
===============
RobBERTje is a collection of distilled models based on RobBERT. There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch the repository for updates.
News
====
* February 21, 2022: Our paper about RobBERTje has been published in volume 11 of CLIN journal!
* July 2, 2021: Publicly released 4 RobBERTje models.
* May 12, 2021: RobBERTje was accepted at CLIN31 for an oral presentation!
The models
==========
Results
=======
Intrinsic results
-----------------
We calculated the *pseudo perplexity* (PPPL) from cite, which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
Extrinsic results
-----------------
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a Dutch NLI task named SICK-NL was also released and we evaluated our models with it as well.
| [] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
93
] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] |
fill-mask | transformers |
<p align="center">
<img src="https://github.com/iPieter/robbertje/raw/master/images/robbertje_logo_with_name.png" alt="RobBERTje: A collection of distilled Dutch BERT-based models" width="75%">
</p>
# About RobBERTje
RobBERTje is a collection of distilled models based on [RobBERT](http://github.com/iPieter/robbert). There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch [the repository](http://github.com/iPieter/robbertje) for updates.
# News
- **February 21, 2022**: Our paper about RobBERTje has been published in [volume 11 of CLIN journal](https://www.clinjournal.org/clinj/article/view/131)!
- **July 2, 2021**: Publicly released 4 RobBERTje models.
- **May 12, 2021**: RobBERTje was accepted at [CLIN31](https://www.clin31.ugent.be) for an oral presentation!
# The models
| Model | Description | Parameters | Training size | Huggingface id |
|--------------|-------------|------------------|-------------------|------------------------------------------------------------------------------------|
| Non-shuffled | Trained on the non-shuffled variant of the oscar corpus, without any operations to preserve this order during training and distillation. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-non-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-non-shuffled) |
| Shuffled | Trained on the publicly available and shuffled OSCAR corpus. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-shuffled) |
| Merged (p=0.5) | Same as the non-shuffled variant, but sequential sentences of the same document are merged with a probability of 50%. | 74 M | 1 GB | this model |
| BORT | A smaller version with 8 attention heads instead of 12 and 4 layers instead of 6 (and 12 for RobBERT). | 46 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-bort](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-bort) |
# Results
## Intrinsic results
We calculated the _pseudo perplexity_ (PPPL) from [cite](), which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
| Model | PPPL |
|-------------------|-----------|
| RobBERT (teacher) | 7.76 |
| Non-shuffled | 12.95 |
| Shuffled | 18.74 |
| Merged (p=0.5) | 17.10 |
| BORT | 26.44 |
## Extrinsic results
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a [Dutch NLI task named SICK-NL](https://arxiv.org/abs/2101.05716) was also released and we evaluated our models with it as well.
| Model | DBRD | DIE-DAT | NER | POS |SICK-NL |
|------------------|-----------|-----------|-----------|-----------|----------|
| RobBERT (teacher)|94.4 | 99.2 |89.1 |96.4 | 84.2 |
| Non-shuffled |90.2 | 98.4 |82.9 |95.5 | 83.4 |
| Shuffled |92.5 | 98.2 |82.7 |95.6 | 83.4 |
| Merged (p=0.5) |92.9 | 96.5 |81.8 |95.2 | 82.8 |
| BORT |89.6 | 92.2 |79.7 |94.3 | 81.0 |
| {"language": "nl", "license": "mit", "tags": ["Dutch", "Flemish", "RoBERTa", "RobBERT", "RobBERTje"], "datasets": ["oscar", "oscar (NL)", "dbrd", "lassy-ud", "europarl-mono", "conll2002"], "thumbnail": "https://github.com/iPieter/RobBERT/raw/master/res/robbert_logo.png", "widget": [{"text": "Hallo, ik ben RobBERTje, een gedistilleerd <mask> taalmodel van de KU Leuven."}]} | DTAI-KULeuven/robbertje-1-gb-merged | null | [
"transformers",
"pytorch",
"roberta",
"fill-mask",
"Dutch",
"Flemish",
"RoBERTa",
"RobBERT",
"RobBERTje",
"nl",
"arxiv:2101.05716",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2101.05716"
] | [
"nl"
] | TAGS
#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
![](URL alt=)
About RobBERTje
===============
RobBERTje is a collection of distilled models based on RobBERT. There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch the repository for updates.
News
====
* February 21, 2022: Our paper about RobBERTje has been published in volume 11 of CLIN journal!
* July 2, 2021: Publicly released 4 RobBERTje models.
* May 12, 2021: RobBERTje was accepted at CLIN31 for an oral presentation!
The models
==========
Results
=======
Intrinsic results
-----------------
We calculated the *pseudo perplexity* (PPPL) from cite, which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
Extrinsic results
-----------------
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a Dutch NLI task named SICK-NL was also released and we evaluated our models with it as well.
| [] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
58
] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] |
fill-mask | transformers |
<p align="center">
<img src="https://github.com/iPieter/robbertje/raw/master/images/robbertje_logo_with_name.png" alt="RobBERTje: A collection of distilled Dutch BERT-based models" width="75%">
</p>
# About RobBERTje
RobBERTje is a collection of distilled models based on [RobBERT](http://github.com/iPieter/robbert). There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch [the repository](http://github.com/iPieter/robbertje) for updates.
# News
- **February 21, 2022**: Our paper about RobBERTje has been published in [volume 11 of CLIN journal](https://www.clinjournal.org/clinj/article/view/131)!
- **July 2, 2021**: Publicly released 4 RobBERTje models.
- **May 12, 2021**: RobBERTje was accepted at [CLIN31](https://www.clin31.ugent.be) for an oral presentation!
# The models
| Model | Description | Parameters | Training size | Huggingface id |
|--------------|-------------|------------------|-------------------|------------------------------------------------------------------------------------|
| Non-shuffled | Trained on the non-shuffled variant of the oscar corpus, without any operations to preserve this order during training and distillation. | 74 M | 1 GB | this model |
| Shuffled | Trained on the publicly available and shuffled OSCAR corpus. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-shuffled) |
| Merged (p=0.5) | Same as the non-shuffled variant, but sequential sentences of the same document are merged with a probability of 50%. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-merged](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-merged) |
| BORT | A smaller version with 8 attention heads instead of 12 and 4 layers instead of 6 (and 12 for RobBERT). | 46 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-bort](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-bort) |
# Results
## Intrinsic results
We calculated the _pseudo perplexity_ (PPPL) from [cite](), which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
| Model | PPPL |
|-------------------|-----------|
| RobBERT (teacher) | 7.76 |
| Non-shuffled | 12.95 |
| Shuffled | 18.74 |
| Merged (p=0.5) | 17.10 |
| BORT | 26.44 |
## Extrinsic results
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a [Dutch NLI task named SICK-NL](https://arxiv.org/abs/2101.05716) was also released and we evaluated our models with it as well.
| Model | DBRD | DIE-DAT | NER | POS |SICK-NL |
|------------------|-----------|-----------|-----------|-----------|----------|
| RobBERT (teacher)|94.4 | 99.2 |89.1 |96.4 | 84.2 |
| Non-shuffled |90.2 | 98.4 |82.9 |95.5 | 83.4 |
| Shuffled |92.5 | 98.2 |82.7 |95.6 | 83.4 |
| Merged (p=0.5) |92.9 | 96.5 |81.8 |95.2 | 82.8 |
| BORT |89.6 | 92.2 |79.7 |94.3 | 81.0 |
| {"language": "nl", "license": "mit", "tags": ["Dutch", "Flemish", "RoBERTa", "RobBERT", "RobBERTje"], "datasets": ["oscar", "dbrd", "lassy-ud", "europarl-mono", "conll2002"], "thumbnail": "https://github.com/iPieter/robbertje/raw/master/images/robbertje_logo_with_name.png", "widget": [{"text": "Hallo, ik ben RobBERTje, een gedistilleerd <mask> taalmodel van de KU Leuven."}]} | DTAI-KULeuven/robbertje-1-gb-non-shuffled | null | [
"transformers",
"pytorch",
"roberta",
"fill-mask",
"Dutch",
"Flemish",
"RoBERTa",
"RobBERT",
"RobBERTje",
"nl",
"dataset:oscar",
"dataset:dbrd",
"dataset:lassy-ud",
"dataset:europarl-mono",
"dataset:conll2002",
"arxiv:2101.05716",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2101.05716"
] | [
"nl"
] | TAGS
#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
![](URL alt=)
About RobBERTje
===============
RobBERTje is a collection of distilled models based on RobBERT. There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch the repository for updates.
News
====
* February 21, 2022: Our paper about RobBERTje has been published in volume 11 of CLIN journal!
* July 2, 2021: Publicly released 4 RobBERTje models.
* May 12, 2021: RobBERTje was accepted at CLIN31 for an oral presentation!
The models
==========
Results
=======
Intrinsic results
-----------------
We calculated the *pseudo perplexity* (PPPL) from cite, which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
Extrinsic results
-----------------
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a Dutch NLI task named SICK-NL was also released and we evaluated our models with it as well.
| [] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
93
] | [
"TAGS\n#transformers #pytorch #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #dataset-oscar #dataset-dbrd #dataset-lassy-ud #dataset-europarl-mono #dataset-conll2002 #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] |
fill-mask | transformers |
<p align="center">
<img src="https://github.com/iPieter/robbertje/raw/master/images/robbertje_logo_with_name.png" alt="RobBERTje: A collection of distilled Dutch BERT-based models" width="75%">
</p>
# About RobBERTje
RobBERTje is a collection of distilled models based on [RobBERT](http://github.com/iPieter/robbert). There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch [the repository](http://github.com/iPieter/robbertje) for updates.
# News
- **February 21, 2022**: Our paper about RobBERTje has been published in [volume 11 of CLIN journal](https://www.clinjournal.org/clinj/article/view/131)!
- **July 2, 2021**: Publicly released 4 RobBERTje models.
- **May 12, 2021**: RobBERTje was accepted at [CLIN31](https://www.clin31.ugent.be) for an oral presentation!
# The models
| Model | Description | Parameters | Training size | Huggingface id |
|--------------|-------------|------------------|-------------------|------------------------------------------------------------------------------------|
| Non-shuffled | Trained on the non-shuffled variant of the oscar corpus, without any operations to preserve this order during training and distillation. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-non-shuffled](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-non-shuffled) |
| Shuffled | Trained on the publicly available and shuffled OSCAR corpus. | 74 M | 1 GB | this model |
| Merged (p=0.5) | Same as the non-shuffled variant, but sequential sentences of the same document are merged with a probability of 50%. | 74 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-merged](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-merged) |
| BORT | A smaller version with 8 attention heads instead of 12 and 4 layers instead of 6 (and 12 for RobBERT). | 46 M | 1 GB | [DTAI-KULeuven/robbertje-1-gb-bort](https://huggingface.co/DTAI-KULeuven/robbertje-1-gb-bort) |
# Results
## Intrinsic results
We calculated the _pseudo perplexity_ (PPPL) from [cite](), which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
| Model | PPPL |
|-------------------|-----------|
| RobBERT (teacher) | 7.76 |
| Non-shuffled | 12.95 |
| Shuffled | 18.74 |
| Merged (p=0.5) | 17.10 |
| BORT | 26.44 |
## Extrinsic results
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a [Dutch NLI task named SICK-NL](https://arxiv.org/abs/2101.05716) was also released and we evaluated our models with it as well.
| Model | DBRD | DIE-DAT | NER | POS |SICK-NL |
|------------------|-----------|-----------|-----------|-----------|----------|
| RobBERT (teacher)|94.4 | 99.2 |89.1 |96.4 | 84.2 |
| Non-shuffled |90.2 | 98.4 |82.9 |95.5 | 83.4 |
| Shuffled |92.5 | 98.2 |82.7 |95.6 | 83.4 |
| Merged (p=0.5) |92.9 | 96.5 |81.8 |95.2 | 82.8 |
| BORT |89.6 | 92.2 |79.7 |94.3 | 81.0 |
| {"language": "nl", "license": "mit", "tags": ["Dutch", "Flemish", "RoBERTa", "RobBERT", "RobBERTje"], "datasets": ["oscar", "oscar (NL)", "dbrd", "lassy-ud", "europarl-mono", "conll2002"], "thumbnail": "https://github.com/iPieter/RobBERT/raw/master/res/robbert_logo.png", "widget": [{"text": "Hallo, ik ben RobBERTje, een gedistilleerd <mask> taalmodel van de KU Leuven."}]} | DTAI-KULeuven/robbertje-1-gb-shuffled | null | [
"transformers",
"pytorch",
"safetensors",
"roberta",
"fill-mask",
"Dutch",
"Flemish",
"RoBERTa",
"RobBERT",
"RobBERTje",
"nl",
"arxiv:2101.05716",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2101.05716"
] | [
"nl"
] | TAGS
#transformers #pytorch #safetensors #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
![](URL alt=)
About RobBERTje
===============
RobBERTje is a collection of distilled models based on RobBERT. There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch the repository for updates.
News
====
* February 21, 2022: Our paper about RobBERTje has been published in volume 11 of CLIN journal!
* July 2, 2021: Publicly released 4 RobBERTje models.
* May 12, 2021: RobBERTje was accepted at CLIN31 for an oral presentation!
The models
==========
Results
=======
Intrinsic results
-----------------
We calculated the *pseudo perplexity* (PPPL) from cite, which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
Extrinsic results
-----------------
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a Dutch NLI task named SICK-NL was also released and we evaluated our models with it as well.
| [] | [
"TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
62
] | [
"TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #Dutch #Flemish #RoBERTa #RobBERT #RobBERTje #nl #arxiv-2101.05716 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] |
text-classification | transformers |
# Danish BERT for emotion detection
The BERT Emotion model detects whether a Danish text is emotional or not.
It is based on the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO which has been fine-tuned on social media data.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/sentiment_analysis.html#bert-emotion) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-binary-emotion-classification-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-binary-emotion-classification-base")
```
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Der er et tr\u00e6 i haven."}]} | alexandrainst/da-binary-emotion-classification-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Danish BERT for emotion detection
The BERT Emotion model detects whether a Danish text is emotional or not.
It is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | [
"# Danish BERT for emotion detection\n\nThe BERT Emotion model detects whether a Danish text is emotional or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Danish BERT for emotion detection\n\nThe BERT Emotion model detects whether a Danish text is emotional or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
45,
64,
33
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Danish BERT for emotion detection\n\nThe BERT Emotion model detects whether a Danish text is emotional or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] |
text-classification | transformers |
# Danish BERT for emotion classification
The BERT Emotion model classifies a Danish text in one of the following class:
* Glæde/Sindsro
* Tillid/Accept
* Forventning/Interrese
* Overasket/Målløs
* Vrede/Irritation
* Foragt/Modvilje
* Sorg/trist
* Frygt/Bekymret
It is based on the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO which has been fine-tuned on social media data.
This model should be used after detecting whether the text contains emotion or not, using the binary [BERT Emotion model](https://huggingface.co/alexandrainst/da-binary-emotion-classification-base).
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/sentiment_analysis.html#bert-emotion) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-emotion-classification-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-emotion-classification-base")
```
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Jeg ejer en r\u00f8d bil og det er en god bil."}]} | alexandrainst/da-emotion-classification-base | null | [
"transformers",
"pytorch",
"tf",
"bert",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Danish BERT for emotion classification
The BERT Emotion model classifies a Danish text in one of the following class:
* Glæde/Sindsro
* Tillid/Accept
* Forventning/Interrese
* Overasket/Målløs
* Vrede/Irritation
* Foragt/Modvilje
* Sorg/trist
* Frygt/Bekymret
It is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data.
This model should be used after detecting whether the text contains emotion or not, using the binary BERT Emotion model.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | [
"# Danish BERT for emotion classification\n\nThe BERT Emotion model classifies a Danish text in one of the following class:\n* Glæde/Sindsro\n* Tillid/Accept\n* Forventning/Interrese\n* Overasket/Målløs\n* Vrede/Irritation\n* Foragt/Modvilje\n* Sorg/trist\n* Frygt/Bekymret\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nThis model should be used after detecting whether the text contains emotion or not, using the binary BERT Emotion model.\n\nSee the DaNLP documentation for more details. \n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
"TAGS\n#transformers #pytorch #tf #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Danish BERT for emotion classification\n\nThe BERT Emotion model classifies a Danish text in one of the following class:\n* Glæde/Sindsro\n* Tillid/Accept\n* Forventning/Interrese\n* Overasket/Målløs\n* Vrede/Irritation\n* Foragt/Modvilje\n* Sorg/trist\n* Frygt/Bekymret\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nThis model should be used after detecting whether the text contains emotion or not, using the binary BERT Emotion model.\n\nSee the DaNLP documentation for more details. \n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
45,
144,
33
] | [
"TAGS\n#transformers #pytorch #tf #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Danish BERT for emotion classification\n\nThe BERT Emotion model classifies a Danish text in one of the following class:\n* Glæde/Sindsro\n* Tillid/Accept\n* Forventning/Interrese\n* Overasket/Målløs\n* Vrede/Irritation\n* Foragt/Modvilje\n* Sorg/trist\n* Frygt/Bekymret\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nThis model should be used after detecting whether the text contains emotion or not, using the binary BERT Emotion model.\n\nSee the DaNLP documentation for more details. \n\nHere is how to use the model:## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] |
text-classification | transformers |
# Danish BERT for hate speech classification
The BERT HateSpeech model classifies offensive Danish text into 4 categories:
* `Særlig opmærksomhed` (special attention, e.g. threat)
* `Personangreb` (personal attack)
* `Sprogbrug` (offensive language)
* `Spam & indhold` (spam)
This model is intended to be used after the [BERT HateSpeech detection model](https://huggingface.co/alexandrainst/da-hatespeech-detection-base).
It is based on the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO which has been fine-tuned on social media data.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/hatespeech.html#bertdr) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-hatespeech-classification-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-hatespeech-classification-base")
```
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Senile gamle idiot"}]} | alexandrainst/da-hatespeech-classification-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Danish BERT for hate speech classification
The BERT HateSpeech model classifies offensive Danish text into 4 categories:
* 'Særlig opmærksomhed' (special attention, e.g. threat)
* 'Personangreb' (personal attack)
* 'Sprogbrug' (offensive language)
* 'Spam & indhold' (spam)
This model is intended to be used after the BERT HateSpeech detection model.
It is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | [
"# Danish BERT for hate speech classification\n\nThe BERT HateSpeech model classifies offensive Danish text into 4 categories: \n * 'Særlig opmærksomhed' (special attention, e.g. threat)\n * 'Personangreb' (personal attack) \n * 'Sprogbrug' (offensive language)\n * 'Spam & indhold' (spam)\nThis model is intended to be used after the BERT HateSpeech detection model.\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Danish BERT for hate speech classification\n\nThe BERT HateSpeech model classifies offensive Danish text into 4 categories: \n * 'Særlig opmærksomhed' (special attention, e.g. threat)\n * 'Personangreb' (personal attack) \n * 'Sprogbrug' (offensive language)\n * 'Spam & indhold' (spam)\nThis model is intended to be used after the BERT HateSpeech detection model.\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
45,
138,
33
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Danish BERT for hate speech classification\n\nThe BERT HateSpeech model classifies offensive Danish text into 4 categories: \n * 'Særlig opmærksomhed' (special attention, e.g. threat)\n * 'Personangreb' (personal attack) \n * 'Sprogbrug' (offensive language)\n * 'Spam & indhold' (spam)\nThis model is intended to be used after the BERT HateSpeech detection model.\n\nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] |
text-classification | transformers |
# Danish BERT for hate speech (offensive language) detection
The BERT HateSpeech model detects whether a Danish text is offensive or not.
It is based on the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO which has been fine-tuned on social media data.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/hatespeech.html#bertdr) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-hatespeech-detection-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-hatespeech-detection-base")
```
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Senile gamle idiot"}]} | alexandrainst/da-hatespeech-detection-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Danish BERT for hate speech (offensive language) detection
The BERT HateSpeech model detects whether a Danish text is offensive or not.
It is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | [
"# Danish BERT for hate speech (offensive language) detection\n\nThe BERT HateSpeech model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Danish BERT for hate speech (offensive language) detection\n\nThe BERT HateSpeech model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
45,
71,
33
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Danish BERT for hate speech (offensive language) detection\n\nThe BERT HateSpeech model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish BERT model by BotXO which has been fine-tuned on social media data. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] |
token-classification | transformers |
# BERT fine-tuned for Named Entity Recognition in Danish
The model tags tokens (in Danish sentences) with named entity tags (BIO format) [PER, ORG, LOC, MISC].
The pretrained language model used for fine-tuning is the [Danish BERT](https://github.com/certainlyio/nordic_bert) by BotXO.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/ner.html#bert) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForTokenClassification
model = BertForTokenClassification.from_pretrained("alexandrainst/da-ner-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-ner-base")
```
## Training Data
The model has been trained on the [DaNE](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#dane). | {"language": ["da"], "license": "apache-2.0", "datasets": ["dane"], "widget": [{"text": "Jens Peter Hansen kommer fra Danmark"}]} | alexandrainst/da-ner-base | null | [
"transformers",
"pytorch",
"tf",
"bert",
"token-classification",
"da",
"dataset:dane",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #bert #token-classification #da #dataset-dane #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# BERT fine-tuned for Named Entity Recognition in Danish
The model tags tokens (in Danish sentences) with named entity tags (BIO format) [PER, ORG, LOC, MISC].
The pretrained language model used for fine-tuning is the Danish BERT by BotXO.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training Data
The model has been trained on the DaNE. | [
"# BERT fine-tuned for Named Entity Recognition in Danish\n\nThe model tags tokens (in Danish sentences) with named entity tags (BIO format) [PER, ORG, LOC, MISC].\nThe pretrained language model used for fine-tuning is the Danish BERT by BotXO. \n\nSee the DaNLP documentation for more details.\n\nHere is how to use the model:",
"## Training Data \n\nThe model has been trained on the DaNE."
] | [
"TAGS\n#transformers #pytorch #tf #bert #token-classification #da #dataset-dane #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# BERT fine-tuned for Named Entity Recognition in Danish\n\nThe model tags tokens (in Danish sentences) with named entity tags (BIO format) [PER, ORG, LOC, MISC].\nThe pretrained language model used for fine-tuning is the Danish BERT by BotXO. \n\nSee the DaNLP documentation for more details.\n\nHere is how to use the model:",
"## Training Data \n\nThe model has been trained on the DaNE."
] | [
46,
78,
13
] | [
"TAGS\n#transformers #pytorch #tf #bert #token-classification #da #dataset-dane #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# BERT fine-tuned for Named Entity Recognition in Danish\n\nThe model tags tokens (in Danish sentences) with named entity tags (BIO format) [PER, ORG, LOC, MISC].\nThe pretrained language model used for fine-tuning is the Danish BERT by BotXO. \n\nSee the DaNLP documentation for more details.\n\nHere is how to use the model:## Training Data \n\nThe model has been trained on the DaNE."
] |
text-classification | transformers |
# Model Card for Danish BERT
Danish BERT Tone for sentiment polarity detection
# Model Details
## Model Description
The BERT Tone model detects sentiment polarity (positive, neutral or negative) in Danish texts. It has been finetuned on the pretrained Danish BERT model by BotXO.
- **Developed by:** DaNLP
- **Shared by [Optional]:** Hugging Face
- **Model type:** Text Classification
- **Language(s) (NLP):** Danish (da)
- **License:** cc-by-sa-4.0
- **Related Models:** More information needed
- **Parent Model:** BERT
- **Resources for more information:**
- [GitHub Repo](https://github.com/certainlyio/nordic_bert)
- [Associated Documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/sentiment_analysis.html#bert-tone)
# Uses
## Direct Use
This model can be used for text classification
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The data used for training come from the [Twitter Sentiment](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#twitsent) and [EuroParl sentiment 2](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#europarl-sentiment2) datasets.
## Training Procedure
### Preprocessing
It has been finetuned on the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO.
### Speeds, Sizes, Times
More information needed.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed.
### Factors
### Metrics
F1
## Results
More information needed.
# Model Examination
More information needed.
# Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** More information needed.
- **Hours used:** More information needed.
- **Cloud Provider:** More information needed.
- **Compute Region:** More information needed.
- **Carbon Emitted:** More information needed.
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed.
## Compute Infrastructure
More information needed.
### Hardware
More information needed.
### Software
More information needed.
# Citation
**BibTeX:**
More information needed.
**APA:**
More information needed.
# Glossary [optional]
More information needed.
# More Information [optional]
More information needed.
# Model Card Authors [optional]
DaNLP in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed.
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-sentiment-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-sentiment-base")
```
</details> | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Det er super godt"}]} | alexandrainst/da-sentiment-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"text-classification",
"da",
"arxiv:1910.09700",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1910.09700"
] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #bert #text-classification #da #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Danish BERT
Danish BERT Tone for sentiment polarity detection
# Model Details
## Model Description
The BERT Tone model detects sentiment polarity (positive, neutral or negative) in Danish texts. It has been finetuned on the pretrained Danish BERT model by BotXO.
- Developed by: DaNLP
- Shared by [Optional]: Hugging Face
- Model type: Text Classification
- Language(s) (NLP): Danish (da)
- License: cc-by-sa-4.0
- Related Models: More information needed
- Parent Model: BERT
- Resources for more information:
- GitHub Repo
- Associated Documentation
# Uses
## Direct Use
This model can be used for text classification
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets.
## Training Procedure
### Preprocessing
It has been finetuned on the pretrained Danish BERT model by BotXO.
### Speeds, Sizes, Times
More information needed.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed.
### Factors
### Metrics
F1
## Results
More information needed.
# Model Examination
More information needed.
# Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: More information needed.
- Hours used: More information needed.
- Cloud Provider: More information needed.
- Compute Region: More information needed.
- Carbon Emitted: More information needed.
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed.
## Compute Infrastructure
More information needed.
### Hardware
More information needed.
### Software
More information needed.
BibTeX:
More information needed.
APA:
More information needed.
# Glossary [optional]
More information needed.
# More Information [optional]
More information needed.
# Model Card Authors [optional]
DaNLP in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed.
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
</details> | [
"# Model Card for Danish BERT\n Danish BERT Tone for sentiment polarity detection",
"# Model Details",
"## Model Description\n \nThe BERT Tone model detects sentiment polarity (positive, neutral or negative) in Danish texts. It has been finetuned on the pretrained Danish BERT model by BotXO.\n \n- Developed by: DaNLP\n- Shared by [Optional]: Hugging Face\n- Model type: Text Classification\n- Language(s) (NLP): Danish (da)\n- License: cc-by-sa-4.0\n- Related Models: More information needed\n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo \n - Associated Documentation",
"# Uses",
"## Direct Use\n \nThis model can be used for text classification",
"## Downstream Use [Optional]\n \n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets.",
"## Training Procedure",
"### Preprocessing\n \nIt has been finetuned on the pretrained Danish BERT model by BotXO.",
"### Speeds, Sizes, Times\nMore information needed.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed.",
"### Factors",
"### Metrics\n \nF1",
"## Results \n \nMore information needed.",
"# Model Examination\n \nMore information needed.",
"# Environmental Impact\n \n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed.\n- Hours used: More information needed.\n- Cloud Provider: More information needed.\n- Compute Region: More information needed.\n- Carbon Emitted: More information needed.",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed.",
"## Compute Infrastructure\n \nMore information needed.",
"### Hardware\n \nMore information needed.",
"### Software\n \nMore information needed.\n \nBibTeX:\n \nMore information needed.\n \nAPA:\n \nMore information needed.",
"# Glossary [optional]\n \nMore information needed.",
"# More Information [optional]\n \nMore information needed.",
"# Model Card Authors [optional]\n \nDaNLP in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed.",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Danish BERT\n Danish BERT Tone for sentiment polarity detection",
"# Model Details",
"## Model Description\n \nThe BERT Tone model detects sentiment polarity (positive, neutral or negative) in Danish texts. It has been finetuned on the pretrained Danish BERT model by BotXO.\n \n- Developed by: DaNLP\n- Shared by [Optional]: Hugging Face\n- Model type: Text Classification\n- Language(s) (NLP): Danish (da)\n- License: cc-by-sa-4.0\n- Related Models: More information needed\n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo \n - Associated Documentation",
"# Uses",
"## Direct Use\n \nThis model can be used for text classification",
"## Downstream Use [Optional]\n \n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets.",
"## Training Procedure",
"### Preprocessing\n \nIt has been finetuned on the pretrained Danish BERT model by BotXO.",
"### Speeds, Sizes, Times\nMore information needed.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed.",
"### Factors",
"### Metrics\n \nF1",
"## Results \n \nMore information needed.",
"# Model Examination\n \nMore information needed.",
"# Environmental Impact\n \n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed.\n- Hours used: More information needed.\n- Cloud Provider: More information needed.\n- Compute Region: More information needed.\n- Carbon Emitted: More information needed.",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed.",
"## Compute Infrastructure\n \nMore information needed.",
"### Hardware\n \nMore information needed.",
"### Software\n \nMore information needed.\n \nBibTeX:\n \nMore information needed.\n \nAPA:\n \nMore information needed.",
"# Glossary [optional]\n \nMore information needed.",
"# More Information [optional]\n \nMore information needed.",
"# Model Card Authors [optional]\n \nDaNLP in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed.",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] | [
55,
14,
3,
117,
2,
12,
11,
25,
70,
33,
3,
23,
4,
26,
12,
2,
9,
9,
4,
6,
7,
7,
68,
6,
10,
8,
8,
23,
10,
10,
22,
8,
36
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Danish BERT\n Danish BERT Tone for sentiment polarity detection# Model Details## Model Description\n \nThe BERT Tone model detects sentiment polarity (positive, neutral or negative) in Danish texts. It has been finetuned on the pretrained Danish BERT model by BotXO.\n \n- Developed by: DaNLP\n- Shared by [Optional]: Hugging Face\n- Model type: Text Classification\n- Language(s) (NLP): Danish (da)\n- License: cc-by-sa-4.0\n- Related Models: More information needed\n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo \n - Associated Documentation# Uses## Direct Use\n \nThis model can be used for text classification## Downstream Use [Optional]\n \n \nMore information needed.## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.# Training Details## Training Data\n \nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets.## Training Procedure### Preprocessing\n \nIt has been finetuned on the pretrained Danish BERT model by BotXO.### Speeds, Sizes, Times\nMore information needed.# Evaluation## Testing Data, Factors & Metrics### Testing Data\n \nMore information needed.### Factors### Metrics\n \nF1## Results \n \nMore information needed.# Model Examination\n \nMore information needed.# Environmental Impact\n \n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed.\n- Hours used: More information needed.\n- Cloud Provider: More information needed.\n- Compute Region: More information needed.\n- Carbon Emitted: More information needed.# Technical Specifications [optional]## Model Architecture and Objective\n \nMore information needed.## Compute Infrastructure\n \nMore information needed.### Hardware\n \nMore information needed.### Software\n \nMore information needed.\n \nBibTeX:\n \nMore information needed.\n \nAPA:\n \nMore information needed.# Glossary [optional]\n \nMore information needed.# More Information [optional]\n \nMore information needed.# Model Card Authors [optional]\n \nDaNLP in collaboration with Ezi Ozoani and the Hugging Face team# Model Card Contact\n \nMore information needed.# How to Get Started with the Model\n \nUse the code below to get started with the model.\n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
text-classification | transformers |
# Danish BERT Tone for the detection of subjectivity/objectivity
The BERT Tone model detects whether a text (in Danish) is subjective or objective.
The model is based on the finetuning of the pretrained [Danish BERT](https://github.com/certainlyio/nordic_bert) model by BotXO.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/sentiment_analysis.html#bert-tone) for more details.
Here is how to use the model:
```python
from transformers import BertTokenizer, BertForSequenceClassification
model = BertForSequenceClassification.from_pretrained("alexandrainst/da-subjectivivity-classification-base")
tokenizer = BertTokenizer.from_pretrained("alexandrainst/da-subjectivivity-classification-base")
```
## Training data
The data used for training come from the [Twitter Sentiment](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#twitsent) and [EuroParl sentiment 2](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#europarl-sentiment2) datasets. | {"language": ["da"], "license": "apache-2.0", "datasets": ["DDSC/twitter-sent", "DDSC/europarl"], "widget": [{"text": "Jeg tror alligvel, det bliver godt"}]} | alexandrainst/da-subjectivivity-classification-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"text-classification",
"da",
"dataset:DDSC/twitter-sent",
"dataset:DDSC/europarl",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #bert #text-classification #da #dataset-DDSC/twitter-sent #dataset-DDSC/europarl #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Danish BERT Tone for the detection of subjectivity/objectivity
The BERT Tone model detects whether a text (in Danish) is subjective or objective.
The model is based on the finetuning of the pretrained Danish BERT model by BotXO.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets. | [
"# Danish BERT Tone for the detection of subjectivity/objectivity\n\nThe BERT Tone model detects whether a text (in Danish) is subjective or objective. \nThe model is based on the finetuning of the pretrained Danish BERT model by BotXO. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #dataset-DDSC/twitter-sent #dataset-DDSC/europarl #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Danish BERT Tone for the detection of subjectivity/objectivity\n\nThe BERT Tone model detects whether a text (in Danish) is subjective or objective. \nThe model is based on the finetuning of the pretrained Danish BERT model by BotXO. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets."
] | [
64,
70,
23
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #text-classification #da #dataset-DDSC/twitter-sent #dataset-DDSC/europarl #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Danish BERT Tone for the detection of subjectivity/objectivity\n\nThe BERT Tone model detects whether a text (in Danish) is subjective or objective. \nThe model is based on the finetuning of the pretrained Danish BERT model by BotXO. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:## Training data\n\nThe data used for training come from the Twitter Sentiment and EuroParl sentiment 2 datasets."
] |
text-classification | transformers |
# Danish ELECTRA for hate speech (offensive language) detection
The ELECTRA Offensive model detects whether a Danish text is offensive or not.
It is based on the pretrained [Danish Ælæctra](Maltehb/aelaectra-danish-electra-small-cased) model.
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/hatespeech.html#electra) for more details.
Here is how to use the model:
```python
from transformers import ElectraTokenizer, ElectraForSequenceClassification
model = ElectraForSequenceClassification.from_pretrained("alexandrainst/da-hatespeech-detection-small")
tokenizer = ElectraTokenizer.from_pretrained("alexandrainst/da-hatespeech-detection-small")
```
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | {"language": ["da"], "license": "apache-2.0", "widget": [{"text": "Senile gamle idiot"}]} | alexandrainst/da-hatespeech-detection-small | null | [
"transformers",
"pytorch",
"electra",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #electra #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Danish ELECTRA for hate speech (offensive language) detection
The ELECTRA Offensive model detects whether a Danish text is offensive or not.
It is based on the pretrained Danish Ælæctra model.
See the DaNLP documentation for more details.
Here is how to use the model:
## Training data
The data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio. | [
"# Danish ELECTRA for hate speech (offensive language) detection\n\nThe ELECTRA Offensive model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish Ælæctra model. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
"TAGS\n#transformers #pytorch #electra #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Danish ELECTRA for hate speech (offensive language) detection\n\nThe ELECTRA Offensive model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish Ælæctra model. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:",
"## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] | [
39,
61,
33
] | [
"TAGS\n#transformers #pytorch #electra #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Danish ELECTRA for hate speech (offensive language) detection\n\nThe ELECTRA Offensive model detects whether a Danish text is offensive or not. \nIt is based on the pretrained Danish Ælæctra model. \n\nSee the DaNLP documentation for more details. \n\n\nHere is how to use the model:## Training data\n\nThe data used for training has not been made publicly available. It consists of social media data manually annotated in collaboration with Danmarks Radio."
] |
text-classification | transformers |
# XLM-Roberta fine-tuned for Named Entity Disambiguation
Given a sentence and a knowledge graph context, the model detects whether a specific entity (represented by the knowledge graph context) is mentioned in the sentence (binary classification).
The base language model used is the [xlm-roberta-base](https://huggingface.co/xlm-roberta-base).
Here is how to use the model:
```python
from transformers import XLMRobertaTokenizer, XLMRobertaForSequenceClassification
model = XLMRobertaForSequenceClassification.from_pretrained("alexandrainst/da-ned-base")
tokenizer = XLMRobertaTokenizer.from_pretrained("alexandrainst/da-ned-base")
```
The tokenizer takes 2 strings has input: the sentence and the knowledge graph (KG) context.
Here is an example:
```python
sentence = "Karen Blixen vendte tilbage til Danmark, hvor hun boede resten af sit liv på Rungstedlund, som hun arvede efter sin mor i 1939"
kg_context = "udmærkelser modtaget Kritikerprisen udmærkelser modtaget Tagea Brandts Rejselegat udmærkelser modtaget Ingenio et arti udmærkelser modtaget Holbergmedaljen udmærkelser modtaget De Gyldne Laurbær mor Ingeborg Dinesen ægtefælle Bror von Blixen-Finecke køn kvinde Commons-kategori Karen Blixen LCAuth no95003722 VIAF 90663542 VIAF 121643918 GND-identifikator 118637878 ISNI 0000 0001 2096 6265 ISNI 0000 0003 6863 4408 ISNI 0000 0001 1891 0457 fødested Rungstedlund fødested Rungsted dødssted Rungstedlund dødssted København statsborgerskab Danmark NDL-nummer 00433530 dødsdato +1962-09-07T00:00:00Z dødsdato +1962-01-01T00:00:00Z fødselsdato +1885-04-17T00:00:00Z fødselsdato +1885-01-01T00:00:00Z AUT NKC jn20000600905 AUT NKC jo2015880827 AUT NKC xx0196181 emnets hovedkategori Kategori:Karen Blixen tilfælde af menneske billede Karen Blixen cropped from larger original.jpg IMDb-identifikationsnummer nm0227598 Freebase-ID /m/04ymd8w BNF 118857710 beskæftigelse skribent beskæftigelse selvbiograf beskæftigelse novelleforfatter ..."
```
A KG context, for a specific entity, can be generated from its Wikidata page.
In the previous example, the KG context is a string representation of the Wikidata page of [Karen Blixen (QID=Q182804)](https://www.wikidata.org/wiki/Q182804).
See the [DaNLP documentation](https://danlp-alexandra.readthedocs.io/en/latest/docs/tasks/ned.html#xlmr) for more details about how to generate a KG context.
## Training Data
The model has been trained on the [DaNED](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#daned) and [DaWikiNED](https://danlp-alexandra.readthedocs.io/en/latest/docs/datasets.html#dawikined) datasets. | {"language": ["da"], "license": "apache-2.0"} | alexandrainst/da-ned-base | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"xlm-roberta",
"text-classification",
"da",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"da"
] | TAGS
#transformers #pytorch #tf #safetensors #xlm-roberta #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# XLM-Roberta fine-tuned for Named Entity Disambiguation
Given a sentence and a knowledge graph context, the model detects whether a specific entity (represented by the knowledge graph context) is mentioned in the sentence (binary classification).
The base language model used is the xlm-roberta-base.
Here is how to use the model:
The tokenizer takes 2 strings has input: the sentence and the knowledge graph (KG) context.
Here is an example:
A KG context, for a specific entity, can be generated from its Wikidata page.
In the previous example, the KG context is a string representation of the Wikidata page of Karen Blixen (QID=Q182804).
See the DaNLP documentation for more details about how to generate a KG context.
## Training Data
The model has been trained on the DaNED and DaWikiNED datasets. | [
"# XLM-Roberta fine-tuned for Named Entity Disambiguation\n\nGiven a sentence and a knowledge graph context, the model detects whether a specific entity (represented by the knowledge graph context) is mentioned in the sentence (binary classification). \nThe base language model used is the xlm-roberta-base. \n\nHere is how to use the model: \n\n\n\nThe tokenizer takes 2 strings has input: the sentence and the knowledge graph (KG) context. \nHere is an example:\n\n\nA KG context, for a specific entity, can be generated from its Wikidata page. \nIn the previous example, the KG context is a string representation of the Wikidata page of Karen Blixen (QID=Q182804). \nSee the DaNLP documentation for more details about how to generate a KG context.",
"## Training Data \n\nThe model has been trained on the DaNED and DaWikiNED datasets."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# XLM-Roberta fine-tuned for Named Entity Disambiguation\n\nGiven a sentence and a knowledge graph context, the model detects whether a specific entity (represented by the knowledge graph context) is mentioned in the sentence (binary classification). \nThe base language model used is the xlm-roberta-base. \n\nHere is how to use the model: \n\n\n\nThe tokenizer takes 2 strings has input: the sentence and the knowledge graph (KG) context. \nHere is an example:\n\n\nA KG context, for a specific entity, can be generated from its Wikidata page. \nIn the previous example, the KG context is a string representation of the Wikidata page of Karen Blixen (QID=Q182804). \nSee the DaNLP documentation for more details about how to generate a KG context.",
"## Training Data \n\nThe model has been trained on the DaNED and DaWikiNED datasets."
] | [
48,
170,
22
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #text-classification #da #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# XLM-Roberta fine-tuned for Named Entity Disambiguation\n\nGiven a sentence and a knowledge graph context, the model detects whether a specific entity (represented by the knowledge graph context) is mentioned in the sentence (binary classification). \nThe base language model used is the xlm-roberta-base. \n\nHere is how to use the model: \n\n\n\nThe tokenizer takes 2 strings has input: the sentence and the knowledge graph (KG) context. \nHere is an example:\n\n\nA KG context, for a specific entity, can be generated from its Wikidata page. \nIn the previous example, the KG context is a string representation of the Wikidata page of Karen Blixen (QID=Q182804). \nSee the DaNLP documentation for more details about how to generate a KG context.## Training Data \n\nThe model has been trained on the DaNED and DaWikiNED datasets."
] |
text-generation | transformers |
#Saitama DialoGPT model | {"tags": ["conversational"]} | Daivakai/DialoGPT-small-saitama | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
#Saitama DialoGPT model | [] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
39
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
text-classification | transformers |
# scientific-challenges-and-directions
We present a novel resource to help scientists and medical professionals discover challenges and potential directions across scientific literature, focusing on a broad corpus pertaining to the COVID-19 pandemic and related historical research. At a high level, the _challenges_ and _directions_ are defined as follows:
* **Challenge**: A sentence mentioning a problem, difficulty, flaw, limitation, failure, lack of clarity, or knowledge gap.
* **Research direction**: A sentence mentioning suggestions or needs for further research, hypotheses, speculations, indications or hints that an issue is worthy of exploration.
* This model here is described in our paper: [A Search Engine for Discovery of Scientific Challenges and Directions](https://arxiv.org/abs/2108.13751) (though we've upgraded the infrastructure since the paper was released - there are slight differences in the results).
* Our dataset can be found [here](https://huggingface.co/datasets/DanL/scientific-challenges-and-directions-dataset).
* Please cite our paper if you use our datasets or models in your project. See the [BibTeX](#citation).
* Feel free to [email us](#contact-us).
* Also, check out [our search engine](https://challenges.apps.allenai.org/), as an example application.
## Model description
This model is a fine-tuned version of [PubMedBERT](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext) on the [scientific-challenges-and-directions-dataset](https://huggingface.co/datasets/DanL/scientific-challenges-and-directions-dataset), designed for multi-label text classification.
## Training and evaluation data
The scientific-challenges-and-directions model is trained based on a dataset that is a collection of 2894 sentences and their surrounding contexts, from 1786 full-text papers in the CORD-19 corpus, labeled for classification of challenges and directions by expert annotators with biomedical and bioNLP backgrounds. For full details on the train/test/split of the data see section 3.1 in our [paper](https://arxiv.org/abs/2108.13751)
## Example notebook
We include an example notebook that uses the model for inference in our [repo](https://github.com/Dan-La/scientific-challenges-and-directions). See `Inference_Notebook.ipynb`.
A training notebook is also included.
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning rate: 2e-05
- train batch size: 8
- eval batch size: 4
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr scheduler type: linear
- lr scheduler warmup steps: 500
- num epochs: 30
### Training results
The achieves the following results on the test set:
- Precision Challenge: 0.768719
- Recall Challenge: 0.780405
- F1 Challenge: 0.774518
- Precision Direction: 0.758112
- Recall Direction: 0.774096
- F1 Direction: 0.766021
- Precision (micro avg. on both labels): 0.764894
- Recall (micro avg. on both labels): 0.778139
- F1 (micro avg. on both labels): 0.771459
### Framework versions
- Transformers 4.15.0
- Pytorch 1.10.0+cu111
- Datasets 1.17.0
- Tokenizers 0.10.3
## Citation
If using our dataset and models, please cite:
```
@misc{lahav2021search,
title={A Search Engine for Discovery of Scientific Challenges and Directions},
author={Dan Lahav and Jon Saad Falcon and Bailey Kuehl and Sophie Johnson and Sravanthi Parasa and Noam Shomron and Duen Horng Chau and Diyi Yang and Eric Horvitz and Daniel S. Weld and Tom Hope},
year={2021},
eprint={2108.13751},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Contact us
Please don't hesitate to reach out.
**Email:** `lahav@mail.tau.ac.il`,`tomh@allenai.org`.
| {"language": ["en"], "tags": ["generated_from_trainer", "text-classification"], "datasets": ["DanL/scientific-challenges-and-directions-dataset"], "metrics": ["precision", "recall", "f1"], "widget": [{"text": "severe atypical cases of pneumonia emerged and quickly spread worldwide.", "example_title": "challenge"}, {"text": "we speculate that studying IL-6 will be beneficial.", "example_title": "direction"}, {"text": "in future studies, both PRRs should be tested as the cause for multiple deaths.", "example_title": "both"}, {"text": "IbMADS1-transformed potatoes exhibited tuber morphogenesis in the fibrous roots.", "example_title": "neither"}], "model-index": [{"name": "scientific-challenges-and-directions", "results": []}]} | DanL/scientific-challenges-and-directions | null | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"en",
"dataset:DanL/scientific-challenges-and-directions-dataset",
"arxiv:2108.13751",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2108.13751"
] | [
"en"
] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #en #dataset-DanL/scientific-challenges-and-directions-dataset #arxiv-2108.13751 #autotrain_compatible #endpoints_compatible #region-us
|
# scientific-challenges-and-directions
We present a novel resource to help scientists and medical professionals discover challenges and potential directions across scientific literature, focusing on a broad corpus pertaining to the COVID-19 pandemic and related historical research. At a high level, the _challenges_ and _directions_ are defined as follows:
* Challenge: A sentence mentioning a problem, difficulty, flaw, limitation, failure, lack of clarity, or knowledge gap.
* Research direction: A sentence mentioning suggestions or needs for further research, hypotheses, speculations, indications or hints that an issue is worthy of exploration.
* This model here is described in our paper: A Search Engine for Discovery of Scientific Challenges and Directions (though we've upgraded the infrastructure since the paper was released - there are slight differences in the results).
* Our dataset can be found here.
* Please cite our paper if you use our datasets or models in your project. See the BibTeX.
* Feel free to email us.
* Also, check out our search engine, as an example application.
## Model description
This model is a fine-tuned version of PubMedBERT on the scientific-challenges-and-directions-dataset, designed for multi-label text classification.
## Training and evaluation data
The scientific-challenges-and-directions model is trained based on a dataset that is a collection of 2894 sentences and their surrounding contexts, from 1786 full-text papers in the CORD-19 corpus, labeled for classification of challenges and directions by expert annotators with biomedical and bioNLP backgrounds. For full details on the train/test/split of the data see section 3.1 in our paper
## Example notebook
We include an example notebook that uses the model for inference in our repo. See 'Inference_Notebook.ipynb'.
A training notebook is also included.
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning rate: 2e-05
- train batch size: 8
- eval batch size: 4
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr scheduler type: linear
- lr scheduler warmup steps: 500
- num epochs: 30
### Training results
The achieves the following results on the test set:
- Precision Challenge: 0.768719
- Recall Challenge: 0.780405
- F1 Challenge: 0.774518
- Precision Direction: 0.758112
- Recall Direction: 0.774096
- F1 Direction: 0.766021
- Precision (micro avg. on both labels): 0.764894
- Recall (micro avg. on both labels): 0.778139
- F1 (micro avg. on both labels): 0.771459
### Framework versions
- Transformers 4.15.0
- Pytorch 1.10.0+cu111
- Datasets 1.17.0
- Tokenizers 0.10.3
If using our dataset and models, please cite:
## Contact us
Please don't hesitate to reach out.
Email: 'lahav@URL','tomh@URL'.
| [
"# scientific-challenges-and-directions\n\nWe present a novel resource to help scientists and medical professionals discover challenges and potential directions across scientific literature, focusing on a broad corpus pertaining to the COVID-19 pandemic and related historical research. At a high level, the _challenges_ and _directions_ are defined as follows:\n\n* Challenge: A sentence mentioning a problem, difficulty, flaw, limitation, failure, lack of clarity, or knowledge gap.\n* Research direction: A sentence mentioning suggestions or needs for further research, hypotheses, speculations, indications or hints that an issue is worthy of exploration. \n\n* This model here is described in our paper: A Search Engine for Discovery of Scientific Challenges and Directions (though we've upgraded the infrastructure since the paper was released - there are slight differences in the results).\n* Our dataset can be found here.\n* Please cite our paper if you use our datasets or models in your project. See the BibTeX. \n* Feel free to email us. \n* Also, check out our search engine, as an example application.",
"## Model description\nThis model is a fine-tuned version of PubMedBERT on the scientific-challenges-and-directions-dataset, designed for multi-label text classification.",
"## Training and evaluation data\nThe scientific-challenges-and-directions model is trained based on a dataset that is a collection of 2894 sentences and their surrounding contexts, from 1786 full-text papers in the CORD-19 corpus, labeled for classification of challenges and directions by expert annotators with biomedical and bioNLP backgrounds. For full details on the train/test/split of the data see section 3.1 in our paper",
"## Example notebook\nWe include an example notebook that uses the model for inference in our repo. See 'Inference_Notebook.ipynb'.\nA training notebook is also included.",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning rate: 2e-05\n- train batch size: 8\n- eval batch size: 4\n- seed: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr scheduler type: linear\n- lr scheduler warmup steps: 500\n- num epochs: 30",
"### Training results\nThe achieves the following results on the test set:\n- Precision Challenge: 0.768719 \n- Recall Challenge: 0.780405 \n- F1 Challenge: 0.774518\n- Precision Direction: 0.758112 \n- Recall Direction: 0.774096 \n- F1 Direction: 0.766021\n- Precision (micro avg. on both labels): 0.764894 \n- Recall (micro avg. on both labels): 0.778139 \n- F1 (micro avg. on both labels): 0.771459",
"### Framework versions\n\n- Transformers 4.15.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3\n\nIf using our dataset and models, please cite:",
"## Contact us\n\nPlease don't hesitate to reach out.\n\nEmail: 'lahav@URL','tomh@URL'."
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #en #dataset-DanL/scientific-challenges-and-directions-dataset #arxiv-2108.13751 #autotrain_compatible #endpoints_compatible #region-us \n",
"# scientific-challenges-and-directions\n\nWe present a novel resource to help scientists and medical professionals discover challenges and potential directions across scientific literature, focusing on a broad corpus pertaining to the COVID-19 pandemic and related historical research. At a high level, the _challenges_ and _directions_ are defined as follows:\n\n* Challenge: A sentence mentioning a problem, difficulty, flaw, limitation, failure, lack of clarity, or knowledge gap.\n* Research direction: A sentence mentioning suggestions or needs for further research, hypotheses, speculations, indications or hints that an issue is worthy of exploration. \n\n* This model here is described in our paper: A Search Engine for Discovery of Scientific Challenges and Directions (though we've upgraded the infrastructure since the paper was released - there are slight differences in the results).\n* Our dataset can be found here.\n* Please cite our paper if you use our datasets or models in your project. See the BibTeX. \n* Feel free to email us. \n* Also, check out our search engine, as an example application.",
"## Model description\nThis model is a fine-tuned version of PubMedBERT on the scientific-challenges-and-directions-dataset, designed for multi-label text classification.",
"## Training and evaluation data\nThe scientific-challenges-and-directions model is trained based on a dataset that is a collection of 2894 sentences and their surrounding contexts, from 1786 full-text papers in the CORD-19 corpus, labeled for classification of challenges and directions by expert annotators with biomedical and bioNLP backgrounds. For full details on the train/test/split of the data see section 3.1 in our paper",
"## Example notebook\nWe include an example notebook that uses the model for inference in our repo. See 'Inference_Notebook.ipynb'.\nA training notebook is also included.",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning rate: 2e-05\n- train batch size: 8\n- eval batch size: 4\n- seed: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr scheduler type: linear\n- lr scheduler warmup steps: 500\n- num epochs: 30",
"### Training results\nThe achieves the following results on the test set:\n- Precision Challenge: 0.768719 \n- Recall Challenge: 0.780405 \n- F1 Challenge: 0.774518\n- Precision Direction: 0.758112 \n- Recall Direction: 0.774096 \n- F1 Direction: 0.766021\n- Precision (micro avg. on both labels): 0.764894 \n- Recall (micro avg. on both labels): 0.778139 \n- F1 (micro avg. on both labels): 0.771459",
"### Framework versions\n\n- Transformers 4.15.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3\n\nIf using our dataset and models, please cite:",
"## Contact us\n\nPlease don't hesitate to reach out.\n\nEmail: 'lahav@URL','tomh@URL'."
] | [
63,
221,
37,
89,
38,
4,
95,
126,
55,
32
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #en #dataset-DanL/scientific-challenges-and-directions-dataset #arxiv-2108.13751 #autotrain_compatible #endpoints_compatible #region-us \n# scientific-challenges-and-directions\n\nWe present a novel resource to help scientists and medical professionals discover challenges and potential directions across scientific literature, focusing on a broad corpus pertaining to the COVID-19 pandemic and related historical research. At a high level, the _challenges_ and _directions_ are defined as follows:\n\n* Challenge: A sentence mentioning a problem, difficulty, flaw, limitation, failure, lack of clarity, or knowledge gap.\n* Research direction: A sentence mentioning suggestions or needs for further research, hypotheses, speculations, indications or hints that an issue is worthy of exploration. \n\n* This model here is described in our paper: A Search Engine for Discovery of Scientific Challenges and Directions (though we've upgraded the infrastructure since the paper was released - there are slight differences in the results).\n* Our dataset can be found here.\n* Please cite our paper if you use our datasets or models in your project. See the BibTeX. \n* Feel free to email us. \n* Also, check out our search engine, as an example application.## Model description\nThis model is a fine-tuned version of PubMedBERT on the scientific-challenges-and-directions-dataset, designed for multi-label text classification.## Training and evaluation data\nThe scientific-challenges-and-directions model is trained based on a dataset that is a collection of 2894 sentences and their surrounding contexts, from 1786 full-text papers in the CORD-19 corpus, labeled for classification of challenges and directions by expert annotators with biomedical and bioNLP backgrounds. For full details on the train/test/split of the data see section 3.1 in our paper## Example notebook\nWe include an example notebook that uses the model for inference in our repo. See 'Inference_Notebook.ipynb'.\nA training notebook is also included.## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning rate: 2e-05\n- train batch size: 8\n- eval batch size: 4\n- seed: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr scheduler type: linear\n- lr scheduler warmup steps: 500\n- num epochs: 30### Training results\nThe achieves the following results on the test set:\n- Precision Challenge: 0.768719 \n- Recall Challenge: 0.780405 \n- F1 Challenge: 0.774518\n- Precision Direction: 0.758112 \n- Recall Direction: 0.774096 \n- F1 Direction: 0.766021\n- Precision (micro avg. on both labels): 0.764894 \n- Recall (micro avg. on both labels): 0.778139 \n- F1 (micro avg. on both labels): 0.771459### Framework versions\n\n- Transformers 4.15.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3\n\nIf using our dataset and models, please cite:## Contact us\n\nPlease don't hesitate to reach out.\n\nEmail: 'lahav@URL','tomh@URL'."
] |
text-classification | transformers | Fine-tuned CovidBERT on Med-Marco Dataset for passage ranking
# CovidBERT-MedNLI
This is the model **CovidBERT** trained by DeepSet on AllenAI's [CORD19 Dataset](https://pages.semanticscholar.org/coronavirus-research) of scientific articles about coronaviruses.
The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [1] using the **average pooling strategy** and a **softmax loss**.
It is further fine-tuned Med-Marco Dataset. MacAvaney et.al in their [paper](https://arxiv.org/abs/2010.05987) titled “SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search” used MedSyn a lexicon of layperson and expert terminology for various medical conditions to filter for medical questions. One can also replace this by UMLs ontologies but the beauty of MedSyn is that the terms are more general human conversation lingo and not terms based on scientific literature.
Parameter details for the original training on CORD-19 are available on [DeepSet's MLFlow](https://public-mlflow.deepset.ai/#/experiments/2/runs/ba27d00c30044ef6a33b1d307b4a6cba)
**Base model**: `deepset/covid_bert_base` from HuggingFace's `AutoModel`.
| {} | Darkrider/covidbert_medmarco | null | [
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"arxiv:2010.05987",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2010.05987"
] | [] | TAGS
#transformers #pytorch #jax #bert #text-classification #arxiv-2010.05987 #autotrain_compatible #endpoints_compatible #region-us
| Fine-tuned CovidBERT on Med-Marco Dataset for passage ranking
# CovidBERT-MedNLI
This is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.
The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.
It is further fine-tuned Med-Marco Dataset. MacAvaney URL in their paper titled “SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search” used MedSyn a lexicon of layperson and expert terminology for various medical conditions to filter for medical questions. One can also replace this by UMLs ontologies but the beauty of MedSyn is that the terms are more general human conversation lingo and not terms based on scientific literature.
Parameter details for the original training on CORD-19 are available on DeepSet's MLFlow
Base model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'.
| [
"# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\n\nIt is further fine-tuned Med-Marco Dataset. MacAvaney URL in their paper titled “SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search” used MedSyn a lexicon of layperson and expert terminology for various medical conditions to filter for medical questions. One can also replace this by UMLs ontologies but the beauty of MedSyn is that the terms are more general human conversation lingo and not terms based on scientific literature.\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #arxiv-2010.05987 #autotrain_compatible #endpoints_compatible #region-us \n",
"# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\n\nIt is further fine-tuned Med-Marco Dataset. MacAvaney URL in their paper titled “SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search” used MedSyn a lexicon of layperson and expert terminology for various medical conditions to filter for medical questions. One can also replace this by UMLs ontologies but the beauty of MedSyn is that the terms are more general human conversation lingo and not terms based on scientific literature.\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] | [
41,
238
] | [
"TAGS\n#transformers #pytorch #jax #bert #text-classification #arxiv-2010.05987 #autotrain_compatible #endpoints_compatible #region-us \n# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\n\nIt is further fine-tuned Med-Marco Dataset. MacAvaney URL in their paper titled “SLEDGE-Z: A Zero-Shot Baseline for COVID-19 Literature Search” used MedSyn a lexicon of layperson and expert terminology for various medical conditions to filter for medical questions. One can also replace this by UMLs ontologies but the beauty of MedSyn is that the terms are more general human conversation lingo and not terms based on scientific literature.\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] |
null | transformers | # CovidBERT-MedNLI
This is the model **CovidBERT** trained by DeepSet on AllenAI's [CORD19 Dataset](https://pages.semanticscholar.org/coronavirus-research) of scientific articles about coronaviruses.
The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [1] using the **average pooling strategy** and a **softmax loss**.
It is further fine-tuned on both MedNLI datasets available at Physionet.
[ACL-BIONLP 2019](https://physionet.org/content/mednli-bionlp19/1.0.1/)
[MedNLI from MIMIC](https://physionet.org/content/mednli/1.0.0/)
Parameter details for the original training on CORD-19 are available on [DeepSet's MLFlow](https://public-mlflow.deepset.ai/#/experiments/2/runs/ba27d00c30044ef6a33b1d307b4a6cba)
**Base model**: `deepset/covid_bert_base` from HuggingFace's `AutoModel`.
| {} | Darkrider/covidbert_mednli | null | [
"transformers",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #endpoints_compatible #region-us
| # CovidBERT-MedNLI
This is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.
The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.
It is further fine-tuned on both MedNLI datasets available at Physionet.
ACL-BIONLP 2019
MedNLI from MIMIC
Parameter details for the original training on CORD-19 are available on DeepSet's MLFlow
Base model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'.
| [
"# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\nIt is further fine-tuned on both MedNLI datasets available at Physionet. \n\nACL-BIONLP 2019\n\nMedNLI from MIMIC\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] | [
"TAGS\n#transformers #endpoints_compatible #region-us \n",
"# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\nIt is further fine-tuned on both MedNLI datasets available at Physionet. \n\nACL-BIONLP 2019\n\nMedNLI from MIMIC\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] | [
12,
172
] | [
"TAGS\n#transformers #endpoints_compatible #region-us \n# CovidBERT-MedNLI\n\nThis is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses.\n\nThe model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss.\nIt is further fine-tuned on both MedNLI datasets available at Physionet. \n\nACL-BIONLP 2019\n\nMedNLI from MIMIC\n\n\nParameter details for the original training on CORD-19 are available on DeepSet's MLFlow\n\nBase model: 'deepset/covid_bert_base' from HuggingFace's 'AutoModel'."
] |
fill-mask | transformers |
# Marathi DistilBERT
## Model description
This model is an adaptation of DistilBERT (Victor Sanh et al., 2019) for Marathi language. This version of Marathi-DistilBERT is trained from scratch on approximately 11.2 million sentences.
```
DISCLAIMER
This model has not been thoroughly tested and may contain biased opinions or inappropriate language. User discretion is advised
```
## Training data
The training data has been extracted from a variety of sources, mainly including:
1. Oscar Corpus
2. Marathi Newspapers
3. Marathi storybooks and articles
The data is cleaned by removing all languages other than Marathi, while preserving common punctuations
## Training procedure
The model is trained from scratch using an Adam optimizer with a learning rate of 1e-4 and default β1 and β2 values of 0.9 and 0.999 respectively with a total batch size of 256 on a v3-8 TPU and mask probability of 15%.
## Example
```python
from transformers import pipeline
fill_mask = pipeline(
"fill-mask",
model="DarshanDeshpande/marathi-distilbert",
tokenizer="DarshanDeshpande/marathi-distilbert",
)
fill_mask("हा खरोखर चांगला [MASK] आहे.")
```
### BibTeX entry and citation info
```bibtex
@misc{sanh2020distilbert,
title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter},
author={Victor Sanh and Lysandre Debut and Julien Chaumond and Thomas Wolf},
year={2020},
eprint={1910.01108},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
<h3>Authors </h3>
<h5>1. Darshan Deshpande: <a href="https://github.com/DarshanDeshpande">GitHub</a>, <a href="https://www.linkedin.com/in/darshan-deshpande/">LinkedIn</a><h5>
<h5>2. Harshavardhan Abichandani: <a href="https://github.com/Baras64">GitHub</a>, <a href="http://www.linkedin.com/in/harsh-abhi">LinkedIn</a><h5> | {"language": ["mr"], "license": "apache-2.0", "tags": ["fill-mask"], "datasets": ["Oscar Corpus, News, Stories"], "widget": [{"text": "\u0939\u093e \u0916\u0930\u094b\u0916\u0930 \u091a\u093e\u0902\u0917\u0932\u093e [MASK] \u0906\u0939\u0947."}]} | DarshanDeshpande/marathi-distilbert | null | [
"transformers",
"pytorch",
"tf",
"distilbert",
"fill-mask",
"mr",
"arxiv:1910.01108",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1910.01108"
] | [
"mr"
] | TAGS
#transformers #pytorch #tf #distilbert #fill-mask #mr #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Marathi DistilBERT
## Model description
This model is an adaptation of DistilBERT (Victor Sanh et al., 2019) for Marathi language. This version of Marathi-DistilBERT is trained from scratch on approximately 11.2 million sentences.
## Training data
The training data has been extracted from a variety of sources, mainly including:
1. Oscar Corpus
2. Marathi Newspapers
3. Marathi storybooks and articles
The data is cleaned by removing all languages other than Marathi, while preserving common punctuations
## Training procedure
The model is trained from scratch using an Adam optimizer with a learning rate of 1e-4 and default β1 and β2 values of 0.9 and 0.999 respectively with a total batch size of 256 on a v3-8 TPU and mask probability of 15%.
## Example
### BibTeX entry and citation info
<h3>Authors </h3>
<h5>1. Darshan Deshpande: <a href="URL <a href="URL
<h5>2. Harshavardhan Abichandani: <a href="URL <a href="http://URL | [
"# Marathi DistilBERT",
"## Model description\n\nThis model is an adaptation of DistilBERT (Victor Sanh et al., 2019) for Marathi language. This version of Marathi-DistilBERT is trained from scratch on approximately 11.2 million sentences.",
"## Training data\nThe training data has been extracted from a variety of sources, mainly including:\n1. Oscar Corpus\n2. Marathi Newspapers\n3. Marathi storybooks and articles\n \nThe data is cleaned by removing all languages other than Marathi, while preserving common punctuations",
"## Training procedure\nThe model is trained from scratch using an Adam optimizer with a learning rate of 1e-4 and default β1 and β2 values of 0.9 and 0.999 respectively with a total batch size of 256 on a v3-8 TPU and mask probability of 15%.",
"## Example",
"### BibTeX entry and citation info\n\n\n\n<h3>Authors </h3>\n<h5>1. Darshan Deshpande: <a href=\"URL <a href=\"URL\n\n<h5>2. Harshavardhan Abichandani: <a href=\"URL <a href=\"http://URL"
] | [
"TAGS\n#transformers #pytorch #tf #distilbert #fill-mask #mr #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Marathi DistilBERT",
"## Model description\n\nThis model is an adaptation of DistilBERT (Victor Sanh et al., 2019) for Marathi language. This version of Marathi-DistilBERT is trained from scratch on approximately 11.2 million sentences.",
"## Training data\nThe training data has been extracted from a variety of sources, mainly including:\n1. Oscar Corpus\n2. Marathi Newspapers\n3. Marathi storybooks and articles\n \nThe data is cleaned by removing all languages other than Marathi, while preserving common punctuations",
"## Training procedure\nThe model is trained from scratch using an Adam optimizer with a learning rate of 1e-4 and default β1 and β2 values of 0.9 and 0.999 respectively with a total batch size of 256 on a v3-8 TPU and mask probability of 15%.",
"## Example",
"### BibTeX entry and citation info\n\n\n\n<h3>Authors </h3>\n<h5>1. Darshan Deshpande: <a href=\"URL <a href=\"URL\n\n<h5>2. Harshavardhan Abichandani: <a href=\"URL <a href=\"http://URL"
] | [
53,
5,
47,
53,
64,
3,
84
] | [
"TAGS\n#transformers #pytorch #tf #distilbert #fill-mask #mr #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Marathi DistilBERT## Model description\n\nThis model is an adaptation of DistilBERT (Victor Sanh et al., 2019) for Marathi language. This version of Marathi-DistilBERT is trained from scratch on approximately 11.2 million sentences.## Training data\nThe training data has been extracted from a variety of sources, mainly including:\n1. Oscar Corpus\n2. Marathi Newspapers\n3. Marathi storybooks and articles\n \nThe data is cleaned by removing all languages other than Marathi, while preserving common punctuations## Training procedure\nThe model is trained from scratch using an Adam optimizer with a learning rate of 1e-4 and default β1 and β2 values of 0.9 and 0.999 respectively with a total batch size of 256 on a v3-8 TPU and mask probability of 15%.## Example### BibTeX entry and citation info\n\n\n\n<h3>Authors </h3>\n<h5>1. Darshan Deshpande: <a href=\"URL <a href=\"URL\n\n<h5>2. Harshavardhan Abichandani: <a href=\"URL <a href=\"http://URL"
] |
fill-mask | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-retrained_ru_covid
This model is a fine-tuned version of [blinoff/roberta-base-russian-v0](https://huggingface.co/blinoff/roberta-base-russian-v0) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.8518
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 1
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 25
### Training results
### Framework versions
- Transformers 4.16.2
- Pytorch 1.10.0+cu111
- Datasets 1.18.3
- Tokenizers 0.11.0
| {"tags": ["generated_from_trainer"], "model-index": [{"name": "roberta-retrained_ru_covid", "results": []}]} | Daryaflp/roberta-retrained_ru_covid | null | [
"transformers",
"pytorch",
"tensorboard",
"roberta",
"fill-mask",
"generated_from_trainer",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
|
# roberta-retrained_ru_covid
This model is a fine-tuned version of blinoff/roberta-base-russian-v0 on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.8518
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 1
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 25
### Training results
### Framework versions
- Transformers 4.16.2
- Pytorch 1.10.0+cu111
- Datasets 1.18.3
- Tokenizers 0.11.0
| [
"# roberta-retrained_ru_covid\n\nThis model is a fine-tuned version of blinoff/roberta-base-russian-v0 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 1.8518",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 1\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0"
] | [
"TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n",
"# roberta-retrained_ru_covid\n\nThis model is a fine-tuned version of blinoff/roberta-base-russian-v0 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 1.8518",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 1\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0"
] | [
37,
56,
7,
9,
9,
4,
93,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# roberta-retrained_ru_covid\n\nThis model is a fine-tuned version of blinoff/roberta-base-russian-v0 on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 1.8518## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 1\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 25### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0"
] |
null | transformers | TinyBERT: Distilling BERT for Natural Language Understanding
========
**This model is a copy of [this model repository](https://huggingface.co/huawei-noah/TinyBERT_General_4L_312D) from Huawei Noah at the specific commit `34707a33cd59a94ecde241ac209bf35103691b43`.**
TinyBERT is 7.5x smaller and 9.4x faster on inference than BERT-base and achieves competitive performances in the tasks of natural language understanding. It performs a novel transformer distillation at both the pre-training and task-specific learning stages. In general distillation, we use the original BERT-base without fine-tuning as the teacher and a large-scale text corpus as the learning data. By performing the Transformer distillation on the text from general domain, we obtain a general TinyBERT which provides a good initialization for the task-specific distillation. We here provide the general TinyBERT for your tasks at hand.
For more details about the techniques of TinyBERT, refer to our paper:
[TinyBERT: Distilling BERT for Natural Language Understanding](https://arxiv.org/abs/1909.10351)
Citation
========
If you find TinyBERT useful in your research, please cite the following paper:
```
@article{jiao2019tinybert,
title={Tinybert: Distilling bert for natural language understanding},
author={Jiao, Xiaoqi and Yin, Yichun and Shang, Lifeng and Jiang, Xin and Chen, Xiao and Li, Linlin and Wang, Fang and Liu, Qun},
journal={arXiv preprint arXiv:1909.10351},
year={2019}
}
```
| {} | DataikuNLP/TinyBERT_General_4L_312D | null | [
"transformers",
"pytorch",
"jax",
"bert",
"arxiv:1909.10351",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1909.10351"
] | [] | TAGS
#transformers #pytorch #jax #bert #arxiv-1909.10351 #endpoints_compatible #region-us
| TinyBERT: Distilling BERT for Natural Language Understanding
========
This model is a copy of this model repository from Huawei Noah at the specific commit '34707a33cd59a94ecde241ac209bf35103691b43'.
TinyBERT is 7.5x smaller and 9.4x faster on inference than BERT-base and achieves competitive performances in the tasks of natural language understanding. It performs a novel transformer distillation at both the pre-training and task-specific learning stages. In general distillation, we use the original BERT-base without fine-tuning as the teacher and a large-scale text corpus as the learning data. By performing the Transformer distillation on the text from general domain, we obtain a general TinyBERT which provides a good initialization for the task-specific distillation. We here provide the general TinyBERT for your tasks at hand.
For more details about the techniques of TinyBERT, refer to our paper:
TinyBERT: Distilling BERT for Natural Language Understanding
Citation
========
If you find TinyBERT useful in your research, please cite the following paper:
| [] | [
"TAGS\n#transformers #pytorch #jax #bert #arxiv-1909.10351 #endpoints_compatible #region-us \n"
] | [
30
] | [
"TAGS\n#transformers #pytorch #jax #bert #arxiv-1909.10351 #endpoints_compatible #region-us \n"
] |
sentence-similarity | sentence-transformers |
# average_word_embeddings_glove.6B.300d
**This model is a copy of [this model repository](https://huggingface.co/sentence-transformers/average_word_embeddings_glove.6B.300d) from sentence-transformers at the specific commit `5d2b7d1c127036ae98b9d487eca4d48744edc709`.**
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 300 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('sentence-transformers/average_word_embeddings_glove.6B.300d')
embeddings = model.encode(sentences)
print(embeddings)
```
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/average_word_embeddings_glove.6B.300d)
## Full Model Architecture
```
SentenceTransformer(
(0): WordEmbeddings(
(emb_layer): Embedding(400001, 300)
)
(1): Pooling({'word_embedding_dimension': 300, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
```
## Citing & Authors
This model was trained by [sentence-transformers](https://www.sbert.net/).
If you find this model helpful, feel free to cite our publication [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://arxiv.org/abs/1908.10084):
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "http://arxiv.org/abs/1908.10084",
}
``` | {"license": "apache-2.0", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "sentence-similarity"} | DataikuNLP/average_word_embeddings_glove.6B.300d | null | [
"sentence-transformers",
"feature-extraction",
"sentence-similarity",
"arxiv:1908.10084",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1908.10084"
] | [] | TAGS
#sentence-transformers #feature-extraction #sentence-similarity #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us
|
# average_word_embeddings_glove.6B.300d
This model is a copy of this model repository from sentence-transformers at the specific commit '5d2b7d1c127036ae98b9d487eca4d48744edc709'.
This is a sentence-transformers model: It maps sentences & paragraphs to a 300 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Full Model Architecture
## Citing & Authors
This model was trained by sentence-transformers.
If you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:
| [
"# average_word_embeddings_glove.6B.300d\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '5d2b7d1c127036ae98b9d487eca4d48744edc709'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 300 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
"TAGS\n#sentence-transformers #feature-extraction #sentence-similarity #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# average_word_embeddings_glove.6B.300d\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '5d2b7d1c127036ae98b9d487eca4d48744edc709'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 300 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
40,
100,
30,
26,
5,
43
] | [
"TAGS\n#sentence-transformers #feature-extraction #sentence-similarity #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n# average_word_embeddings_glove.6B.300d\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '5d2b7d1c127036ae98b9d487eca4d48744edc709'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 300 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Full Model Architecture## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] |
fill-mask | transformers |
# CamemBERT: a Tasty French Language Model
**This model is a copy of [this model repository](https://huggingface.co/camembert-base) at the specific commit `482393b6198924f9da270b1aaf37d238aafca99b`.**
## Introduction
[CamemBERT](https://arxiv.org/abs/1911.03894) is a state-of-the-art language model for French based on the RoBERTa model.
It is now available on Hugging Face in 6 different versions with varying number of parameters, amount of pretraining data and pretraining data source domains.
For further information or requests, please go to [Camembert Website](https://camembert-model.fr/)
## Pre-trained models
| Model | #params | Arch. | Training data |
|--------------------------------|--------------------------------|-------|-----------------------------------|
| `camembert-base` | 110M | Base | OSCAR (138 GB of text) |
| `camembert/camembert-large` | 335M | Large | CCNet (135 GB of text) |
| `camembert/camembert-base-ccnet` | 110M | Base | CCNet (135 GB of text) |
| `camembert/camembert-base-wikipedia-4gb` | 110M | Base | Wikipedia (4 GB of text) |
| `camembert/camembert-base-oscar-4gb` | 110M | Base | Subsample of OSCAR (4 GB of text) |
| `camembert/camembert-base-ccnet-4gb` | 110M | Base | Subsample of CCNet (4 GB of text) |
## How to use CamemBERT with HuggingFace
##### Load CamemBERT and its sub-word tokenizer :
```python
from transformers import CamembertModel, CamembertTokenizer
# You can replace "camembert-base" with any other model from the table, e.g. "camembert/camembert-large".
tokenizer = CamembertTokenizer.from_pretrained("camembert-base")
camembert = CamembertModel.from_pretrained("camembert-base")
camembert.eval() # disable dropout (or leave in train mode to finetune)
```
##### Filling masks using pipeline
```python
from transformers import pipeline
camembert_fill_mask = pipeline("fill-mask", model="camembert-base", tokenizer="camembert-base")
results = camembert_fill_mask("Le camembert est <mask> :)")
# results
#[{'sequence': '<s> Le camembert est délicieux :)</s>', 'score': 0.4909103214740753, 'token': 7200},
# {'sequence': '<s> Le camembert est excellent :)</s>', 'score': 0.10556930303573608, 'token': 2183},
# {'sequence': '<s> Le camembert est succulent :)</s>', 'score': 0.03453315049409866, 'token': 26202},
# {'sequence': '<s> Le camembert est meilleur :)</s>', 'score': 0.03303130343556404, 'token': 528},
# {'sequence': '<s> Le camembert est parfait :)</s>', 'score': 0.030076518654823303, 'token': 1654}]
```
##### Extract contextual embedding features from Camembert output
```python
import torch
# Tokenize in sub-words with SentencePiece
tokenized_sentence = tokenizer.tokenize("J'aime le camembert !")
# ['▁J', "'", 'aime', '▁le', '▁ca', 'member', 't', '▁!']
# 1-hot encode and add special starting and end tokens
encoded_sentence = tokenizer.encode(tokenized_sentence)
# [5, 121, 11, 660, 16, 730, 25543, 110, 83, 6]
# NB: Can be done in one step : tokenize.encode("J'aime le camembert !")
# Feed tokens to Camembert as a torch tensor (batch dim 1)
encoded_sentence = torch.tensor(encoded_sentence).unsqueeze(0)
embeddings, _ = camembert(encoded_sentence)
# embeddings.detach()
# embeddings.size torch.Size([1, 10, 768])
# tensor([[[-0.0254, 0.0235, 0.1027, ..., -0.1459, -0.0205, -0.0116],
# [ 0.0606, -0.1811, -0.0418, ..., -0.1815, 0.0880, -0.0766],
# [-0.1561, -0.1127, 0.2687, ..., -0.0648, 0.0249, 0.0446],
# ...,
```
##### Extract contextual embedding features from all Camembert layers
```python
from transformers import CamembertConfig
# (Need to reload the model with new config)
config = CamembertConfig.from_pretrained("camembert-base", output_hidden_states=True)
camembert = CamembertModel.from_pretrained("camembert-base", config=config)
embeddings, _, all_layer_embeddings = camembert(encoded_sentence)
# all_layer_embeddings list of len(all_layer_embeddings) == 13 (input embedding layer + 12 self attention layers)
all_layer_embeddings[5]
# layer 5 contextual embedding : size torch.Size([1, 10, 768])
#tensor([[[-0.0032, 0.0075, 0.0040, ..., -0.0025, -0.0178, -0.0210],
# [-0.0996, -0.1474, 0.1057, ..., -0.0278, 0.1690, -0.2982],
# [ 0.0557, -0.0588, 0.0547, ..., -0.0726, -0.0867, 0.0699],
# ...,
```
## Authors
CamemBERT was trained and evaluated by Louis Martin\*, Benjamin Muller\*, Pedro Javier Ortiz Suárez\*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
## Citation
If you use our work, please cite:
```bibtex
@inproceedings{martin2020camembert,
title={CamemBERT: a Tasty French Language Model},
author={Martin, Louis and Muller, Benjamin and Su{\'a}rez, Pedro Javier Ortiz and Dupont, Yoann and Romary, Laurent and de la Clergerie, {\'E}ric Villemonte and Seddah, Djam{\'e} and Sagot, Beno{\^\i}t},
booktitle={Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics},
year={2020}
}
```
| {"language": "fr", "license": "mit", "datasets": ["oscar"]} | DataikuNLP/camembert-base | null | [
"transformers",
"pytorch",
"tf",
"camembert",
"fill-mask",
"fr",
"dataset:oscar",
"arxiv:1911.03894",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1911.03894"
] | [
"fr"
] | TAGS
#transformers #pytorch #tf #camembert #fill-mask #fr #dataset-oscar #arxiv-1911.03894 #license-mit #autotrain_compatible #endpoints_compatible #region-us
| CamemBERT: a Tasty French Language Model
========================================
This model is a copy of this model repository at the specific commit '482393b6198924f9da270b1aaf37d238aafca99b'.
Introduction
------------
CamemBERT is a state-of-the-art language model for French based on the RoBERTa model.
It is now available on Hugging Face in 6 different versions with varying number of parameters, amount of pretraining data and pretraining data source domains.
For further information or requests, please go to Camembert Website
Pre-trained models
------------------
How to use CamemBERT with HuggingFace
-------------------------------------
##### Load CamemBERT and its sub-word tokenizer :
##### Filling masks using pipeline
##### Extract contextual embedding features from Camembert output
##### Extract contextual embedding features from all Camembert layers
Authors
-------
CamemBERT was trained and evaluated by Louis Martin\*, Benjamin Muller\*, Pedro Javier Ortiz Suárez\*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.
If you use our work, please cite:
| [
"##### Load CamemBERT and its sub-word tokenizer :",
"##### Filling masks using pipeline",
"##### Extract contextual embedding features from Camembert output",
"##### Extract contextual embedding features from all Camembert layers\n\n\nAuthors\n-------\n\n\nCamemBERT was trained and evaluated by Louis Martin\\*, Benjamin Muller\\*, Pedro Javier Ortiz Suárez\\*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.\n\n\nIf you use our work, please cite:"
] | [
"TAGS\n#transformers #pytorch #tf #camembert #fill-mask #fr #dataset-oscar #arxiv-1911.03894 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"##### Load CamemBERT and its sub-word tokenizer :",
"##### Filling masks using pipeline",
"##### Extract contextual embedding features from Camembert output",
"##### Extract contextual embedding features from all Camembert layers\n\n\nAuthors\n-------\n\n\nCamemBERT was trained and evaluated by Louis Martin\\*, Benjamin Muller\\*, Pedro Javier Ortiz Suárez\\*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.\n\n\nIf you use our work, please cite:"
] | [
55,
17,
9,
17,
90
] | [
"TAGS\n#transformers #pytorch #tf #camembert #fill-mask #fr #dataset-oscar #arxiv-1911.03894 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n##### Load CamemBERT and its sub-word tokenizer :##### Filling masks using pipeline##### Extract contextual embedding features from Camembert output##### Extract contextual embedding features from all Camembert layers\n\n\nAuthors\n-------\n\n\nCamemBERT was trained and evaluated by Louis Martin\\*, Benjamin Muller\\*, Pedro Javier Ortiz Suárez\\*, Yoann Dupont, Laurent Romary, Éric Villemonte de la Clergerie, Djamé Seddah and Benoît Sagot.\n\n\nIf you use our work, please cite:"
] |
sentence-similarity | sentence-transformers |
# DataikuNLP/distiluse-base-multilingual-cased-v1
**This model is a copy of [this model repository](https://huggingface.co/sentence-transformers/distiluse-base-multilingual-cased-v1) from sentence-transformers at the specific commit `3a706e4d65c04f868c4684adfd4da74141be8732`.**
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('sentence-transformers/distiluse-base-multilingual-cased-v1')
embeddings = model.encode(sentences)
print(embeddings)
```
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/distiluse-base-multilingual-cased-v1)
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: DistilBertModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
(2): Dense({'in_features': 768, 'out_features': 512, 'bias': True, 'activation_function': 'torch.nn.modules.activation.Tanh'})
)
```
## Citing & Authors
This model was trained by [sentence-transformers](https://www.sbert.net/).
If you find this model helpful, feel free to cite our publication [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://arxiv.org/abs/1908.10084):
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "http://arxiv.org/abs/1908.10084",
}
``` | {"license": "apache-2.0", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers"], "pipeline_tag": "sentence-similarity"} | DataikuNLP/distiluse-base-multilingual-cased-v1 | null | [
"sentence-transformers",
"pytorch",
"distilbert",
"feature-extraction",
"sentence-similarity",
"transformers",
"arxiv:1908.10084",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1908.10084"
] | [] | TAGS
#sentence-transformers #pytorch #distilbert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us
|
# DataikuNLP/distiluse-base-multilingual-cased-v1
This model is a copy of this model repository from sentence-transformers at the specific commit '3a706e4d65c04f868c4684adfd4da74141be8732'.
This is a sentence-transformers model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Full Model Architecture
## Citing & Authors
This model was trained by sentence-transformers.
If you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:
| [
"# DataikuNLP/distiluse-base-multilingual-cased-v1\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '3a706e4d65c04f868c4684adfd4da74141be8732'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
"TAGS\n#sentence-transformers #pytorch #distilbert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# DataikuNLP/distiluse-base-multilingual-cased-v1\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '3a706e4d65c04f868c4684adfd4da74141be8732'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
51,
105,
30,
26,
5,
43
] | [
"TAGS\n#sentence-transformers #pytorch #distilbert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n# DataikuNLP/distiluse-base-multilingual-cased-v1\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '3a706e4d65c04f868c4684adfd4da74141be8732'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 512 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Full Model Architecture## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] |
sentence-similarity | sentence-transformers |
# DataikuNLP/paraphrase-MiniLM-L6-v2
**This model is a copy of [this model repository](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L6-v2/) from sentence-transformers at the specific commit `c4dfcde8a3e3e17e85cd4f0ec1925a266187f48e`.**
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('sentence-transformers/paraphrase-MiniLM-L6-v2')
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
```python
from transformers import AutoTokenizer, AutoModel
import torch
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('sentence-transformers/paraphrase-MiniLM-L6-v2')
model = AutoModel.from_pretrained('sentence-transformers/paraphrase-MiniLM-L6-v2')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling. In this case, max pooling.
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
print("Sentence embeddings:")
print(sentence_embeddings)
```
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/paraphrase-MiniLM-L6-v2)
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
```
## Citing & Authors
This model was trained by [sentence-transformers](https://www.sbert.net/).
If you find this model helpful, feel free to cite our publication [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://arxiv.org/abs/1908.10084):
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "http://arxiv.org/abs/1908.10084",
}
``` | {"license": "apache-2.0", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers"], "pipeline_tag": "sentence-similarity"} | DataikuNLP/paraphrase-MiniLM-L6-v2 | null | [
"sentence-transformers",
"pytorch",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"arxiv:1908.10084",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1908.10084"
] | [] | TAGS
#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us
|
# DataikuNLP/paraphrase-MiniLM-L6-v2
This model is a copy of this model repository from sentence-transformers at the specific commit 'c4dfcde8a3e3e17e85cd4f0ec1925a266187f48e'.
This is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Full Model Architecture
## Citing & Authors
This model was trained by sentence-transformers.
If you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:
| [
"# DataikuNLP/paraphrase-MiniLM-L6-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'c4dfcde8a3e3e17e85cd4f0ec1925a266187f48e'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
"TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# DataikuNLP/paraphrase-MiniLM-L6-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'c4dfcde8a3e3e17e85cd4f0ec1925a266187f48e'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
49,
102,
30,
58,
26,
5,
43
] | [
"TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n# DataikuNLP/paraphrase-MiniLM-L6-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'c4dfcde8a3e3e17e85cd4f0ec1925a266187f48e'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Full Model Architecture## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] |
sentence-similarity | sentence-transformers |
# DataikuNLP/paraphrase-albert-small-v2
**This model is a copy of [this model repository](https://huggingface.co/sentence-transformers/paraphrase-albert-small-v2/) from sentence-transformers at the specific commit `1eb1996223dd90a4c25be2fc52f6f336419a0d52`.**
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('sentence-transformers/paraphrase-albert-small-v2')
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
```python
from transformers import AutoTokenizer, AutoModel
import torch
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('sentence-transformers/paraphrase-albert-small-v2')
model = AutoModel.from_pretrained('sentence-transformers/paraphrase-albert-small-v2')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling. In this case, max pooling.
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
print("Sentence embeddings:")
print(sentence_embeddings)
```
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/paraphrase-albert-small-v2)
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 100, 'do_lower_case': False}) with Transformer model: AlbertModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
```
## Citing & Authors
This model was trained by [sentence-transformers](https://www.sbert.net/).
If you find this model helpful, feel free to cite our publication [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://arxiv.org/abs/1908.10084):
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "http://arxiv.org/abs/1908.10084",
}
``` | {"license": "apache-2.0", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers"], "pipeline_tag": "sentence-similarity"} | DataikuNLP/paraphrase-albert-small-v2 | null | [
"sentence-transformers",
"pytorch",
"albert",
"feature-extraction",
"sentence-similarity",
"transformers",
"arxiv:1908.10084",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1908.10084"
] | [] | TAGS
#sentence-transformers #pytorch #albert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us
|
# DataikuNLP/paraphrase-albert-small-v2
This model is a copy of this model repository from sentence-transformers at the specific commit '1eb1996223dd90a4c25be2fc52f6f336419a0d52'.
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Full Model Architecture
## Citing & Authors
This model was trained by sentence-transformers.
If you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:
| [
"# DataikuNLP/paraphrase-albert-small-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '1eb1996223dd90a4c25be2fc52f6f336419a0d52'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
"TAGS\n#sentence-transformers #pytorch #albert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# DataikuNLP/paraphrase-albert-small-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '1eb1996223dd90a4c25be2fc52f6f336419a0d52'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
49,
98,
30,
58,
26,
5,
43
] | [
"TAGS\n#sentence-transformers #pytorch #albert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n# DataikuNLP/paraphrase-albert-small-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit '1eb1996223dd90a4c25be2fc52f6f336419a0d52'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Full Model Architecture## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] |
sentence-similarity | sentence-transformers |
# DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2
**This model is a copy of [this model repository](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2) from sentence-transformers at the specific commit `d66eff4d8a8598f264f166af8db67f7797164651`.**
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2')
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
```python
from transformers import AutoTokenizer, AutoModel
import torch
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ['This is an example sentence', 'Each sentence is converted']
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2')
model = AutoModel.from_pretrained('sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling. In this case, max pooling.
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
print("Sentence embeddings:")
print(sentence_embeddings)
```
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
)
```
## Citing & Authors
This model was trained by [sentence-transformers](https://www.sbert.net/).
If you find this model helpful, feel free to cite our publication [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://arxiv.org/abs/1908.10084):
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "http://arxiv.org/abs/1908.10084",
}
``` | {"license": "apache-2.0", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "transformers"], "pipeline_tag": "sentence-similarity"} | DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2 | null | [
"sentence-transformers",
"pytorch",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"arxiv:1908.10084",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1908.10084"
] | [] | TAGS
#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us
|
# DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2
This model is a copy of this model repository from sentence-transformers at the specific commit 'd66eff4d8a8598f264f166af8db67f7797164651'.
This is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Full Model Architecture
## Citing & Authors
This model was trained by sentence-transformers.
If you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:
| [
"# DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'd66eff4d8a8598f264f166af8db67f7797164651'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
"TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'd66eff4d8a8598f264f166af8db67f7797164651'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Full Model Architecture",
"## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] | [
49,
107,
30,
58,
26,
5,
43
] | [
"TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #transformers #arxiv-1908.10084 #license-apache-2.0 #endpoints_compatible #region-us \n# DataikuNLP/paraphrase-multilingual-MiniLM-L12-v2\n\nThis model is a copy of this model repository from sentence-transformers at the specific commit 'd66eff4d8a8598f264f166af8db67f7797164651'.\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Full Model Architecture## Citing & Authors\n\nThis model was trained by sentence-transformers. \n \nIf you find this model helpful, feel free to cite our publication Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks:"
] |
fill-mask | transformers | Hugging Face's logo
---
language: am
datasets:
---
# bert-base-multilingual-cased-finetuned-amharic
## Model description
**bert-base-multilingual-cased-finetuned-amharic** is a **Amharic BERT** model obtained by replacing mBERT vocabulary by amharic vocabulary because the language was not supported, and fine-tuning **bert-base-multilingual-cased** model on Amharic language texts. It provides **better performance** than the multilingual Amharic on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Amharic corpus using Amharic vocabulary.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-amharic')
>>> unmasker("የአሜሪካ የአፍሪካ ቀንድ ልዩ መልዕክተኛ ጄፈሪ ፌልትማን በአራት አገራት የሚያደጉትን [MASK] መጀመራቸውን የአሜሪካ የውጪ ጉዳይ ሚንስቴር አስታወቀ።")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Amharic CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | am_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 0.0 | 60.89
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-amharic | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
-------------------
language: am
datasets:
---
bert-base-multilingual-cased-finetuned-amharic
==============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-amharic is a Amharic BERT model obtained by replacing mBERT vocabulary by amharic vocabulary because the language was not supported, and fine-tuning bert-base-multilingual-cased model on Amharic language texts. It provides better performance than the multilingual Amharic on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Amharic corpus using Amharic vocabulary.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Amharic CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 0.0, am\_bert F1: 60.89
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 0.0, am\\_bert F1: 60.89",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 0.0, am\\_bert F1: 60.89",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
32,
22,
203,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #has_space #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 0.0, am\\_bert F1: 60.89### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: ha
datasets:
---
# bert-base-multilingual-cased-finetuned-hausa
## Model description
**bert-base-multilingual-cased-finetuned-hausa** is a **Hausa BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Hausa language texts. It provides **better performance** than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Hausa corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-hausa')
>>> unmasker("Shugaban [MASK] Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci")
[{'sequence':
'[CLS] Shugaban Nigeria Muhammadu Buhari ya amince da shawarar da ma [UNK] aikatar sufuri karkashin jagoranci [SEP]',
'score': 0.9762618541717529,
'token': 22045,
'token_str': 'Nigeria'},
{'sequence': '[CLS] Shugaban Ka Muhammadu Buhari ya amince da shawarar da ma [UNK] aikatar sufuri karkashin jagoranci [SEP]', 'score': 0.007239189930260181,
'token': 25444,
'token_str': 'Ka'},
{'sequence': '[CLS] Shugaban, Muhammadu Buhari ya amince da shawarar da ma [UNK] aikatar sufuri karkashin jagoranci [SEP]', 'score': 0.001990817254409194,
'token': 117,
'token_str': ','},
{'sequence': '[CLS] Shugaban Ghana Muhammadu Buhari ya amince da shawarar da ma [UNK] aikatar sufuri karkashin jagoranci [SEP]', 'score': 0.001566368737258017,
'token': 28682,
'token_str': 'Ghana'},
{'sequence': '[CLS] Shugabanmu Muhammadu Buhari ya amince da shawarar da ma [UNK] aikatar sufuri karkashin jagoranci [SEP]', 'score': 0.0009375187801197171,
'token': 11717,
'token_str': '##mu'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Hausa CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | ha_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 86.65 | 91.31
[VOA Hausa Textclass](https://huggingface.co/datasets/hausa_voa_topics) | 84.76 | 90.98
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-hausa | null | [
"transformers",
"pytorch",
"tf",
"jax",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: ha
datasets:
---
bert-base-multilingual-cased-finetuned-hausa
============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-hausa is a Hausa BERT model obtained by fine-tuning bert-base-multilingual-cased model on Hausa language texts. It provides better performance than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Hausa corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Hausa CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 86.65, ha\_bert F1: 91.31
Dataset: VOA Hausa Textclass, mBERT F1: 84.76, ha\_bert F1: 90.98
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.65, ha\\_bert F1: 91.31\nDataset: VOA Hausa Textclass, mBERT F1: 84.76, ha\\_bert F1: 90.98",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.65, ha\\_bert F1: 91.31\nDataset: VOA Hausa Textclass, mBERT F1: 84.76, ha\\_bert F1: 90.98",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
33,
22,
229,
15
] | [
"TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.65, ha\\_bert F1: 91.31\nDataset: VOA Hausa Textclass, mBERT F1: 84.76, ha\\_bert F1: 90.98### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: ig
datasets:
---
# bert-base-multilingual-cased-finetuned-igbo
## Model description
**bert-base-multilingual-cased-finetuned-igbo** is a **Igbo BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Igbo language texts. It provides **better performance** than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Igbo corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-igbo')
>>> unmasker("Reno Omokri na Gọọmentị [MASK] enweghị ihe ha ga-eji hiwe ya bụ mmachi.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + OPUS CC-Align + [IGBO NLP Corpus](https://github.com/IgnatiusEzeani/IGBONLP) +[Igbo CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | ig_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 85.11 | 86.75
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-igbo | null | [
"transformers",
"pytorch",
"safetensors",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: ig
datasets:
---
bert-base-multilingual-cased-finetuned-igbo
===========================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-igbo is a Igbo BERT model obtained by fine-tuning bert-base-multilingual-cased model on Igbo language texts. It provides better performance than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Igbo corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 85.11, ig\_bert F1: 86.75
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 85.11, ig\\_bert F1: 86.75",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 85.11, ig\\_bert F1: 86.75",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
32,
22,
221,
15
] | [
"TAGS\n#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 85.11, ig\\_bert F1: 86.75### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: rw
datasets:
---
# bert-base-multilingual-cased-finetuned-kinyarwanda
## Model description
**bert-base-multilingual-cased-finetuned-kinyarwanda** is a **Kinyarwanda BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Kinyarwanda language texts. It provides **better performance** than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Kinyarwanda corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-kinyarwanda')
>>> unmasker("Twabonye ko igihe mu [MASK] hazaba hari ikirango abantu bakunze")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [KIRNEWS](https://github.com/Andrews2017/KINNEWS-and-KIRNEWS-Corpus) + [BBC Gahuza](https://www.bbc.com/gahuza)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | rw_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 72.20 | 77.57
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-kinyarwanda | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: rw
datasets:
---
bert-base-multilingual-cased-finetuned-kinyarwanda
==================================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-kinyarwanda is a Kinyarwanda BERT model obtained by fine-tuning bert-base-multilingual-cased model on Kinyarwanda language texts. It provides better performance than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Kinyarwanda corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 72.20, rw\_bert F1: 77.57
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 72.20, rw\\_bert F1: 77.57",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 72.20, rw\\_bert F1: 77.57",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
28,
22,
211,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 72.20, rw\\_bert F1: 77.57### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: lg
datasets:
---
# bert-base-multilingual-cased-finetuned-luganda
## Model description
**bert-base-multilingual-cased-finetuned-luganda** is a **Luganda BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Luganda language texts. It provides **better performance** than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Luganda corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-luganda')
>>> unmasker("Ffe tulwanyisa abo abaagala okutabangula [MASK], Kimuli bwe yategeezezza.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [BUKKEDDE](https://github.com/masakhane-io/masakhane-ner/tree/main/text_by_language/luganda) +[Luganda CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | lg_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 80.36 | 84.70
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-luganda | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: lg
datasets:
---
bert-base-multilingual-cased-finetuned-luganda
==============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-luganda is a Luganda BERT model obtained by fine-tuning bert-base-multilingual-cased model on Luganda language texts. It provides better performance than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Luganda corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 80.36, lg\_bert F1: 84.70
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 80.36, lg\\_bert F1: 84.70",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 80.36, lg\\_bert F1: 84.70",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
28,
22,
214,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 80.36, lg\\_bert F1: 84.70### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: luo
datasets:
---
# bert-base-multilingual-cased-finetuned-luo
## Model description
**bert-base-multilingual-cased-finetuned-luo** is a **Luo BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Luo language texts. It provides **better performance** than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Luo corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-luo')
>>> unmasker("Obila ma Changamwe [MASK] pedho achije angwen mag njore")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | luo_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 74.22 | 75.59
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-luo | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: luo
datasets:
---
bert-base-multilingual-cased-finetuned-luo
==========================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-luo is a Luo BERT model obtained by fine-tuning bert-base-multilingual-cased model on Luo language texts. It provides better performance than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Luo corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 74.22, luo\_bert F1: 75.59
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 74.22, luo\\_bert F1: 75.59",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 74.22, luo\\_bert F1: 75.59",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
28,
22,
202,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 74.22, luo\\_bert F1: 75.59### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: pcm
datasets:
---
# bert-base-multilingual-cased-finetuned-naija
## Model description
**bert-base-multilingual-cased-finetuned-naija** is a **Nigerian-Pidgin BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Nigerian-Pidgin language texts. It provides **better performance** than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Nigerian-Pidgin corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-naija')
>>> unmasker("Another attack on ambulance happen for Koforidua in March [MASK] year where robbers kill Ambulance driver")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [BBC Pidgin](https://www.bbc.com/pidgin)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | pcm_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 87.23 | 89.95
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-naija | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: pcm
datasets:
---
bert-base-multilingual-cased-finetuned-naija
============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-naija is a Nigerian-Pidgin BERT model obtained by fine-tuning bert-base-multilingual-cased model on Nigerian-Pidgin language texts. It provides better performance than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Nigerian-Pidgin corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + BBC Pidgin
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 87.23, pcm\_bert F1: 89.95
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 87.23, pcm\\_bert F1: 89.95",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 87.23, pcm\\_bert F1: 89.95",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
28,
22,
207,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 87.23, pcm\\_bert F1: 89.95### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: ha
datasets:
---
# bert-base-multilingual-cased-finetuned-swahili
## Model description
**bert-base-multilingual-cased-finetuned-swahili** is a **Swahili BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Swahili language texts. It provides **better performance** than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Swahili corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-swahili')
>>> unmasker("Jumatatu, Bwana Kagame alielezea shirika la France24 huko [MASK] kwamba "hakuna uhalifu ulitendwa")
[{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Paris kwamba hakuna uhalifu ulitendwa',
'score': 0.31642526388168335,
'token': 10728,
'token_str': 'Paris'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Rwanda kwamba hakuna uhalifu ulitendwa',
'score': 0.15753623843193054,
'token': 57557,
'token_str': 'Rwanda'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Burundi kwamba hakuna uhalifu ulitendwa',
'score': 0.07211585342884064,
'token': 57824,
'token_str': 'Burundi'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko France kwamba hakuna uhalifu ulitendwa',
'score': 0.029844321310520172,
'token': 10688,
'token_str': 'France'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Senegal kwamba hakuna uhalifu ulitendwa',
'score': 0.0265930388122797,
'token': 38052,
'token_str': 'Senegal'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Swahili CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | sw_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 86.80 | 89.36
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-swahili | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: ha
datasets:
---
bert-base-multilingual-cased-finetuned-swahili
==============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-swahili is a Swahili BERT model obtained by fine-tuning bert-base-multilingual-cased model on Swahili language texts. It provides better performance than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Swahili corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Swahili CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 86.80, sw\_bert F1: 89.36
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.80, sw\\_bert F1: 89.36",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.80, sw\\_bert F1: 89.36",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
35,
22,
203,
15
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 86.80, sw\\_bert F1: 89.36### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: wo
datasets:
---
# bert-base-multilingual-cased-finetuned-wolof
## Model description
**bert-base-multilingual-cased-finetuned-wolof** is a **Wolof BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Wolof language texts. It provides **better performance** than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Wolof corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-wolof')
>>> unmasker("Màkki Sàll feeñal na ay xalaatam ci mbir yu am solo yu soxal [MASK] ak Afrik.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Bible OT](http://biblewolof.com/) + [OPUS](https://opus.nlpl.eu/) + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | wo_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 64.52 | 69.43
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-wolof | null | [
"transformers",
"pytorch",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: wo
datasets:
---
bert-base-multilingual-cased-finetuned-wolof
============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-wolof is a Wolof BERT model obtained by fine-tuning bert-base-multilingual-cased model on Wolof language texts. It provides better performance than the multilingual BERT on named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Wolof corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 64.52, wo\_bert F1: 69.43
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 64.52, wo\\_bert F1: 69.43",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 64.52, wo\\_bert F1: 69.43",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
28,
22,
222,
15
] | [
"TAGS\n#transformers #pytorch #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 64.52, wo\\_bert F1: 69.43### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: yo
datasets:
---
# bert-base-multilingual-cased-finetuned-yoruba
## Model description
**bert-base-multilingual-cased-finetuned-yoruba** is a **Yoruba BERT** model obtained by fine-tuning **bert-base-multilingual-cased** model on Yorùbá language texts. It provides **better performance** than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Yorùbá corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/bert-base-multilingual-cased-finetuned-yoruba')
>>> unmasker("Arẹmọ Phillip to jẹ ọkọ [MASK] Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun")
[{'sequence': '[CLS] Arẹmọ Phillip to jẹ ọkọ Mary Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun [SEP]', 'score': 0.1738305538892746,
'token': 12176,
'token_str': 'Mary'},
{'sequence': '[CLS] Arẹmọ Phillip to jẹ ọkọ Queen Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun [SEP]', 'score': 0.16382873058319092,
'token': 13704,
'token_str': 'Queen'},
{'sequence': '[CLS] Arẹmọ Phillip to jẹ ọkọ ti Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun [SEP]', 'score': 0.13272495567798615,
'token': 14382,
'token_str': 'ti'},
{'sequence': '[CLS] Arẹmọ Phillip to jẹ ọkọ King Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun [SEP]', 'score': 0.12823280692100525,
'token': 11515,
'token_str': 'King'},
{'sequence': '[CLS] Arẹmọ Phillip to jẹ ọkọ Lady Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun [SEP]', 'score': 0.07841219753026962,
'token': 14005,
'token_str': 'Lady'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on Bible, JW300, [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt), [Yoruba Embedding corpus](https://huggingface.co/datasets/yoruba_text_c3) and [CC-Aligned](https://opus.nlpl.eu/), Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| mBERT F1 | yo_bert F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 78.97 | 82.58
[BBC Yorùbá Textclass](https://huggingface.co/datasets/yoruba_bbc_topics) | 75.13 | 79.11
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/bert-base-multilingual-cased-finetuned-yoruba | null | [
"transformers",
"pytorch",
"tf",
"jax",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: yo
datasets:
---
bert-base-multilingual-cased-finetuned-yoruba
=============================================
Model description
-----------------
bert-base-multilingual-cased-finetuned-yoruba is a Yoruba BERT model obtained by fine-tuning bert-base-multilingual-cased model on Yorùbá language texts. It provides better performance than the multilingual BERT on text classification and named entity recognition datasets.
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on Yorùbá corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, mBERT F1: 78.97, yo\_bert F1: 82.58
Dataset: BBC Yorùbá Textclass, mBERT F1: 75.13, yo\_bert F1: 79.11
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 78.97, yo\\_bert F1: 82.58\nDataset: BBC Yorùbá Textclass, mBERT F1: 75.13, yo\\_bert F1: 79.11",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 78.97, yo\\_bert F1: 82.58\nDataset: BBC Yorùbá Textclass, mBERT F1: 75.13, yo\\_bert F1: 79.11",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
33,
22,
284,
15
] | [
"TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, mBERT F1: 78.97, yo\\_bert F1: 82.58\nDataset: BBC Yorùbá Textclass, mBERT F1: 75.13, yo\\_bert F1: 79.11### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ha
- ig
- rw
- lg
- luo
- pcm
- sw
- wo
- yo
- multilingual
datasets:
- masakhaner
---
# bert-base-multilingual-cased-masakhaner
## Model description
**bert-base-multilingual-cased-masakhaner** is the first **Named Entity Recognition** model for 9 African languages (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) based on a fine-tuned mBERT base model. It achieves the **state-of-the-art performance** for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/bert-base-multilingual-cased-masakhaner")
model = AutoModelForTokenClassification.from_pretrained("Davlan/bert-base-multilingual-cased-masakhaner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-DATE |Beginning of a DATE entity right after another DATE entity
I-DATE |DATE entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the [original MasakhaNER paper](https://arxiv.org/abs/2103.11811) which trained & evaluated the model on MasakhaNER corpus.
## Eval results on Test set (F-score)
language|F1-score
-|-
hau |88.66
ibo |85.72
kin |71.94
lug |81.73
luo |77.39
pcm |88.96
swa |88.23
wol |66.27
yor |80.09
### BibTeX entry and citation info
```
@article{adelani21tacl,
title = {Masakha{NER}: Named Entity Recognition for African Languages},
author = {David Ifeoluwa Adelani and Jade Abbott and Graham Neubig and Daniel D'souza and Julia Kreutzer and Constantine Lignos and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and Israel Abebe Azime and Shamsuddeen Muhammad and Chris Chinenye Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and Jesujoba Alabi and Seid Muhie Yimam and Tajuddeen Gwadabe and Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and Verrah Otiende and Iroro Orife and Davis David and Samba Ngom and Tosin Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and Chiamaka Chukwuneke and Nkiruka Odu and Eric Peter Wairagala and Samuel Oyerinde and Clemencia Siro and Tobius Saul Bateesa and Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and Ayodele Awokoya and Mouhamadane MBOUP and Dibora Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and Thierno Ibrahima DIOP and Abdoulaye Diallo and Adewale Akinfaderin and Tendai Marengereke and Salomey Osei},
journal = {Transactions of the Association for Computational Linguistics (TACL)},
month = {},
url = {https://arxiv.org/abs/2103.11811},
year = {2021}
}
```
| {} | Davlan/bert-base-multilingual-cased-masakhaner | null | [
"transformers",
"pytorch",
"tf",
"bert",
"token-classification",
"arxiv:2103.11811",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.11811"
] | [] | TAGS
#transformers #pytorch #tf #bert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* ha
* ig
* rw
* lg
* luo
* pcm
* sw
* wo
* yo
* multilingual
datasets:
* masakhaner
---
bert-base-multilingual-cased-masakhaner
=======================================
Model description
-----------------
bert-base-multilingual-cased-masakhaner is the first Named Entity Recognition model for 9 African languages (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) based on a fine-tuned mBERT base model. It achieves the state-of-the-art performance for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane MasakhaNER dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.
Eval results on Test set (F-score)
----------------------------------
### BibTeX entry and citation info
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #bert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
41,
21,
282,
10
] | [
"TAGS\n#transformers #pytorch #tf #bert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------### BibTeX entry and citation info"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ar
- de
- en
- es
- fr
- it
- lv
- nl
- pt
- zh
- multilingual
---
# bert-base-multilingual-cased-ner-hrl
## Model description
**bert-base-multilingual-cased-ner-hrl** is a **Named Entity Recognition** model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned mBERT base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on an aggregation of 10 high-resourced languages
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/bert-base-multilingual-cased-ner-hrl")
model = AutoModelForTokenClassification.from_pretrained("Davlan/bert-base-multilingual-cased-ner-hrl")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Nader Jokhadar had given Syria the lead with a well-struck header in the seventh minute."
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
The training data for the 10 languages are from:
Language|Dataset
-|-
Arabic | [ANERcorp](https://camel.abudhabi.nyu.edu/anercorp/)
German | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
English | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
Spanish | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
French | [Europeana Newspapers](https://github.com/EuropeanaNewspapers/ner-corpora/tree/master/enp_FR.bnf.bio)
Italian | [Italian I-CAB](https://ontotext.fbk.eu/icab.html)
Latvian | [Latvian NER](https://github.com/LUMII-AILab/FullStack/tree/master/NamedEntities)
Dutch | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
Portuguese |[Paramopama + Second Harem](https://github.com/davidsbatista/NER-datasets/tree/master/Portuguese)
Chinese | [MSRA](https://huggingface.co/datasets/msra_ner)
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code. | {"license": "afl-3.0"} | Davlan/bert-base-multilingual-cased-ner-hrl | null | [
"transformers",
"pytorch",
"tf",
"onnx",
"bert",
"token-classification",
"license:afl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #onnx #bert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
-------------------
language:
* ar
* de
* en
* es
* fr
* it
* lv
* nl
* pt
* zh
* multilingual
---
bert-base-multilingual-cased-ner-hrl
====================================
Model description
-----------------
bert-base-multilingual-cased-ner-hrl is a Named Entity Recognition model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned mBERT base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *bert-base-multilingual-cased* model that was fine-tuned on an aggregation of 10 high-resourced languages
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
The training data for the 10 languages are from:
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code.
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
"TAGS\n#transformers #pytorch #tf #onnx #bert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
46,
21,
170
] | [
"TAGS\n#transformers #pytorch #tf #onnx #bert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# byt5-base-eng-yor-mt
## Model description
**byt5-base-eng-yor-mt** is a **machine translation** model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a **strong baseline** for automatically translating texts from English to Yorùbá.
Specifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning byt5-base achieves **12.23 BLEU** on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/byt5-base-eng-yor-mt | null | [
"transformers",
"pytorch",
"safetensors",
"t5",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #safetensors #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# byt5-base-eng-yor-mt
## Model description
byt5-base-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá.
Specifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning byt5-base achieves 12.23 BLEU on Menyo-20k test set while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
| [
"# byt5-base-eng-yor-mt",
"## Model description\nbyt5-base-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 12.23 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #safetensors #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n",
"# byt5-base-eng-yor-mt",
"## Model description\nbyt5-base-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 12.23 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
56,
13,
92,
31,
25,
16,
48,
15
] | [
"TAGS\n#transformers #pytorch #safetensors #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# byt5-base-eng-yor-mt## Model description\nbyt5-base-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 12.23 BLEU on Menyo-20k test set while mt5-base achieves 9.82### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# byt5-base-yor-eng-mt
## Model description
**byt5-base-yor-eng-mt** is a **machine translation** model from Yorùbá language to English language based on a fine-tuned byt5-base model. It establishes a **strong baseline** for automatically translating texts from Yorùbá to English.
Specifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning byt5-base achieves 14.05 BLEU on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/byt5-base-yor-eng-mt | null | [
"transformers",
"pytorch",
"t5",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# byt5-base-yor-eng-mt
## Model description
byt5-base-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English.
Specifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning byt5-base achieves 14.05 BLEU on Menyo-20k test set while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
| [
"# byt5-base-yor-eng-mt",
"## Model description\nbyt5-base-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 14.05 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# byt5-base-yor-eng-mt",
"## Model description\nbyt5-base-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 14.05 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
48,
13,
92,
31,
25,
16,
48,
15
] | [
"TAGS\n#transformers #pytorch #t5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# byt5-base-yor-eng-mt## Model description\nbyt5-base-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned byt5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *byt5-base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning byt5-base achieves 14.05 BLEU on Menyo-20k test set while mt5-base achieves 15.57### BibTeX entry and citation info\nBy David Adelani"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ha
- ig
- rw
- lg
- luo
- pcm
- sw
- wo
- yo
- multilingual
datasets:
- masakhaner
---
# bert-base-multilingual-cased-masakhaner
## Model description
**distilbert-base-multilingual-cased-masakhaner** is the first **Named Entity Recognition** model for 9 African languages (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) based on a fine-tuned BERT base model. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *distilbert-base-multilingual-cased* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/distilbert-base-multilingual-cased-masakhaner")
model = AutoModelForTokenClassification.from_pretrained("Davlan/distilbert-base-multilingual-cased-masakhaner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-DATE |Beginning of a DATE entity right after another DATE entity
I-DATE |DATE entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the [original MasakhaNER paper](https://arxiv.org/abs/2103.11811) which trained & evaluated the model on MasakhaNER corpus.
## Eval results on Test set (F-score)
language|F1-score
-|-
hau |88.88
ibo |84.87
kin |74.19
lug |78.43
luo |73.32
pcm |87.98
swa |86.20
wol |64.67
yor |78.10
### BibTeX entry and citation info
```
@article{adelani21tacl,
title = {Masakha{NER}: Named Entity Recognition for African Languages},
author = {David Ifeoluwa Adelani and Jade Abbott and Graham Neubig and Daniel D'souza and Julia Kreutzer and Constantine Lignos and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and Israel Abebe Azime and Shamsuddeen Muhammad and Chris Chinenye Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and Jesujoba Alabi and Seid Muhie Yimam and Tajuddeen Gwadabe and Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and Verrah Otiende and Iroro Orife and Davis David and Samba Ngom and Tosin Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and Chiamaka Chukwuneke and Nkiruka Odu and Eric Peter Wairagala and Samuel Oyerinde and Clemencia Siro and Tobius Saul Bateesa and Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and Ayodele Awokoya and Mouhamadane MBOUP and Dibora Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and Thierno Ibrahima DIOP and Abdoulaye Diallo and Adewale Akinfaderin and Tendai Marengereke and Salomey Osei},
journal = {Transactions of the Association for Computational Linguistics (TACL)},
month = {},
url = {https://arxiv.org/abs/2103.11811},
year = {2021}
}
```
| {} | Davlan/distilbert-base-multilingual-cased-masakhaner | null | [
"transformers",
"pytorch",
"tf",
"distilbert",
"token-classification",
"arxiv:2103.11811",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.11811"
] | [] | TAGS
#transformers #pytorch #tf #distilbert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* ha
* ig
* rw
* lg
* luo
* pcm
* sw
* wo
* yo
* multilingual
datasets:
* masakhaner
---
bert-base-multilingual-cased-masakhaner
=======================================
Model description
-----------------
distilbert-base-multilingual-cased-masakhaner is the first Named Entity Recognition model for 9 African languages (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) based on a fine-tuned BERT base model. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *distilbert-base-multilingual-cased* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane MasakhaNER dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.
Eval results on Test set (F-score)
----------------------------------
### BibTeX entry and citation info
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #distilbert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
43,
21,
282,
10
] | [
"TAGS\n#transformers #pytorch #tf #distilbert #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 9 African NER datasets (Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahilu, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------### BibTeX entry and citation info"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ar
- de
- en
- es
- fr
- it
- lv
- nl
- pt
- zh
- multilingual
---
# distilbert-base-multilingual-cased-ner-hrl
## Model description
**distilbert-base-multilingual-cased-ner-hrl** is a **Named Entity Recognition** model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned Distiled BERT base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *distilbert-base-multilingual-cased* model that was fine-tuned on an aggregation of 10 high-resourced languages
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/distilbert-base-multilingual-cased-ner-hrl")
model = AutoModelForTokenClassification.from_pretrained("Davlan/distilbert-base-multilingual-cased-ner-hrl")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Nader Jokhadar had given Syria the lead with a well-struck header in the seventh minute."
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
The training data for the 10 languages are from:
Language|Dataset
-|-
Arabic | [ANERcorp](https://camel.abudhabi.nyu.edu/anercorp/)
German | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
English | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
Spanish | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
French | [Europeana Newspapers](https://github.com/EuropeanaNewspapers/ner-corpora/tree/master/enp_FR.bnf.bio)
Italian | [Italian I-CAB](https://ontotext.fbk.eu/icab.html)
Latvian | [Latvian NER](https://github.com/LUMII-AILab/FullStack/tree/master/NamedEntities)
Dutch | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
Portuguese |[Paramopama + Second Harem](https://github.com/davidsbatista/NER-datasets/tree/master/Portuguese)
Chinese | [MSRA](https://huggingface.co/datasets/msra_ner)
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code. | {"license": "afl-3.0"} | Davlan/distilbert-base-multilingual-cased-ner-hrl | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"distilbert",
"token-classification",
"license:afl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #safetensors #distilbert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
-------------------
language:
* ar
* de
* en
* es
* fr
* it
* lv
* nl
* pt
* zh
* multilingual
---
distilbert-base-multilingual-cased-ner-hrl
==========================================
Model description
-----------------
distilbert-base-multilingual-cased-ner-hrl is a Named Entity Recognition model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned Distiled BERT base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *distilbert-base-multilingual-cased* model that was fine-tuned on an aggregation of 10 high-resourced languages
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
The training data for the 10 languages are from:
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code.
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #distilbert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
49,
21,
170
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #distilbert #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# m2m100_418M-eng-yor-mt
## Model description
**m2m100_418M-eng-yor-mt** is a **machine translation** model from English language to Yorùbá language based on a fine-tuned facebook/m2m100_418M model. It establishes a **strong baseline** for automatically translating texts from English to Yorùbá.
Specifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt).
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning m2m100_418M achieves **13.39 BLEU** on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/m2m100_418M-eng-yor-mt | null | [
"transformers",
"pytorch",
"m2m_100",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# m2m100_418M-eng-yor-mt
## Model description
m2m100_418M-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from English to Yorùbá.
Specifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning m2m100_418M achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
| [
"# m2m100_418M-eng-yor-mt",
"## Model description\nm2m100_418M-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us \n",
"# m2m100_418M-eng-yor-mt",
"## Model description\nm2m100_418M-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
44,
15,
103,
31,
24,
16,
50,
15
] | [
"TAGS\n#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us \n# m2m100_418M-eng-yor-mt## Model description\nm2m100_418M-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# m2m100_418M-eng-yor-mt
## Model description
**m2m100_418M-yor-eng-mt** is a **machine translation** model from Yorùbá language to English language based on a fine-tuned facebook/m2m100_418M model. It establishes a **strong baseline** for automatically translating texts from Yorùbá to English.
Specifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt).
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning m2m100_418M achieves **16.76 BLEU** on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/m2m100_418M-yor-eng-mt | null | [
"transformers",
"pytorch",
"m2m_100",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# m2m100_418M-eng-yor-mt
## Model description
m2m100_418M-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from Yorùbá to English.
Specifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning m2m100_418M achieves 16.76 BLEU on Menyo-20k test set while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
| [
"# m2m100_418M-eng-yor-mt",
"## Model description\nm2m100_418M-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 16.76 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# m2m100_418M-eng-yor-mt",
"## Model description\nm2m100_418M-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 16.76 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
48,
15,
103,
31,
24,
16,
50,
15
] | [
"TAGS\n#transformers #pytorch #m2m_100 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# m2m100_418M-eng-yor-mt## Model description\nm2m100_418M-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/m2m100_418M model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *facebook/m2m100_418M* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k.#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning m2m100_418M achieves 16.76 BLEU on Menyo-20k test set while mt5-base achieves 15.57### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language: yo
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# mT5_base_yoruba_adr
## Model description
**mT5_base_yoruba_adr** is a **automatic diacritics restoration** model for Yorùbá language based on a fine-tuned mT5-base model. It achieves the **state-of-the-art performance** for adding the correct diacritics or tonal marks to Yorùbá texts.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for ADR.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("")
model = AutoModelForTokenClassification.from_pretrained("")
nlp = pipeline("", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
64.63 BLEU on [Global Voices test set](https://arxiv.org/abs/2003.10564)
70.27 BLEU on [Menyo-20k test set](https://arxiv.org/abs/2103.08647)
### BibTeX entry and citation info
By Jesujoba Alabi and David Adelani
```
```
| {} | Davlan/mT5_base_yoruba_adr | null | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"arxiv:2003.10564",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2003.10564",
"2103.08647"
] | [] | TAGS
#transformers #pytorch #mt5 #text2text-generation #arxiv-2003.10564 #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Hugging Face's logo
---
language: yo
datasets:
- JW300 + Menyo-20k
---
# mT5_base_yoruba_adr
## Model description
mT5_base_yoruba_adr is a automatic diacritics restoration model for Yorùbá language based on a fine-tuned mT5-base model. It achieves the state-of-the-art performance for adding the correct diacritics or tonal marks to Yorùbá texts.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for ADR.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
64.63 BLEU on Global Voices test set
70.27 BLEU on Menyo-20k test set
### BibTeX entry and citation info
By Jesujoba Alabi and David Adelani
| [
"# mT5_base_yoruba_adr",
"## Model description\nmT5_base_yoruba_adr is a automatic diacritics restoration model for Yorùbá language based on a fine-tuned mT5-base model. It achieves the state-of-the-art performance for adding the correct diacritics or tonal marks to Yorùbá texts. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for ADR.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n64.63 BLEU on Global Voices test set\n70.27 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy Jesujoba Alabi and David Adelani"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2003.10564 #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# mT5_base_yoruba_adr",
"## Model description\nmT5_base_yoruba_adr is a automatic diacritics restoration model for Yorùbá language based on a fine-tuned mT5-base model. It achieves the state-of-the-art performance for adding the correct diacritics or tonal marks to Yorùbá texts. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for ADR.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n64.63 BLEU on Global Voices test set\n70.27 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy Jesujoba Alabi and David Adelani"
] | [
57,
12,
104,
6,
21,
45,
28,
18,
39,
22
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2003.10564 #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mT5_base_yoruba_adr## Model description\nmT5_base_yoruba_adr is a automatic diacritics restoration model for Yorùbá language based on a fine-tuned mT5-base model. It achieves the state-of-the-art performance for adding the correct diacritics or tonal marks to Yorùbá texts. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k## Intended uses & limitations#### How to use\nYou can use this model with Transformers *pipeline* for ADR.#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset## Training procedure\nThis model was trained on a single NVIDIA V100 GPU## Eval results on Test set (BLEU score)\n64.63 BLEU on Global Voices test set\n70.27 BLEU on Menyo-20k test set### BibTeX entry and citation info\nBy Jesujoba Alabi and David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# mbart50-large-eng-yor-mt
## Model description
**mbart50-large-eng-yor-mt** is a **machine translation** model from English language to Yorùbá language based on a fine-tuned facebook/mbart-large-50 model. It establishes a **strong baseline** for automatically translating texts from English to Yorùbá.
Specifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt). The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning mbarr50-large achieves **13.39 BLEU** on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/mbart50-large-eng-yor-mt | null | [
"transformers",
"pytorch",
"mbart",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# mbart50-large-eng-yor-mt
## Model description
mbart50-large-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from English to Yorùbá.
Specifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning mbarr50-large achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82
### BibTeX entry and citation info
By David Adelani
| [
"# mbart50-large-eng-yor-mt",
"## Model description\nmbart50-large-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning mbarr50-large achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# mbart50-large-eng-yor-mt",
"## Model description\nmbart50-large-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning mbarr50-large achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
46,
13,
145,
31,
25,
16,
48,
15
] | [
"TAGS\n#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# mbart50-large-eng-yor-mt## Model description\nmbart50-large-eng-yor-mt is a machine translation model from English language to Yorùbá language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning mbarr50-large achieves 13.39 BLEU on Menyo-20k test set while mt5-base achieves 9.82### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# mbart50-large-yor-eng-mt
## Model description
**mbart50-large-yor-eng-mt** is a **machine translation** model from Yorùbá language to English language based on a fine-tuned facebook/mbart-large-50 model. It establishes a **strong baseline** for automatically translating texts from Yorùbá to English.
Specifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt). The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning mbart50-large achieves **15.88 BLEU** on [Menyo-20k test set](https://arxiv.org/abs/2103.08647) while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/mbart50-large-yor-eng-mt | null | [
"transformers",
"pytorch",
"mbart",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# mbart50-large-yor-eng-mt
## Model description
mbart50-large-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from Yorùbá to English.
Specifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.
#### Limitations and bias
This model is limited by its training dataset. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
Fine-tuning mbart50-large achieves 15.88 BLEU on Menyo-20k test set while mt5-base achieves 15.57
### BibTeX entry and citation info
By David Adelani
| [
"# mbart50-large-yor-eng-mt",
"## Model description\nmbart50-large-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning mbart50-large achieves 15.88 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us \n",
"# mbart50-large-yor-eng-mt",
"## Model description\nmbart50-large-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.",
"#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\nFine-tuning mbart50-large achieves 15.88 BLEU on Menyo-20k test set while mt5-base achieves 15.57",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
42,
13,
145,
31,
25,
16,
48,
15
] | [
"TAGS\n#transformers #pytorch #mbart #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #region-us \n# mbart50-large-yor-eng-mt## Model description\nmbart50-large-yor-eng-mt is a machine translation model from Yorùbá language to English language based on a fine-tuned facebook/mbart-large-50 model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mbart-large-50* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k. The model was trained using Swahili(sw_KE) as the language since the pre-trained model does not initially support Yorùbá. Thus, you need to use the sw_KE for language code when evaluating the model.#### Limitations and bias\nThis model is limited by its training dataset. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on NVIDIA V100 GPU## Eval results on Test set (BLEU score)\nFine-tuning mbart50-large achieves 15.88 BLEU on Menyo-20k test set while mt5-base achieves 15.57### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# mT5_base_eng_yor_mt
## Model description
**mT5_base_yor_eng_mt** is a **machine translation** model from English language to Yorùbá language based on a fine-tuned mT5-base model. It establishes a **strong baseline** for automatically translating texts from English to Yorùbá.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for MT.
```python
from transformers import MT5ForConditionalGeneration, T5Tokenizer
model = MT5ForConditionalGeneration.from_pretrained("Davlan/mt5_base_eng_yor_mt")
tokenizer = T5Tokenizer.from_pretrained("google/mt5-base")
input_string = "Where are you?"
inputs = tokenizer.encode(input_string, return_tensors="pt")
generated_tokens = model.generate(inputs)
results = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
print(results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
9.82 BLEU on [Menyo-20k test set](https://arxiv.org/abs/2103.08647)
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/mt5_base_eng_yor_mt | null | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# mT5_base_eng_yor_mt
## Model description
mT5_base_yor_eng_mt is a machine translation model from English language to Yorùbá language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for MT.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 corpus and Menyo-20k dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
9.82 BLEU on Menyo-20k test set
### BibTeX entry and citation info
By David Adelani
| [
"# mT5_base_eng_yor_mt",
"## Model description\nmT5_base_yor_eng_mt is a machine translation model from English language to Yorùbá language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for MT.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n9.82 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# mT5_base_eng_yor_mt",
"## Model description\nmT5_base_yor_eng_mt is a machine translation model from English language to Yorùbá language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for MT.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n9.82 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
48,
12,
89,
6,
20,
45,
25,
18,
28,
15
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mT5_base_eng_yor_mt## Model description\nmT5_base_yor_eng_mt is a machine translation model from English language to Yorùbá language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from English to Yorùbá. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k## Intended uses & limitations#### How to use\nYou can use this model with Transformers *pipeline* for MT.#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 corpus and Menyo-20k dataset## Training procedure\nThis model was trained on a single NVIDIA V100 GPU## Eval results on Test set (BLEU score)\n9.82 BLEU on Menyo-20k test set### BibTeX entry and citation info\nBy David Adelani"
] |
text2text-generation | transformers | Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
---
# mT5_base_yor_eng_mt
## Model description
**mT5_base_yor_eng_mt** is a **machine translation** model from Yorùbá language to English language based on a fine-tuned mT5-base model. It establishes a **strong baseline** for automatically translating texts from Yorùbá to English.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt)
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for MT.
```python
from transformers import MT5ForConditionalGeneration, T5Tokenizer
model = MT5ForConditionalGeneration.from_pretrained("Davlan/mt5_base_yor_eng_mt")
tokenizer = T5Tokenizer.from_pretrained("google/mt5-base")
input_string = "Akọni ajìjàgbara obìnrin tó sun àtìmalé torí owó orí"
inputs = tokenizer.encode(input_string, return_tensors="pt")
generated_tokens = model.generate(inputs)
results = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
print(results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 Yorùbá corpus and [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt) dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
15.57 BLEU on [Menyo-20k test set](https://arxiv.org/abs/2103.08647)
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/mt5_base_yor_eng_mt | null | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"arxiv:2103.08647",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.08647"
] | [] | TAGS
#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Hugging Face's logo
---
language:
- yo
- en
datasets:
- JW300 + Menyo-20k
---
# mT5_base_yor_eng_mt
## Model description
mT5_base_yor_eng_mt is a machine translation model from Yorùbá language to English language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English.
Specifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for MT.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (BLEU score)
15.57 BLEU on Menyo-20k test set
### BibTeX entry and citation info
By David Adelani
| [
"# mT5_base_yor_eng_mt",
"## Model description\nmT5_base_yor_eng_mt is a machine translation model from Yorùbá language to English language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for MT.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n15.57 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# mT5_base_yor_eng_mt",
"## Model description\nmT5_base_yor_eng_mt is a machine translation model from Yorùbá language to English language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k",
"## Intended uses & limitations",
"#### How to use\nYou can use this model with Transformers *pipeline* for MT.",
"#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.",
"## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset",
"## Training procedure\nThis model was trained on a single NVIDIA V100 GPU",
"## Eval results on Test set (BLEU score)\n15.57 BLEU on Menyo-20k test set",
"### BibTeX entry and citation info\nBy David Adelani"
] | [
48,
12,
89,
6,
20,
45,
28,
18,
28,
15
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #arxiv-2103.08647 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mT5_base_yor_eng_mt## Model description\nmT5_base_yor_eng_mt is a machine translation model from Yorùbá language to English language based on a fine-tuned mT5-base model. It establishes a strong baseline for automatically translating texts from Yorùbá to English. \n\nSpecifically, this model is a *mT5_base* model that was fine-tuned on JW300 Yorùbá corpus and Menyo-20k## Intended uses & limitations#### How to use\nYou can use this model with Transformers *pipeline* for MT.#### Limitations and bias\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.## Training data\nThis model was fine-tuned on on JW300 Yorùbá corpus and Menyo-20k dataset## Training procedure\nThis model was trained on a single NVIDIA V100 GPU## Eval results on Test set (BLEU score)\n15.57 BLEU on Menyo-20k test set### BibTeX entry and citation info\nBy David Adelani"
] |
text-classification | transformers | Hugging Face's logo
---
language:
- hau
- ibo
- pcm
- yor
- multilingual
---
# naija-twitter-sentiment-afriberta-large
## Model description
**naija-twitter-sentiment-afriberta-large** is the first multilingual twitter **sentiment classification** model for four (4) Nigerian languages (Hausa, Igbo, Nigerian Pidgin, and Yorùbá) based on a fine-tuned castorini/afriberta_large large model.
It achieves the **state-of-the-art performance** for the twitter sentiment classification task trained on the [NaijaSenti corpus](https://github.com/hausanlp/NaijaSenti).
The model has been trained to classify tweets into 3 sentiment classes: negative, neutral and positive
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of 4 Nigerian language datasets obtained from [NaijaSenti](https://github.com/hausanlp/NaijaSenti) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers for Sentiment Classification.
```python
from transformers import AutoModelForSequenceClassification
from transformers import AutoTokenizer
import numpy as np
from scipy.special import softmax
MODEL = "Davlan/naija-twitter-sentiment-afriberta-large"
tokenizer = AutoTokenizer.from_pretrained(MODEL)
# PT
model = AutoModelForSequenceClassification.from_pretrained(MODEL)
text = "I like you"
encoded_input = tokenizer(text, return_tensors='pt')
output = model(**encoded_input)
scores = output[0][0].detach().numpy()
scores = softmax(scores)
id2label = {0:"positive", 1:"neutral", 2:"negative"}
ranking = np.argsort(scores)
ranking = ranking[::-1]
for i in range(scores.shape[0]):
l = id2label[ranking[i]]
s = scores[ranking[i]]
print(f"{i+1}) {l} {np.round(float(s), 4)}")
```
#### Limitations and bias
This model is limited by its training dataset and domain i.e Twitter. This may not generalize well for all use cases in different domains.
## Training procedure
This model was trained on a single Nvidia RTX 2080 GPU with recommended hyperparameters from the [original NaijaSenti paper](https://arxiv.org/abs/2201.08277).
## Eval results on Test set (F-score), average over 5 runs.
language|F1-score
-|-
hau |81.2
ibo |80.8
pcm |74.5
yor |80.4
### BibTeX entry and citation info
```
@inproceedings{Muhammad2022NaijaSentiAN,
title={NaijaSenti: A Nigerian Twitter Sentiment Corpus for Multilingual Sentiment Analysis},
author={Shamsuddeen Hassan Muhammad and David Ifeoluwa Adelani and Sebastian Ruder and Ibrahim Said Ahmad and Idris Abdulmumin and Bello Shehu Bello and Monojit Choudhury and Chris C. Emezue and Saheed Salahudeen Abdullahi and Anuoluwapo Aremu and Alipio Jeorge and Pavel B. Brazdil},
year={2022}
}
```
| {} | Davlan/naija-twitter-sentiment-afriberta-large | null | [
"transformers",
"pytorch",
"tf",
"xlm-roberta",
"text-classification",
"arxiv:2201.08277",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2201.08277"
] | [] | TAGS
#transformers #pytorch #tf #xlm-roberta #text-classification #arxiv-2201.08277 #autotrain_compatible #endpoints_compatible #has_space #region-us
| Hugging Face's logo
-------------------
language:
* hau
* ibo
* pcm
* yor
* multilingual
---
naija-twitter-sentiment-afriberta-large
=======================================
Model description
-----------------
naija-twitter-sentiment-afriberta-large is the first multilingual twitter sentiment classification model for four (4) Nigerian languages (Hausa, Igbo, Nigerian Pidgin, and Yorùbá) based on a fine-tuned castorini/afriberta\_large large model.
It achieves the state-of-the-art performance for the twitter sentiment classification task trained on the NaijaSenti corpus.
The model has been trained to classify tweets into 3 sentiment classes: negative, neutral and positive
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of 4 Nigerian language datasets obtained from NaijaSenti dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers for Sentiment Classification.
#### Limitations and bias
This model is limited by its training dataset and domain i.e Twitter. This may not generalize well for all use cases in different domains.
Training procedure
------------------
This model was trained on a single Nvidia RTX 2080 GPU with recommended hyperparameters from the original NaijaSenti paper.
Eval results on Test set (F-score), average over 5 runs.
--------------------------------------------------------
### BibTeX entry and citation info
| [
"#### How to use\n\n\nYou can use this model with Transformers for Sentiment Classification.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset and domain i.e Twitter. This may not generalize well for all use cases in different domains.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single Nvidia RTX 2080 GPU with recommended hyperparameters from the original NaijaSenti paper.\n\n\nEval results on Test set (F-score), average over 5 runs.\n--------------------------------------------------------",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #xlm-roberta #text-classification #arxiv-2201.08277 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers for Sentiment Classification.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset and domain i.e Twitter. This may not generalize well for all use cases in different domains.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single Nvidia RTX 2080 GPU with recommended hyperparameters from the original NaijaSenti paper.\n\n\nEval results on Test set (F-score), average over 5 runs.\n--------------------------------------------------------",
"### BibTeX entry and citation info"
] | [
49,
18,
161,
10
] | [
"TAGS\n#transformers #pytorch #tf #xlm-roberta #text-classification #arxiv-2201.08277 #autotrain_compatible #endpoints_compatible #has_space #region-us \n#### How to use\n\n\nYou can use this model with Transformers for Sentiment Classification.#### Limitations and bias\n\n\nThis model is limited by its training dataset and domain i.e Twitter. This may not generalize well for all use cases in different domains.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single Nvidia RTX 2080 GPU with recommended hyperparameters from the original NaijaSenti paper.\n\n\nEval results on Test set (F-score), average over 5 runs.\n--------------------------------------------------------### BibTeX entry and citation info"
] |
fill-mask | transformers | Hugging Face's logo
---
language: am
datasets:
---
# xlm-roberta-base-finetuned-amharic
## Model description
**xlm-roberta-base-finetuned-amharic** is a **Amharic RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Amharic language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Amharic corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-hausa')
>>> unmasker("የአሜሪካ የአፍሪካ ቀንድ ልዩ መልዕክተኛ ጄፈሪ ፌልትማን በአራት አገራት የሚያደጉትን <mask> መጀመራቸውን የአሜሪካ የውጪ ጉዳይ ሚንስቴር አስታወቀ።")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Amharic CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | am_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 70.96 | 77.97
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-amharic | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: am
datasets:
---
xlm-roberta-base-finetuned-amharic
==================================
Model description
-----------------
xlm-roberta-base-finetuned-amharic is a Amharic RoBERTa model obtained by fine-tuning xlm-roberta-base model on Amharic language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Amharic corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Amharic CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 70.96, am\_roberta F1: 77.97
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 70.96, am\\_roberta F1: 77.97",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 70.96, am\\_roberta F1: 77.97",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
205,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Amharic CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 70.96, am\\_roberta F1: 77.97### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: ha
datasets:
---
# xlm-roberta-base-finetuned-hausa
## Model description
**xlm-roberta-base-finetuned-hausa** is a **Hausa RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Hausa language texts. It provides **better performance** than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Hausa corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-hausa')
>>> unmasker("Shugaban <mask> Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci")
[{'sequence': '<s> Shugaban kasa Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci</s>',
'score': 0.8104371428489685,
'token': 29762,
'token_str': '▁kasa'},
{'sequence': '<s> Shugaban Najeriya Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci</s>', 'score': 0.17371904850006104,
'token': 49173,
'token_str': '▁Najeriya'},
{'sequence': '<s> Shugaban kasar Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci</s>', 'score': 0.006917025428265333,
'token': 21221,
'token_str': '▁kasar'},
{'sequence': '<s> Shugaban Nigeria Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci</s>', 'score': 0.005785710643976927,
'token': 72620,
'token_str': '▁Nigeria'},
{'sequence': '<s> Shugaban Kasar Muhammadu Buhari ya amince da shawarar da ma’aikatar sufuri karkashin jagoranci</s>', 'score': 0.0010596115607768297,
'token': 170255,
'token_str': '▁Kasar'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Hausa CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | ha_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 86.10 | 91.47
[VOA Hausa Textclass](https://huggingface.co/datasets/hausa_voa_topics) | |
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-hausa | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: ha
datasets:
---
xlm-roberta-base-finetuned-hausa
================================
Model description
-----------------
xlm-roberta-base-finetuned-hausa is a Hausa RoBERTa model obtained by fine-tuning xlm-roberta-base model on Hausa language texts. It provides better performance than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Hausa corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Hausa CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 86.10, ha\_roberta F1: 91.47
Dataset: VOA Hausa Textclass, XLM-R F1: , ha\_roberta F1:
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 86.10, ha\\_roberta F1: 91.47\nDataset: VOA Hausa Textclass, XLM-R F1: , ha\\_roberta F1:",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 86.10, ha\\_roberta F1: 91.47\nDataset: VOA Hausa Textclass, XLM-R F1: , ha\\_roberta F1:",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
227,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Hausa CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 86.10, ha\\_roberta F1: 91.47\nDataset: VOA Hausa Textclass, XLM-R F1: , ha\\_roberta F1:### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: ig
datasets:
---
# xlm-roberta-base-finetuned-igbo
## Model description
**xlm-roberta-base-finetuned-igbo** is a **Igbo RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Hausa language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Igbo corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-igbo')
>>> unmasker("Reno Omokri na Gọọmentị <mask> enweghị ihe ha ga-eji hiwe ya bụ mmachi.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + OPUS CC-Align + [IGBO NLP Corpus](https://github.com/IgnatiusEzeani/IGBONLP) +[Igbo CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | ig_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 84.51 | 87.74
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-igbo | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: ig
datasets:
---
xlm-roberta-base-finetuned-igbo
===============================
Model description
-----------------
xlm-roberta-base-finetuned-igbo is a Igbo RoBERTa model obtained by fine-tuning xlm-roberta-base model on Hausa language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Igbo corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 84.51, ig\_roberta F1: 87.74
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 84.51, ig\\_roberta F1: 87.74",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 84.51, ig\\_roberta F1: 87.74",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
223,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + OPUS CC-Align + IGBO NLP Corpus +Igbo CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 84.51, ig\\_roberta F1: 87.74### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: rw
datasets:
---
# xlm-roberta-base-finetuned-kinyarwanda
## Model description
**xlm-roberta-base-finetuned-kinyarwanda** is a **Kinyarwanda RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Kinyarwanda language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Kinyarwanda corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-kinyarwanda')
>>> unmasker("Twabonye ko igihe mu <mask> hazaba hari ikirango abantu bakunze")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [KIRNEWS](https://github.com/Andrews2017/KINNEWS-and-KIRNEWS-Corpus) + [BBC Gahuza](https://www.bbc.com/gahuza)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | rw_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 73.22 | 77.76
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-kinyarwanda | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: rw
datasets:
---
xlm-roberta-base-finetuned-kinyarwanda
======================================
Model description
-----------------
xlm-roberta-base-finetuned-kinyarwanda is a Kinyarwanda RoBERTa model obtained by fine-tuning xlm-roberta-base model on Kinyarwanda language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Kinyarwanda corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 73.22, rw\_roberta F1: 77.76
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 73.22, rw\\_roberta F1: 77.76",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 73.22, rw\\_roberta F1: 77.76",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
213,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + KIRNEWS + BBC Gahuza\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 73.22, rw\\_roberta F1: 77.76### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: lg
datasets:
---
# xlm-roberta-base-finetuned-luganda
## Model description
**xlm-roberta-base-finetuned-luganda** is a **Luganda RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Luganda language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Luganda corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-luganda')
>>> unmasker("Ffe tulwanyisa abo abaagala okutabangula <mask>, Kimuli bwe yategeezezza.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [BUKKEDDE](https://github.com/masakhane-io/masakhane-ner/tree/main/text_by_language/luganda) +[Luganda CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | lg_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 79.69 | 84.70
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-luganda | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: lg
datasets:
---
xlm-roberta-base-finetuned-luganda
==================================
Model description
-----------------
xlm-roberta-base-finetuned-luganda is a Luganda RoBERTa model obtained by fine-tuning xlm-roberta-base model on Luganda language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Luganda corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 79.69, lg\_roberta F1: 84.70
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 79.69, lg\\_roberta F1: 84.70",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 79.69, lg\\_roberta F1: 84.70",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
216,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BUKKEDDE +Luganda CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 79.69, lg\\_roberta F1: 84.70### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: luo
datasets:
---
# xlm-roberta-base-finetuned-luo
## Model description
**xlm-roberta-base-finetuned-luo** is a **Luo RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Luo language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Luo corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-luo')
>>> unmasker("Obila ma Changamwe <mask> pedho achije angwen mag njore")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | luo_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 74.86 | 75.27
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-luo | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: luo
datasets:
---
xlm-roberta-base-finetuned-luo
==============================
Model description
-----------------
xlm-roberta-base-finetuned-luo is a Luo RoBERTa model obtained by fine-tuning xlm-roberta-base model on Luo language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Luo corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 74.86, luo\_roberta F1: 75.27
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 74.86, luo\\_roberta F1: 75.27",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 74.86, luo\\_roberta F1: 75.27",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
204,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 74.86, luo\\_roberta F1: 75.27### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: pcm
datasets:
---
# xlm-roberta-base-finetuned-naija
## Model description
**xlm-roberta-base-finetuned-naija** is a **Nigerian Pidgin RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Nigerian Pidgin language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Nigerian Pidgin corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-naija')
>>> unmasker("Another attack on ambulance happen for Koforidua in March <mask> year where robbers kill Ambulance driver")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on JW300 + [BBC Pidgin](https://www.bbc.com/pidgin)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | pcm_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 87.26 | 90.00
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-naija | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: pcm
datasets:
---
xlm-roberta-base-finetuned-naija
================================
Model description
-----------------
xlm-roberta-base-finetuned-naija is a Nigerian Pidgin RoBERTa model obtained by fine-tuning xlm-roberta-base model on Nigerian Pidgin language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Nigerian Pidgin corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on JW300 + BBC Pidgin
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 87.26, pcm\_roberta F1: 90.00
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.26, pcm\\_roberta F1: 90.00",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.26, pcm\\_roberta F1: 90.00",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
209,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on JW300 + BBC Pidgin\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.26, pcm\\_roberta F1: 90.00### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: sw
datasets:
---
# xlm-roberta-base-finetuned-swahili
## Model description
**xlm-roberta-base-finetuned-swahili** is a **Swahili RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Swahili language texts. It provides **better performance** than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Swahili corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-swahili')
>>> unmasker("Jumatatu, Bwana Kagame alielezea shirika la France24 huko <mask> kwamba hakuna uhalifu ulitendwa")
[{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Ufaransa kwamba hakuna uhalifu ulitendwa',
'score': 0.5077782273292542,
'token': 190096,
'token_str': 'Ufaransa'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Paris kwamba hakuna uhalifu ulitendwa',
'score': 0.3657738268375397,
'token': 7270,
'token_str': 'Paris'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Gabon kwamba hakuna uhalifu ulitendwa',
'score': 0.01592041552066803,
'token': 176392,
'token_str': 'Gabon'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko France kwamba hakuna uhalifu ulitendwa',
'score': 0.010881908237934113,
'token': 9942,
'token_str': 'France'},
{'sequence': 'Jumatatu, Bwana Kagame alielezea shirika la France24 huko Marseille kwamba hakuna uhalifu ulitendwa',
'score': 0.009554869495332241,
'token': 185918,
'token_str': 'Marseille'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Swahili CC-100](http://data.statmt.org/cc-100/)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | sw_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 87.55 | 89.46
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-swahili | null | [
"transformers",
"pytorch",
"safetensors",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #safetensors #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: sw
datasets:
---
xlm-roberta-base-finetuned-swahili
==================================
Model description
-----------------
xlm-roberta-base-finetuned-swahili is a Swahili RoBERTa model obtained by fine-tuning xlm-roberta-base model on Swahili language texts. It provides better performance than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Swahili corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Swahili CC-100
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 87.55, sw\_roberta F1: 89.46
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.55, sw\\_roberta F1: 89.46",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.55, sw\\_roberta F1: 89.46",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
35,
22,
205,
15
] | [
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Swahili CC-100\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 87.55, sw\\_roberta F1: 89.46### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: wo
datasets:
---
# xlm-roberta-base-finetuned-wolof
## Model description
**xlm-roberta-base-finetuned-luganda** is a **Wolof RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Wolof language texts. It provides **better performance** than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Wolof corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-wolof')
>>> unmasker("Màkki Sàll feeñal na ay xalaatam ci mbir yu am solo yu soxal <mask> ak Afrik.")
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on [Bible OT](http://biblewolof.com/) + [OPUS](https://opus.nlpl.eu/) + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | wo_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 63.86 | 68.31
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-wolof | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: wo
datasets:
---
xlm-roberta-base-finetuned-wolof
================================
Model description
-----------------
xlm-roberta-base-finetuned-luganda is a Wolof RoBERTa model obtained by fine-tuning xlm-roberta-base model on Wolof language texts. It provides better performance than the XLM-RoBERTa on named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Wolof corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 63.86, wo\_roberta F1: 68.31
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 63.86, wo\\_roberta F1: 68.31",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 63.86, wo\\_roberta F1: 68.31",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
224,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible OT + OPUS + News Corpora (Lu Defu Waxu, Saabal, and Wolof Online)\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 63.86, wo\\_roberta F1: 68.31### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
fill-mask | transformers | Hugging Face's logo
---
language: yo
datasets:
---
# xlm-roberta-base-finetuned-yoruba
## Model description
**xlm-roberta-base-finetuned-yoruba** is a **Yoruba RoBERTa** model obtained by fine-tuning **xlm-roberta-base** model on Yorùbá language texts. It provides **better performance** than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Yorùbá corpus.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='Davlan/xlm-roberta-base-finetuned-yoruba')
>>> unmasker("Arẹmọ Phillip to jẹ ọkọ <mask> Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun")
[{'sequence': '<s> Arẹmọ Phillip to jẹ ọkọ Queen Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun</s>', 'score': 0.24844281375408173,
'token': 44109,
'token_str': '▁Queen'},
{'sequence': '<s> Arẹmọ Phillip to jẹ ọkọ ile Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun</s>', 'score': 0.1665010154247284,
'token': 1350,
'token_str': '▁ile'},
{'sequence': '<s> Arẹmọ Phillip to jẹ ọkọ ti Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun</s>', 'score': 0.07604238390922546,
'token': 1053,
'token_str': '▁ti'},
{'sequence': '<s> Arẹmọ Phillip to jẹ ọkọ baba Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun</s>', 'score': 0.06353845447301865,
'token': 12878,
'token_str': '▁baba'},
{'sequence': '<s> Arẹmọ Phillip to jẹ ọkọ Oba Elizabeth to ti wa lori aisan ti dagbere faye lẹni ọdun mọkandilọgọrun</s>', 'score': 0.03836742788553238,
'token': 82879,
'token_str': '▁Oba'}]
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on Bible, JW300, [Menyo-20k](https://huggingface.co/datasets/menyo20k_mt), [Yoruba Embedding corpus](https://huggingface.co/datasets/yoruba_text_c3) and [CC-Aligned](https://opus.nlpl.eu/), Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.
## Training procedure
This model was trained on a single NVIDIA V100 GPU
## Eval results on Test set (F-score, average over 5 runs)
Dataset| XLM-R F1 | yo_roberta F1
-|-|-
[MasakhaNER](https://github.com/masakhane-io/masakhane-ner) | 77.58 | 83.66
[BBC Yorùbá Textclass](https://huggingface.co/datasets/yoruba_bbc_topics) | |
### BibTeX entry and citation info
By David Adelani
```
```
| {} | Davlan/xlm-roberta-base-finetuned-yoruba | null | [
"transformers",
"pytorch",
"xlm-roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language: yo
datasets:
---
xlm-roberta-base-finetuned-yoruba
=================================
Model description
-----------------
xlm-roberta-base-finetuned-yoruba is a Yoruba RoBERTa model obtained by fine-tuning xlm-roberta-base model on Yorùbá language texts. It provides better performance than the XLM-RoBERTa on text classification and named entity recognition datasets.
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on Yorùbá corpus.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for masked token prediction.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU
Eval results on Test set (F-score, average over 5 runs)
-------------------------------------------------------
Dataset: MasakhaNER, XLM-R F1: 77.58, yo\_roberta F1: 83.66
Dataset: BBC Yorùbá Textclass, XLM-R F1: , yo\_roberta F1:
### BibTeX entry and citation info
By David Adelani
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 77.58, yo\\_roberta F1: 83.66\nDataset: BBC Yorùbá Textclass, XLM-R F1: , yo\\_roberta F1:",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 77.58, yo\\_roberta F1: 83.66\nDataset: BBC Yorùbá Textclass, XLM-R F1: , yo\\_roberta F1:",
"### BibTeX entry and citation info\n\n\nBy David Adelani"
] | [
31,
22,
282,
15
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for masked token prediction.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on Bible, JW300, Menyo-20k, Yoruba Embedding corpus and CC-Aligned, Wikipedia, news corpora (BBC Yoruba, VON Yoruba, Asejere, Alaroye), and other small datasets curated from friends.\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU\n\n\nEval results on Test set (F-score, average over 5 runs)\n-------------------------------------------------------\n\n\nDataset: MasakhaNER, XLM-R F1: 77.58, yo\\_roberta F1: 83.66\nDataset: BBC Yorùbá Textclass, XLM-R F1: , yo\\_roberta F1:### BibTeX entry and citation info\n\n\nBy David Adelani"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- am
- ha
- ig
- rw
- lg
- luo
- pcm
- sw
- wo
- yo
- multilingual
datasets:
- masakhaner
---
# xlm-roberta-base-masakhaner
## Model description
**xlm-roberta-base-masakhaner** is the first **Named Entity Recognition** model for 10 African languages (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) based on a fine-tuned XLM-RoBERTa large model. It achieves the **state-of-the-art performance** for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/xlm-roberta-base-masakhaner")
model = AutoModelForTokenClassification.from_pretrained("Davlan/xlm-roberta-base-masakhaner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-DATE |Beginning of a DATE entity right after another DATE entity
I-DATE |DATE entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the [original MasakhaNER paper](https://arxiv.org/abs/2103.11811) which trained & evaluated the model on MasakhaNER corpus.
### BibTeX entry and citation info
```
@article{adelani21tacl,
title = {Masakha{NER}: Named Entity Recognition for African Languages},
author = {David Ifeoluwa Adelani and Jade Abbott and Graham Neubig and Daniel D'souza and Julia Kreutzer and Constantine Lignos and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and Israel Abebe Azime and Shamsuddeen Muhammad and Chris Chinenye Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and Jesujoba Alabi and Seid Muhie Yimam and Tajuddeen Gwadabe and Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and Verrah Otiende and Iroro Orife and Davis David and Samba Ngom and Tosin Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and Chiamaka Chukwuneke and Nkiruka Odu and Eric Peter Wairagala and Samuel Oyerinde and Clemencia Siro and Tobius Saul Bateesa and Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and Ayodele Awokoya and Mouhamadane MBOUP and Dibora Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and Thierno Ibrahima DIOP and Abdoulaye Diallo and Adewale Akinfaderin and Tendai Marengereke and Salomey Osei},
journal = {Transactions of the Association for Computational Linguistics (TACL)},
month = {},
url = {https://arxiv.org/abs/2103.11811},
year = {2021}
}
```
| {} | Davlan/xlm-roberta-base-masakhaner | null | [
"transformers",
"pytorch",
"xlm-roberta",
"token-classification",
"arxiv:2103.11811",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.11811"
] | [] | TAGS
#transformers #pytorch #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* am
* ha
* ig
* rw
* lg
* luo
* pcm
* sw
* wo
* yo
* multilingual
datasets:
* masakhaner
---
xlm-roberta-base-masakhaner
===========================
Model description
-----------------
xlm-roberta-base-masakhaner is the first Named Entity Recognition model for 10 African languages (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) based on a fine-tuned XLM-RoBERTa large model. It achieves the state-of-the-art performance for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane MasakhaNER dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.
### BibTeX entry and citation info
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.",
"### BibTeX entry and citation info"
] | [
41,
21,
240,
10
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.### BibTeX entry and citation info"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ar
- de
- en
- es
- fr
- it
- lv
- nl
- pt
- zh
- multilingual
---
# xlm-roberta-base-ner-hrl
## Model description
**xlm-roberta-base-ner-hrl** is a **Named Entity Recognition** model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned XLM-RoBERTa base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on an aggregation of 10 high-resourced languages
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/xlm-roberta-base-ner-hrl")
model = AutoModelForTokenClassification.from_pretrained("Davlan/xlm-roberta-base-ner-hrl")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Nader Jokhadar had given Syria the lead with a well-struck header in the seventh minute."
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
The training data for the 10 languages are from:
Language|Dataset
-|-
Arabic | [ANERcorp](https://camel.abudhabi.nyu.edu/anercorp/)
German | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
English | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
Spanish | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
French | [Europeana Newspapers](https://github.com/EuropeanaNewspapers/ner-corpora/tree/master/enp_FR.bnf.bio)
Italian | [Italian I-CAB](https://ontotext.fbk.eu/icab.html)
Latvian | [Latvian NER](https://github.com/LUMII-AILab/FullStack/tree/master/NamedEntities)
Dutch | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
Portuguese |[Paramopama + Second Harem](https://github.com/davidsbatista/NER-datasets/tree/master/Portuguese)
Chinese | [MSRA](https://huggingface.co/datasets/msra_ner)
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code. | {"license": "afl-3.0"} | Davlan/xlm-roberta-base-ner-hrl | null | [
"transformers",
"pytorch",
"safetensors",
"xlm-roberta",
"token-classification",
"license:afl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* ar
* de
* en
* es
* fr
* it
* lv
* nl
* pt
* zh
* multilingual
---
xlm-roberta-base-ner-hrl
========================
Model description
-----------------
xlm-roberta-base-ner-hrl is a Named Entity Recognition model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned XLM-RoBERTa base model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-base* model that was fine-tuned on an aggregation of 10 high-resourced languages
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
The training data for the 10 languages are from:
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code.
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
43,
21,
170
] | [
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] |
token-classification | transformers | Hugging Face's logo
---
language:
- af
- nr
- nso
- ss
- st
- tn
- ts
- ve
- xh
- zu
- multilingual
datasets:
- masakhaner
---
# xlm-roberta-base-sadilar-ner
## Model description
**xlm-roberta-base-sadilar-ner** is the first **Named Entity Recognition** model for 10 South African languages (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) based on a fine-tuned XLM-RoBERTa large model. It achieves the **state-of-the-art performance** for the NER task. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of South African languages datasets obtained from [SADILAR](https://www.sadilar.org/index.php/en/) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/xlm-roberta-base-sadilar-ner")
model = AutoModelForTokenClassification.from_pretrained("Davlan/xlm-roberta-base-sadilar-ner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Kuchaza kona ukuthi uMengameli uMnuz Cyril Ramaphosa, usebatshelile ukuthi uzosikhipha maduze isitifiketi."
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 10 African NER datasets (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) [SADILAR](https://www.sadilar.org/index.php/en/) dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
### BibTeX entry and citation info
```
| {} | Davlan/xlm-roberta-base-sadilar-ner | null | [
"transformers",
"pytorch",
"xlm-roberta",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* af
* nr
* nso
* ss
* st
* tn
* ts
* ve
* xh
* zu
* multilingual
datasets:
* masakhaner
---
xlm-roberta-base-sadilar-ner
============================
Model description
-----------------
xlm-roberta-base-sadilar-ner is the first Named Entity Recognition model for 10 South African languages (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) based on a fine-tuned XLM-RoBERTa large model. It achieves the state-of-the-art performance for the NER task. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of South African languages datasets obtained from SADILAR dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 10 African NER datasets (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) SADILAR dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
### BibTeX entry and citation info
'''
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) SADILAR dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:",
"### BibTeX entry and citation info\n\n\n'''"
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) SADILAR dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:",
"### BibTeX entry and citation info\n\n\n'''"
] | [
31,
21,
178,
13
] | [
"TAGS\n#transformers #pytorch #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Afrikaans, isiNdebele, isiXhosa, isiZulu, Sepedi, Sesotho, Setswana, siSwati, Tshivenda and Xitsonga) SADILAR dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:### BibTeX entry and citation info\n\n\n'''"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ar
- as
- bn
- ca
- en
- es
- eu
- fr
- gu
- hi
- id
- ig
- mr
- pa
- pt
- sw
- ur
- vi
- yo
- zh
- multilingual
datasets:
- wikiann
---
# xlm-roberta-base-wikiann-ner
## Model description
**xlm-roberta-base-wikiann-ner** is the first **Named Entity Recognition** model for 20 languages (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese) based on a fine-tuned XLM-RoBERTa large model. It achieves the **state-of-the-art performance** for the NER task. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of languages datasets obtained from [WikiANN](https://huggingface.co/datasets/wikiann) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("xlm-roberta-base-wikiann-ner")
model = AutoModelForTokenClassification.from_pretrained("xlm-roberta-base-wikiann-ner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Ìbọn ń ró kù kù gẹ́gẹ́ bí ọwọ́ ọ̀pọ̀ aráàlù ṣe tẹ ìbọn ní Kyiv láti dojú kọ Russia"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 20 NER datasets (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese)[wikiann](https://huggingface.co/datasets/wikiann).
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
### BibTeX entry and citation info
```
| {} | Davlan/xlm-roberta-base-wikiann-ner | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"xlm-roberta",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* ar
* as
* bn
* ca
* en
* es
* eu
* fr
* gu
* hi
* id
* ig
* mr
* pa
* pt
* sw
* ur
* vi
* yo
* zh
* multilingual
datasets:
* wikiann
---
xlm-roberta-base-wikiann-ner
============================
Model description
-----------------
xlm-roberta-base-wikiann-ner is the first Named Entity Recognition model for 20 languages (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese) based on a fine-tuned XLM-RoBERTa large model. It achieves the state-of-the-art performance for the NER task. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of languages datasets obtained from WikiANN dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 20 NER datasets (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese)wikiann.
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
### BibTeX entry and citation info
'''
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 20 NER datasets (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese)wikiann.\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:",
"### BibTeX entry and citation info\n\n\n'''"
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 20 NER datasets (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese)wikiann.\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:",
"### BibTeX entry and citation info\n\n\n'''"
] | [
38,
21,
184,
13
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 20 NER datasets (Arabic, Assamese, Bengali, Catalan, English, Spanish, Basque, French, Gujarati, Hindi, Indonesia, Igbo, Marathi, Punjabi, Portugues and Swahili, Urdu, Vietnamese, Yoruba, Chinese)wikiann.\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:### BibTeX entry and citation info\n\n\n'''"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- amh
- hau
- ibo
- kin
- lug
- luo
- pcm
- swa
- wol
- yor
- multilingual
datasets:
- masakhaner
---
# xlm-roberta-large-masakhaner
## Model description
**xlm-roberta-large-masakhaner** is the first **Named Entity Recognition** model for 10 African languages (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) based on a fine-tuned XLM-RoBERTa large model. It achieves the **state-of-the-art performance** for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset.
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/xlm-roberta-large-masakhaner")
model = AutoModelForTokenClassification.from_pretrained("Davlan/xlm-roberta-large-masakhaner")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
This model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane [MasakhaNER](https://github.com/masakhane-io/masakhane-ner) dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-DATE |Beginning of a DATE entity right after another DATE entity
I-DATE |DATE entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the [original MasakhaNER paper](https://arxiv.org/abs/2103.11811) which trained & evaluated the model on MasakhaNER corpus.
## Eval results on Test set (F-score)
language|F1-score
-|-
amh |75.76
hau |91.75
ibo |86.26
kin |76.38
lug |84.64
luo |80.65
pcm |89.55
swa |89.48
wol |70.70
yor |82.05
### BibTeX entry and citation info
```
@article{adelani21tacl,
title = {Masakha{NER}: Named Entity Recognition for African Languages},
author = {David Ifeoluwa Adelani and Jade Abbott and Graham Neubig and Daniel D'souza and Julia Kreutzer and Constantine Lignos and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and Israel Abebe Azime and Shamsuddeen Muhammad and Chris Chinenye Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and Jesujoba Alabi and Seid Muhie Yimam and Tajuddeen Gwadabe and Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and Verrah Otiende and Iroro Orife and Davis David and Samba Ngom and Tosin Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and Chiamaka Chukwuneke and Nkiruka Odu and Eric Peter Wairagala and Samuel Oyerinde and Clemencia Siro and Tobius Saul Bateesa and Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and Ayodele Awokoya and Mouhamadane MBOUP and Dibora Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and Thierno Ibrahima DIOP and Abdoulaye Diallo and Adewale Akinfaderin and Tendai Marengereke and Salomey Osei},
journal = {Transactions of the Association for Computational Linguistics (TACL)},
month = {},
url = {https://arxiv.org/abs/2103.11811},
year = {2021}
}
```
| {} | Davlan/xlm-roberta-large-masakhaner | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"xlm-roberta",
"token-classification",
"arxiv:2103.11811",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2103.11811"
] | [] | TAGS
#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* amh
* hau
* ibo
* kin
* lug
* luo
* pcm
* swa
* wol
* yor
* multilingual
datasets:
* masakhaner
---
xlm-roberta-large-masakhaner
============================
Model description
-----------------
xlm-roberta-large-masakhaner is the first Named Entity Recognition model for 10 African languages (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) based on a fine-tuned XLM-RoBERTa large model. It achieves the state-of-the-art performance for the NER task. It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of African language datasets obtained from Masakhane MasakhaNER dataset.
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
This model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.
Eval results on Test set (F-score)
----------------------------------
### BibTeX entry and citation info
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------",
"### BibTeX entry and citation info"
] | [
48,
21,
285,
10
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #arxiv-2103.11811 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThis model was fine-tuned on 10 African NER datasets (Amharic, Hausa, Igbo, Kinyarwanda, Luganda, Nigerian Pidgin, Swahili, Wolof, and Yorùbá) Masakhane MasakhaNER dataset\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.\n\n\nEval results on Test set (F-score)\n----------------------------------### BibTeX entry and citation info"
] |
token-classification | transformers | Hugging Face's logo
---
language:
- ar
- de
- en
- es
- fr
- it
- lv
- nl
- pt
- zh
- multilingual
---
# xlm-roberta-large-ner-hrl
## Model description
**xlm-roberta-large-ner-hrl** is a **Named Entity Recognition** model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned XLM-RoBERTa large model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of 10 high-resourced languages
## Intended uses & limitations
#### How to use
You can use this model with Transformers *pipeline* for NER.
```python
from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("Davlan/xlm-roberta-large-ner-hrl")
model = AutoModelForTokenClassification.from_pretrained("Davlan/xlm-roberta-large-ner-hrl")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Nader Jokhadar had given Syria the lead with a well-struck header in the seventh minute."
ner_results = nlp(example)
print(ner_results)
```
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
## Training data
The training data for the 10 languages are from:
Language|Dataset
-|-
Arabic | [ANERcorp](https://camel.abudhabi.nyu.edu/anercorp/)
German | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
English | [conll 2003](https://www.clips.uantwerpen.be/conll2003/ner/)
Spanish | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
French | [Europeana Newspapers](https://github.com/EuropeanaNewspapers/ner-corpora/tree/master/enp_FR.bnf.bio)
Italian | [Italian I-CAB](https://ontotext.fbk.eu/icab.html)
Latvian | [Latvian NER](https://github.com/LUMII-AILab/FullStack/tree/master/NamedEntities)
Dutch | [conll 2002](https://www.clips.uantwerpen.be/conll2002/ner/)
Portuguese |[Paramopama + Second Harem](https://github.com/davidsbatista/NER-datasets/tree/master/Portuguese)
Chinese | [MSRA](https://huggingface.co/datasets/msra_ner)
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Abbreviation|Description
-|-
O|Outside of a named entity
B-PER |Beginning of a person’s name right after another person’s name
I-PER |Person’s name
B-ORG |Beginning of an organisation right after another organisation
I-ORG |Organisation
B-LOC |Beginning of a location right after another location
I-LOC |Location
## Training procedure
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code. | {"license": "afl-3.0"} | Davlan/xlm-roberta-large-ner-hrl | null | [
"transformers",
"pytorch",
"tf",
"safetensors",
"xlm-roberta",
"token-classification",
"license:afl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us
| Hugging Face's logo
-------------------
language:
* ar
* de
* en
* es
* fr
* it
* lv
* nl
* pt
* zh
* multilingual
---
xlm-roberta-large-ner-hrl
=========================
Model description
-----------------
xlm-roberta-large-ner-hrl is a Named Entity Recognition model for 10 high resourced languages (Arabic, German, English, Spanish, French, Italian, Latvian, Dutch, Portuguese and Chinese) based on a fine-tuned XLM-RoBERTa large model. It has been trained to recognize three types of entities: location (LOC), organizations (ORG), and person (PER).
Specifically, this model is a *xlm-roberta-large* model that was fine-tuned on an aggregation of 10 high-resourced languages
Intended uses & limitations
---------------------------
#### How to use
You can use this model with Transformers *pipeline* for NER.
#### Limitations and bias
This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.
Training data
-------------
The training data for the 10 languages are from:
The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:
Training procedure
------------------
This model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code.
| [
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.",
"#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] | [
46,
21,
170
] | [
"TAGS\n#transformers #pytorch #tf #safetensors #xlm-roberta #token-classification #license-afl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nYou can use this model with Transformers *pipeline* for NER.#### Limitations and bias\n\n\nThis model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.\n\n\nTraining data\n-------------\n\n\nThe training data for the 10 languages are from:\n\n\n\nThe training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:\n\n\n\nTraining procedure\n------------------\n\n\nThis model was trained on NVIDIA V100 GPU with recommended hyperparameters from HuggingFace code."
] |
text-generation | transformers |
# Iron Man DialoGPT Model | {"tags": ["conversational"]} | Dawit/DialogGPT-small-ironman | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Iron Man DialoGPT Model | [
"# Iron Man DialoGPT Model"
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Iron Man DialoGPT Model"
] | [
39,
7
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Iron Man DialoGPT Model"
] |
text-generation | transformers |
# My Awesome Model
| {"tags": ["conversational"]} | Daymarebait/Discord_BOT_RICK | null | [
"transformers",
"conversational",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #conversational #endpoints_compatible #region-us
|
# My Awesome Model
| [
"# My Awesome Model"
] | [
"TAGS\n#transformers #conversational #endpoints_compatible #region-us \n",
"# My Awesome Model"
] | [
15,
4
] | [
"TAGS\n#transformers #conversational #endpoints_compatible #region-us \n# My Awesome Model"
] |
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# emoBERTTamil
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the tamilmixsentiment dataset.
It achieves the following results on the evaluation set:
- Loss: 0.9666
- Accuracy: 0.671
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 1.1128 | 1.0 | 250 | 1.0290 | 0.672 |
| 1.0226 | 2.0 | 500 | 1.0172 | 0.686 |
| 0.9137 | 3.0 | 750 | 0.9666 | 0.671 |
### Framework versions
- Transformers 4.9.2
- Pytorch 1.9.0+cu102
- Datasets 1.11.0
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["tamilmixsentiment"], "metrics": ["accuracy"], "model_index": [{"name": "emoBERTTamil", "results": [{"task": {"name": "Text Classification", "type": "text-classification"}, "dataset": {"name": "tamilmixsentiment", "type": "tamilmixsentiment", "args": "default"}, "metric": {"name": "Accuracy", "type": "accuracy", "value": 0.671}}]}]} | DeadBeast/emoBERTTamil | null | [
"transformers",
"pytorch",
"tensorboard",
"bert",
"text-classification",
"generated_from_trainer",
"dataset:tamilmixsentiment",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #dataset-tamilmixsentiment #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| emoBERTTamil
============
This model is a fine-tuned version of bert-base-uncased on the tamilmixsentiment dataset.
It achieves the following results on the evaluation set:
* Loss: 0.9666
* Accuracy: 0.671
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Training results
### Framework versions
* Transformers 4.9.2
* Pytorch 1.9.0+cu102
* Datasets 1.11.0
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #dataset-tamilmixsentiment #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] | [
55,
103,
5,
44
] | [
"TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #dataset-tamilmixsentiment #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0### Training results### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] |
text-classification | transformers |
# **Korean-mBERT**
This model is a fine-tune checkpoint of mBERT-base-cased over **Hugging Face Kore_Scm** dataset for Text classification.
### **How to use?**
**Task**: binary-classification
- LABEL_1: Sarcasm (*Sarcasm means tweets contains sarcasm*)
- LABEL_0: Not Sarcasm (*Not Sarcasm means tweets do not contain sarcasm*)
Click on **Use in Transformers**!
| {"language": "korean", "license": "apache-2.0", "datasets": ["Korean-Sarcasm"]} | DeadBeast/korscm-mBERT | null | [
"transformers",
"pytorch",
"bert",
"text-classification",
"dataset:Korean-Sarcasm",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"korean"
] | TAGS
#transformers #pytorch #bert #text-classification #dataset-Korean-Sarcasm #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Korean-mBERT
This model is a fine-tune checkpoint of mBERT-base-cased over Hugging Face Kore_Scm dataset for Text classification.
### How to use?
Task: binary-classification
- LABEL_1: Sarcasm (*Sarcasm means tweets contains sarcasm*)
- LABEL_0: Not Sarcasm (*Not Sarcasm means tweets do not contain sarcasm*)
Click on Use in Transformers!
| [
"# Korean-mBERT\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Hugging Face Kore_Scm dataset for Text classification.",
"### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Sarcasm (*Sarcasm means tweets contains sarcasm*)\n- LABEL_0: Not Sarcasm (*Not Sarcasm means tweets do not contain sarcasm*)\n\nClick on Use in Transformers!"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #dataset-Korean-Sarcasm #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Korean-mBERT\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Hugging Face Kore_Scm dataset for Text classification.",
"### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Sarcasm (*Sarcasm means tweets contains sarcasm*)\n- LABEL_0: Not Sarcasm (*Not Sarcasm means tweets do not contain sarcasm*)\n\nClick on Use in Transformers!"
] | [
43,
35,
56
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #dataset-Korean-Sarcasm #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Korean-mBERT\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Hugging Face Kore_Scm dataset for Text classification.### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Sarcasm (*Sarcasm means tweets contains sarcasm*)\n- LABEL_0: Not Sarcasm (*Not Sarcasm means tweets do not contain sarcasm*)\n\nClick on Use in Transformers!"
] |
text-classification | transformers |
# **mBERT-base-cased-finetuned-bengali-fakenews**
This model is a fine-tune checkpoint of mBERT-base-cased over **[Bengali-fake-news Dataset](https://www.kaggle.com/cryptexcode/banfakenews)** for Text classification. This model reaches an accuracy of 96.3 with an f1-score of 79.1 on the dev set.
### **How to use?**
**Task**: binary-classification
- LABEL_1: Authentic (*Authentic means news is authentic*)
- LABEL_0: Fake (*Fake means news is fake*)
```
from transformers import pipeline
print(pipeline("sentiment-analysis",model="DeadBeast/mbert-base-cased-finetuned-bengali-fakenews",tokenizer="DeadBeast/mbert-base-cased-finetuned-bengali-fakenews")("অভিনেতা আফজাল শরীফকে ২০ লাখ টাকার অনুদান অসুস্থ অভিনেতা আফজাল শরীফকে চিকিৎসার জন্য ২০ লাখ টাকা অনুদান দিয়েছেন প্রধানমন্ত্রী শেখ হাসিনা।"))
``` | {"language": "bengali", "license": "apache-2.0", "datasets": ["BanFakeNews"]} | DeadBeast/mbert-base-cased-finetuned-bengali-fakenews | null | [
"transformers",
"pytorch",
"bert",
"text-classification",
"dataset:BanFakeNews",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"bengali"
] | TAGS
#transformers #pytorch #bert #text-classification #dataset-BanFakeNews #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# mBERT-base-cased-finetuned-bengali-fakenews
This model is a fine-tune checkpoint of mBERT-base-cased over Bengali-fake-news Dataset for Text classification. This model reaches an accuracy of 96.3 with an f1-score of 79.1 on the dev set.
### How to use?
Task: binary-classification
- LABEL_1: Authentic (*Authentic means news is authentic*)
- LABEL_0: Fake (*Fake means news is fake*)
| [
"# mBERT-base-cased-finetuned-bengali-fakenews\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Bengali-fake-news Dataset for Text classification. This model reaches an accuracy of 96.3 with an f1-score of 79.1 on the dev set.",
"### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Authentic (*Authentic means news is authentic*)\n- LABEL_0: Fake (*Fake means news is fake*)"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #dataset-BanFakeNews #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# mBERT-base-cased-finetuned-bengali-fakenews\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Bengali-fake-news Dataset for Text classification. This model reaches an accuracy of 96.3 with an f1-score of 79.1 on the dev set.",
"### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Authentic (*Authentic means news is authentic*)\n- LABEL_0: Fake (*Fake means news is fake*)"
] | [
45,
69,
42
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #dataset-BanFakeNews #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# mBERT-base-cased-finetuned-bengali-fakenews\n\nThis model is a fine-tune checkpoint of mBERT-base-cased over Bengali-fake-news Dataset for Text classification. This model reaches an accuracy of 96.3 with an f1-score of 79.1 on the dev set.### How to use?\n\nTask: binary-classification\n\n- LABEL_1: Authentic (*Authentic means news is authentic*)\n- LABEL_0: Fake (*Fake means news is fake*)"
] |
null | null | ---
Summarisation model summarsiation | {} | Dean/summarsiation | null | [
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#region-us
| ---
Summarisation model summarsiation | [] | [
"TAGS\n#region-us \n"
] | [
5
] | [
"TAGS\n#region-us \n"
] |
text-generation | transformers |
#Scaramouche DialoGPT Model | {"tags": ["conversational"]} | DecafNosebleed/DialoGPT-small-ScaraBot | null | [
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
#Scaramouche DialoGPT Model | [] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
39
] | [
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
null | transformers |
# Model Card for ChemBERTa-10M-MTR
# Model Details
## Model Description
More information needed
- **Developed by:** DeepChem
- **Shared by [Optional]:** DeepChem
- **Model type:** Token Classification
- **Language(s) (NLP):** More information needed
- **License:** More information needed
- **Parent Model:** [RoBERTa](https://huggingface.co/roberta-base?text=The+goal+of+life+is+%3Cmask%3E.)
- **Resources for more information:** More information needed
# Uses
## Direct Use
More information needed.
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
More information needed
## Training Procedure
### Preprocessing
More information needed
### Speeds, Sizes, Times
More information needed
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** More information needed
- **Hours used:** More information needed
- **Cloud Provider:** More information needed
- **Compute Region:** More information needed
- **Carbon Emitted:** More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed.
# Citation
**BibTeX:**
```bibtex
@book{Ramsundar-et-al-2019,
title={Deep Learning for the Life Sciences},
author={Bharath Ramsundar and Peter Eastman and Patrick Walters and Vijay Pande and Karl Leswing and Zhenqin Wu},
publisher={O'Reilly Media},
note={\url{https://www.amazon.com/Deep-Learning-Life-Sciences-Microscopy/dp/1492039837}},
year={2019}
}
```
**APA:**
More information needed
# Glossary [optional]
More information needed
# More Information [optional]
More information needed
# Model Card Authors [optional]
DeepChem in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
```python
from transformers import AutoTokenizer, RobertaForRegression
tokenizer = AutoTokenizer.from_pretrained("DeepChem/ChemBERTa-10M-MTR")
model = RobertaForRegression.from_pretrained("DeepChem/ChemBERTa-10M-MTR")
```
</details>
| {"tags": ["roberta"]} | DeepChem/ChemBERTa-10M-MTR | null | [
"transformers",
"pytorch",
"roberta",
"arxiv:1910.09700",
"endpoints_compatible",
"has_space",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #pytorch #roberta #arxiv-1910.09700 #endpoints_compatible #has_space #region-us
|
# Model Card for ChemBERTa-10M-MTR
# Model Details
## Model Description
More information needed
- Developed by: DeepChem
- Shared by [Optional]: DeepChem
- Model type: Token Classification
- Language(s) (NLP): More information needed
- License: More information needed
- Parent Model: RoBERTa
- Resources for more information: More information needed
# Uses
## Direct Use
More information needed.
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
More information needed
## Training Procedure
### Preprocessing
More information needed
### Speeds, Sizes, Times
More information needed
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: More information needed
- Hours used: More information needed
- Cloud Provider: More information needed
- Compute Region: More information needed
- Carbon Emitted: More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed.
BibTeX:
APA:
More information needed
# Glossary [optional]
More information needed
# More Information [optional]
More information needed
# Model Card Authors [optional]
DeepChem in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
</details>
| [
"# Model Card for ChemBERTa-10M-MTR",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: DeepChem\n- Shared by [Optional]: DeepChem\n\n- Model type: Token Classification\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: RoBERTa\n- Resources for more information: More information needed",
"# Uses",
"## Direct Use\nMore information needed.",
"## Downstream Use [Optional]\n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nMore information needed",
"## Training Procedure",
"### Preprocessing\n \nMore information needed",
"### Speeds, Sizes, Times\n \nMore information needed",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\nMore information needed",
"### Metrics\n \nMore information needed",
"## Results \n \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \n \nMore information needed",
"### Software\n \nMore information needed.\n \nBibTeX:\n \n\n\n\n\nAPA:\n\nMore information needed",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\nMore information needed",
"# Model Card Authors [optional]\n \nDeepChem in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] | [
"TAGS\n#transformers #pytorch #roberta #arxiv-1910.09700 #endpoints_compatible #has_space #region-us \n",
"# Model Card for ChemBERTa-10M-MTR",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: DeepChem\n- Shared by [Optional]: DeepChem\n\n- Model type: Token Classification\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: RoBERTa\n- Resources for more information: More information needed",
"# Uses",
"## Direct Use\nMore information needed.",
"## Downstream Use [Optional]\n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nMore information needed",
"## Training Procedure",
"### Preprocessing\n \nMore information needed",
"### Speeds, Sizes, Times\n \nMore information needed",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\nMore information needed",
"### Metrics\n \nMore information needed",
"## Results \n \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \n \nMore information needed",
"### Software\n \nMore information needed.\n \nBibTeX:\n \n\n\n\n\nAPA:\n\nMore information needed",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\nMore information needed",
"# Model Card Authors [optional]\n \nDeepChem in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] | [
33,
13,
3,
63,
2,
8,
11,
25,
70,
33,
3,
7,
4,
10,
11,
2,
9,
8,
7,
8,
6,
6,
63,
6,
9,
7,
7,
18,
9,
9,
23,
7,
36
] | [
"TAGS\n#transformers #pytorch #roberta #arxiv-1910.09700 #endpoints_compatible #has_space #region-us \n# Model Card for ChemBERTa-10M-MTR# Model Details## Model Description\n \nMore information needed\n \n- Developed by: DeepChem\n- Shared by [Optional]: DeepChem\n\n- Model type: Token Classification\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: RoBERTa\n- Resources for more information: More information needed# Uses## Direct Use\nMore information needed.## Downstream Use [Optional]\n \nMore information needed.## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.# Training Details## Training Data\n \nMore information needed## Training Procedure### Preprocessing\n \nMore information needed### Speeds, Sizes, Times\n \nMore information needed# Evaluation## Testing Data, Factors & Metrics### Testing Data\n \nMore information needed### Factors\nMore information needed### Metrics\n \nMore information needed## Results \n \nMore information needed# Model Examination\n \nMore information needed# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed# Technical Specifications [optional]## Model Architecture and Objective\n \nMore information needed## Compute Infrastructure\n \nMore information needed### Hardware\n \n \nMore information needed### Software\n \nMore information needed.\n \nBibTeX:\n \n\n\n\n\nAPA:\n\nMore information needed# Glossary [optional]\n \nMore information needed# More Information [optional]\nMore information needed# Model Card Authors [optional]\n \nDeepChem in collaboration with Ezi Ozoani and the Hugging Face team# Model Card Contact\n \nMore information needed# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
feature-extraction | transformers | RoBERTa model trained on 1M SMILES from PubChem 77M set in MoleculeNet. Uses Smiles-Tokenizer | {} | DeepChem/SmilesTokenizer_PubChem_1M | null | [
"transformers",
"pytorch",
"roberta",
"feature-extraction",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [] | TAGS
#transformers #pytorch #roberta #feature-extraction #endpoints_compatible #region-us
| RoBERTa model trained on 1M SMILES from PubChem 77M set in MoleculeNet. Uses Smiles-Tokenizer | [] | [
"TAGS\n#transformers #pytorch #roberta #feature-extraction #endpoints_compatible #region-us \n"
] | [
23
] | [
"TAGS\n#transformers #pytorch #roberta #feature-extraction #endpoints_compatible #region-us \n"
] |
text-generation | transformers | # GPT2-Spanish
GPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the medium version of the original OpenAI GPT2 model.
## Corpus
This model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).
## Tokenizer
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.
This tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.
Apart from the special token "<|endoftext|>" for text ending in the OpenAI GPT-2 models, the tokens "<|talk|>", "<|ax1|>", "<|ax2|>" (..)"<|ax9|>" were included so that they can serve as prompts in future training.
## Training
The model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.
## Authors
The model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).
Thanks to the members of the community who collaborated with funding for the initial tests.
## Cautions
The model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content.
| {"language": "es", "license": "mit", "tags": ["GPT-2", "Spanish", "ebooks", "nlg"], "datasets": ["ebooks"], "widget": [{"text": "Quisiera saber que va a suceder"}]} | DeepESP/gpt2-spanish-medium | null | [
"transformers",
"pytorch",
"tf",
"jax",
"gpt2",
"text-generation",
"GPT-2",
"Spanish",
"ebooks",
"nlg",
"es",
"dataset:ebooks",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"es"
] | TAGS
#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # GPT2-Spanish
GPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the medium version of the original OpenAI GPT2 model.
## Corpus
This model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).
## Tokenizer
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.
This tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.
Apart from the special token "<|endoftext|>" for text ending in the OpenAI GPT-2 models, the tokens "<|talk|>", "<|ax1|>", "<|ax2|>" (..)"<|ax9|>" were included so that they can serve as prompts in future training.
## Training
The model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.
## Authors
The model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).
Thanks to the members of the community who collaborated with funding for the initial tests.
## Cautions
The model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content.
| [
"# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the medium version of the original OpenAI GPT2 model.",
"## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).",
"## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.",
"## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.",
"## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.",
"## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] | [
"TAGS\n#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the medium version of the original OpenAI GPT2 model.",
"## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).",
"## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.",
"## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.",
"## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.",
"## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] | [
66,
66,
48,
178,
35,
87,
37
] | [
"TAGS\n#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the medium version of the original OpenAI GPT2 model.## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] |
text-generation | transformers |
# GPT2-Spanish
GPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the small version of the original OpenAI GPT2 model.
## Corpus
This model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).
## Tokenizer
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.
This tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.
Apart from the special token "<|endoftext|>" for text ending in the OpenAI GPT-2 models, the tokens "<|talk|>", "<|ax1|>", "<|ax2|>" (..)"<|ax9|>" were included so that they can serve as prompts in future training.
## Training
The model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.
## Authors
The model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).
Thanks to the members of the community who collaborated with funding for the initial tests.
## Cautions
The model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content.
| {"language": "es", "license": "mit", "tags": ["GPT-2", "Spanish", "ebooks", "nlg"], "datasets": ["ebooks"], "widget": [{"text": "Quisiera saber que va a suceder"}]} | DeepESP/gpt2-spanish | null | [
"transformers",
"pytorch",
"tf",
"jax",
"gpt2",
"text-generation",
"GPT-2",
"Spanish",
"ebooks",
"nlg",
"es",
"dataset:ebooks",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"es"
] | TAGS
#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
|
# GPT2-Spanish
GPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the small version of the original OpenAI GPT2 model.
## Corpus
This model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).
## Tokenizer
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.
This tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.
Apart from the special token "<|endoftext|>" for text ending in the OpenAI GPT-2 models, the tokens "<|talk|>", "<|ax1|>", "<|ax2|>" (..)"<|ax9|>" were included so that they can serve as prompts in future training.
## Training
The model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.
## Authors
The model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).
Thanks to the members of the community who collaborated with funding for the initial tests.
## Cautions
The model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content.
| [
"# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the small version of the original OpenAI GPT2 model.",
"## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).",
"## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.",
"## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.",
"## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.",
"## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] | [
"TAGS\n#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n",
"# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the small version of the original OpenAI GPT2 model.",
"## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).",
"## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.",
"## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.",
"## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.",
"## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] | [
70,
66,
48,
178,
35,
87,
37
] | [
"TAGS\n#transformers #pytorch #tf #jax #gpt2 #text-generation #GPT-2 #Spanish #ebooks #nlg #es #dataset-ebooks #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# GPT2-Spanish\nGPT2-Spanish is a language generation model trained from scratch with 11.5GB of Spanish texts and with a Byte Pair Encoding (BPE) tokenizer that was trained for this purpose. The parameters used are the same as the small version of the original OpenAI GPT2 model.## Corpus\nThis model was trained with a corpus of 11.5GB of texts corresponding to 3.5GB of Wikipedia articles and 8GB of books (narrative, short stories, theater, poetry, essays, and popularization).## Tokenizer\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for Unicode characters) and a vocabulary size of 50257. The inputs are sequences of 1024 consecutive tokens.\n\nThis tokenizer was trained from scratch with the Spanish corpus, since it was evidenced that the tokenizer of the English models presented limitations to capture the semantic relations of Spanish, due to the morphosyntactic differences between both languages.\n\nApart from the special token \"<|endoftext|>\" for text ending in the OpenAI GPT-2 models, the tokens \"<|talk|>\", \"<|ax1|>\", \"<|ax2|>\" (..)\"<|ax9|>\" were included so that they can serve as prompts in future training.## Training\nThe model and tokenizer were trained using the Hugging Face libraries with an Nvidia Tesla V100 GPU with 16GB memory on Google Colab servers.## Authors\nThe model was trained by Alejandro Oñate Latorre (Spain) and Jorge Ortiz Fuentes (Chile), members of -Deep ESP-, an open-source community on Natural Language Processing in Spanish (https://t.me/joinchat/VoEp1bPrDYEexc6h).\n\nThanks to the members of the community who collaborated with funding for the initial tests.## Cautions\nThe model generates texts according to the patterns learned in the training corpus. These data were not filtered, therefore, the model could generate offensive or discriminatory content."
] |
feature-extraction | transformers |
# bert-base-bg-cs-pl-ru-cased
SlavicBERT\[1\] \(Slavic \(bg, cs, pl, ru\), cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\) was trained on Russian News and four Wikipedias: Bulgarian, Czech, Polish, and Russian. Subtoken vocabulary was built using this data. Multilingual BERT was used as an initialization for SlavicBERT.
08.11.2021: upload model with MLM and NSP heads
\[1\]: Arkhipov M., Trofimova M., Kuratov Y., Sorokin A. \(2019\). [Tuning Multilingual Transformers for Language-Specific Named Entity Recognition](https://www.aclweb.org/anthology/W19-3712/). ACL anthology W19-3712.
| {"language": ["bg", "cs", "pl", "ru"]} | DeepPavlov/bert-base-bg-cs-pl-ru-cased | null | [
"transformers",
"pytorch",
"jax",
"bert",
"feature-extraction",
"bg",
"cs",
"pl",
"ru",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"bg",
"cs",
"pl",
"ru"
] | TAGS
#transformers #pytorch #jax #bert #feature-extraction #bg #cs #pl #ru #endpoints_compatible #region-us
|
# bert-base-bg-cs-pl-ru-cased
SlavicBERT\[1\] \(Slavic \(bg, cs, pl, ru\), cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\) was trained on Russian News and four Wikipedias: Bulgarian, Czech, Polish, and Russian. Subtoken vocabulary was built using this data. Multilingual BERT was used as an initialization for SlavicBERT.
08.11.2021: upload model with MLM and NSP heads
\[1\]: Arkhipov M., Trofimova M., Kuratov Y., Sorokin A. \(2019\). Tuning Multilingual Transformers for Language-Specific Named Entity Recognition. ACL anthology W19-3712.
| [
"# bert-base-bg-cs-pl-ru-cased\n\nSlavicBERT\\[1\\] \\(Slavic \\(bg, cs, pl, ru\\), cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) was trained on Russian News and four Wikipedias: Bulgarian, Czech, Polish, and Russian. Subtoken vocabulary was built using this data. Multilingual BERT was used as an initialization for SlavicBERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Arkhipov M., Trofimova M., Kuratov Y., Sorokin A. \\(2019\\). Tuning Multilingual Transformers for Language-Specific Named Entity Recognition. ACL anthology W19-3712."
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #bg #cs #pl #ru #endpoints_compatible #region-us \n",
"# bert-base-bg-cs-pl-ru-cased\n\nSlavicBERT\\[1\\] \\(Slavic \\(bg, cs, pl, ru\\), cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) was trained on Russian News and four Wikipedias: Bulgarian, Czech, Polish, and Russian. Subtoken vocabulary was built using this data. Multilingual BERT was used as an initialization for SlavicBERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Arkhipov M., Trofimova M., Kuratov Y., Sorokin A. \\(2019\\). Tuning Multilingual Transformers for Language-Specific Named Entity Recognition. ACL anthology W19-3712."
] | [
34,
177
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #bg #cs #pl #ru #endpoints_compatible #region-us \n# bert-base-bg-cs-pl-ru-cased\n\nSlavicBERT\\[1\\] \\(Slavic \\(bg, cs, pl, ru\\), cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) was trained on Russian News and four Wikipedias: Bulgarian, Czech, Polish, and Russian. Subtoken vocabulary was built using this data. Multilingual BERT was used as an initialization for SlavicBERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Arkhipov M., Trofimova M., Kuratov Y., Sorokin A. \\(2019\\). Tuning Multilingual Transformers for Language-Specific Named Entity Recognition. ACL anthology W19-3712."
] |
feature-extraction | transformers |
# bert-base-cased-conversational
Conversational BERT \(English, cased, 12‑layer, 768‑hidden, 12‑heads, 110M parameters\) was trained on the English part of Twitter, Reddit, DailyDialogues\[1\], OpenSubtitles\[2\], Debates\[3\], Blogs\[4\], Facebook News Comments. We used this training data to build the vocabulary of English subtokens and took English cased version of BERT‑base as an initialization for English Conversational BERT.
08.11.2021: upload model with MLM and NSP heads
\[1\]: Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. IJCNLP 2017.
\[2\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \(LREC 2016\)
\[3\]: Justine Zhang, Ravi Kumar, Sujith Ravi, Cristian Danescu-Niculescu-Mizil. Proceedings of NAACL, 2016.
\[4\]: J. Schler, M. Koppel, S. Argamon and J. Pennebaker \(2006\). Effects of Age and Gender on Blogging in Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs.
| {"language": "en"} | DeepPavlov/bert-base-cased-conversational | null | [
"transformers",
"pytorch",
"jax",
"bert",
"feature-extraction",
"en",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [] | [
"en"
] | TAGS
#transformers #pytorch #jax #bert #feature-extraction #en #endpoints_compatible #region-us
|
# bert-base-cased-conversational
Conversational BERT \(English, cased, 12‑layer, 768‑hidden, 12‑heads, 110M parameters\) was trained on the English part of Twitter, Reddit, DailyDialogues\[1\], OpenSubtitles\[2\], Debates\[3\], Blogs\[4\], Facebook News Comments. We used this training data to build the vocabulary of English subtokens and took English cased version of BERT‑base as an initialization for English Conversational BERT.
08.11.2021: upload model with MLM and NSP heads
\[1\]: Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. IJCNLP 2017.
\[2\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \(LREC 2016\)
\[3\]: Justine Zhang, Ravi Kumar, Sujith Ravi, Cristian Danescu-Niculescu-Mizil. Proceedings of NAACL, 2016.
\[4\]: J. Schler, M. Koppel, S. Argamon and J. Pennebaker \(2006\). Effects of Age and Gender on Blogging in Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs.
| [
"# bert-base-cased-conversational\n\nConversational BERT \\(English, cased, 12‑layer, 768‑hidden, 12‑heads, 110M parameters\\) was trained on the English part of Twitter, Reddit, DailyDialogues\\[1\\], OpenSubtitles\\[2\\], Debates\\[3\\], Blogs\\[4\\], Facebook News Comments. We used this training data to build the vocabulary of English subtokens and took English cased version of BERT‑base as an initialization for English Conversational BERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. IJCNLP 2017.\n\n\\[2\\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \\(LREC 2016\\)\n\n\\[3\\]: Justine Zhang, Ravi Kumar, Sujith Ravi, Cristian Danescu-Niculescu-Mizil. Proceedings of NAACL, 2016.\n\n\\[4\\]: J. Schler, M. Koppel, S. Argamon and J. Pennebaker \\(2006\\). Effects of Age and Gender on Blogging in Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs."
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #en #endpoints_compatible #region-us \n",
"# bert-base-cased-conversational\n\nConversational BERT \\(English, cased, 12‑layer, 768‑hidden, 12‑heads, 110M parameters\\) was trained on the English part of Twitter, Reddit, DailyDialogues\\[1\\], OpenSubtitles\\[2\\], Debates\\[3\\], Blogs\\[4\\], Facebook News Comments. We used this training data to build the vocabulary of English subtokens and took English cased version of BERT‑base as an initialization for English Conversational BERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. IJCNLP 2017.\n\n\\[2\\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \\(LREC 2016\\)\n\n\\[3\\]: Justine Zhang, Ravi Kumar, Sujith Ravi, Cristian Danescu-Niculescu-Mizil. Proceedings of NAACL, 2016.\n\n\\[4\\]: J. Schler, M. Koppel, S. Argamon and J. Pennebaker \\(2006\\). Effects of Age and Gender on Blogging in Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs."
] | [
27,
346
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #en #endpoints_compatible #region-us \n# bert-base-cased-conversational\n\nConversational BERT \\(English, cased, 12‑layer, 768‑hidden, 12‑heads, 110M parameters\\) was trained on the English part of Twitter, Reddit, DailyDialogues\\[1\\], OpenSubtitles\\[2\\], Debates\\[3\\], Blogs\\[4\\], Facebook News Comments. We used this training data to build the vocabulary of English subtokens and took English cased version of BERT‑base as an initialization for English Conversational BERT.\n\n08.11.2021: upload model with MLM and NSP heads\n\n\\[1\\]: Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. IJCNLP 2017.\n\n\\[2\\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \\(LREC 2016\\)\n\n\\[3\\]: Justine Zhang, Ravi Kumar, Sujith Ravi, Cristian Danescu-Niculescu-Mizil. Proceedings of NAACL, 2016.\n\n\\[4\\]: J. Schler, M. Koppel, S. Argamon and J. Pennebaker \\(2006\\). Effects of Age and Gender on Blogging in Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs."
] |
feature-extraction | transformers |
# bert-base-multilingual-cased-sentence
Sentence Multilingual BERT \(101 languages, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\) is a representation‑based sentence encoder for 101 languages of Multilingual BERT. It is initialized with Multilingual BERT and then fine‑tuned on english MultiNLI\[1\] and on dev set of multilingual XNLI\[2\]. Sentence representations are mean pooled token embeddings in the same manner as in Sentence‑BERT\[3\].
\[1\]: Williams A., Nangia N. & Bowman S. \(2017\) A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv preprint [arXiv:1704.05426](https://arxiv.org/abs/1704.05426)
\[2\]: Williams A., Bowman S. \(2018\) XNLI: Evaluating Cross-lingual Sentence Representations. arXiv preprint [arXiv:1809.05053](https://arxiv.org/abs/1809.05053)
\[3\]: N. Reimers, I. Gurevych \(2019\) Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint [arXiv:1908.10084](https://arxiv.org/abs/1908.10084)
| {"language": ["multilingual"]} | DeepPavlov/bert-base-multilingual-cased-sentence | null | [
"transformers",
"pytorch",
"jax",
"bert",
"feature-extraction",
"multilingual",
"arxiv:1704.05426",
"arxiv:1809.05053",
"arxiv:1908.10084",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"1704.05426",
"1809.05053",
"1908.10084"
] | [
"multilingual"
] | TAGS
#transformers #pytorch #jax #bert #feature-extraction #multilingual #arxiv-1704.05426 #arxiv-1809.05053 #arxiv-1908.10084 #endpoints_compatible #region-us
|
# bert-base-multilingual-cased-sentence
Sentence Multilingual BERT \(101 languages, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\) is a representation‑based sentence encoder for 101 languages of Multilingual BERT. It is initialized with Multilingual BERT and then fine‑tuned on english MultiNLI\[1\] and on dev set of multilingual XNLI\[2\]. Sentence representations are mean pooled token embeddings in the same manner as in Sentence‑BERT\[3\].
\[1\]: Williams A., Nangia N. & Bowman S. \(2017\) A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv preprint arXiv:1704.05426
\[2\]: Williams A., Bowman S. \(2018\) XNLI: Evaluating Cross-lingual Sentence Representations. arXiv preprint arXiv:1809.05053
\[3\]: N. Reimers, I. Gurevych \(2019\) Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084
| [
"# bert-base-multilingual-cased-sentence\n\nSentence Multilingual BERT \\(101 languages, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) is a representation‑based sentence encoder for 101 languages of Multilingual BERT. It is initialized with Multilingual BERT and then fine‑tuned on english MultiNLI\\[1\\] and on dev set of multilingual XNLI\\[2\\]. Sentence representations are mean pooled token embeddings in the same manner as in Sentence‑BERT\\[3\\].\n\n\n\\[1\\]: Williams A., Nangia N. & Bowman S. \\(2017\\) A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv preprint arXiv:1704.05426\n\n\\[2\\]: Williams A., Bowman S. \\(2018\\) XNLI: Evaluating Cross-lingual Sentence Representations. arXiv preprint arXiv:1809.05053\n\n\\[3\\]: N. Reimers, I. Gurevych \\(2019\\) Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084"
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #multilingual #arxiv-1704.05426 #arxiv-1809.05053 #arxiv-1908.10084 #endpoints_compatible #region-us \n",
"# bert-base-multilingual-cased-sentence\n\nSentence Multilingual BERT \\(101 languages, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) is a representation‑based sentence encoder for 101 languages of Multilingual BERT. It is initialized with Multilingual BERT and then fine‑tuned on english MultiNLI\\[1\\] and on dev set of multilingual XNLI\\[2\\]. Sentence representations are mean pooled token embeddings in the same manner as in Sentence‑BERT\\[3\\].\n\n\n\\[1\\]: Williams A., Nangia N. & Bowman S. \\(2017\\) A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv preprint arXiv:1704.05426\n\n\\[2\\]: Williams A., Bowman S. \\(2018\\) XNLI: Evaluating Cross-lingual Sentence Representations. arXiv preprint arXiv:1809.05053\n\n\\[3\\]: N. Reimers, I. Gurevych \\(2019\\) Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084"
] | [
58,
276
] | [
"TAGS\n#transformers #pytorch #jax #bert #feature-extraction #multilingual #arxiv-1704.05426 #arxiv-1809.05053 #arxiv-1908.10084 #endpoints_compatible #region-us \n# bert-base-multilingual-cased-sentence\n\nSentence Multilingual BERT \\(101 languages, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters\\) is a representation‑based sentence encoder for 101 languages of Multilingual BERT. It is initialized with Multilingual BERT and then fine‑tuned on english MultiNLI\\[1\\] and on dev set of multilingual XNLI\\[2\\]. Sentence representations are mean pooled token embeddings in the same manner as in Sentence‑BERT\\[3\\].\n\n\n\\[1\\]: Williams A., Nangia N. & Bowman S. \\(2017\\) A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv preprint arXiv:1704.05426\n\n\\[2\\]: Williams A., Bowman S. \\(2018\\) XNLI: Evaluating Cross-lingual Sentence Representations. arXiv preprint arXiv:1809.05053\n\n\\[3\\]: N. Reimers, I. Gurevych \\(2019\\) Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. arXiv preprint arXiv:1908.10084"
] |
null | transformers | # distilrubert-base-cased-conversational
Conversational DistilRuBERT \(Russian, cased, 6‑layer, 768‑hidden, 12‑heads, 135.4M parameters\) was trained on OpenSubtitles\[1\], [Dirty](https://d3.ru/), [Pikabu](https://pikabu.ru/), and a Social Media segment of Taiga corpus\[2\] (as [Conversational RuBERT](https://huggingface.co/DeepPavlov/rubert-base-cased-conversational)).
Our DistilRuBERT was highly inspired by \[3\], \[4\]. Namely, we used
* KL loss (between teacher and student output logits)
* MLM loss (between tokens labels and student output logits)
* Cosine embedding loss between mean of two consecutive hidden states of the teacher and one hidden state of the student
The model was trained for about 100 hrs. on 8 nVIDIA Tesla P100-SXM2.0 16Gb.
To evaluate improvements in the inference speed, we ran teacher and student models on random sequences with seq_len=512, batch_size = 16 (for throughput) and batch_size=1 (for latency).
All tests were performed on Intel(R) Xeon(R) CPU E5-2698 v4 @ 2.20GHz and nVIDIA Tesla P100-SXM2.0 16Gb.
| Model | Size, Mb. | CPU latency, sec.| GPU latency, sec. | CPU throughput, samples/sec. | GPU throughput, samples/sec. |
|-------------------------------------------------|------------|------------------|-------------------|------------------------------|------------------------------|
| Teacher (RuBERT-base-cased-conversational) | 679 | 0.655 | 0.031 | 0.3754 | 36.4902 |
| Student (DistilRuBERT-base-cased-conversational)| 517 | 0.3285 | 0.0212 | 0.5803 | 52.2495 |
# Citation
If you found the model useful for your research, we are kindly ask to cite [this](https://arxiv.org/abs/2205.02340) paper:
```
@misc{https://doi.org/10.48550/arxiv.2205.02340,
doi = {10.48550/ARXIV.2205.02340},
url = {https://arxiv.org/abs/2205.02340},
author = {Kolesnikova, Alina and Kuratov, Yuri and Konovalov, Vasily and Burtsev, Mikhail},
keywords = {Computation and Language (cs.CL), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Knowledge Distillation of Russian Language Models with Reduction of Vocabulary},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
\[1\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \(LREC 2016\)
\[2\]: Shavrina T., Shapovalova O. \(2017\) TO THE METHODOLOGY OF CORPUS CONSTRUCTION FOR MACHINE LEARNING: «TAIGA» SYNTAX TREE CORPUS AND PARSER. in proc. of “CORPORA2017”, international conference , Saint-Petersbourg, 2017.
\[3\]: Sanh, V., Debut, L., Chaumond, J., & Wolf, T. \(2019\). DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
\[4\]: <https://github.com/huggingface/transformers/tree/master/examples/research_projects/distillation> | {"language": ["ru"]} | DeepPavlov/distilrubert-base-cased-conversational | null | [
"transformers",
"pytorch",
"distilbert",
"ru",
"arxiv:2205.02340",
"endpoints_compatible",
"region:us"
] | null | 2022-03-02T23:29:04+00:00 | [
"2205.02340"
] | [
"ru"
] | TAGS
#transformers #pytorch #distilbert #ru #arxiv-2205.02340 #endpoints_compatible #region-us
| distilrubert-base-cased-conversational
======================================
Conversational DistilRuBERT (Russian, cased, 6‑layer, 768‑hidden, 12‑heads, 135.4M parameters) was trained on OpenSubtitles[1], Dirty, Pikabu, and a Social Media segment of Taiga corpus[2] (as Conversational RuBERT).
Our DistilRuBERT was highly inspired by [3], [4]. Namely, we used
* KL loss (between teacher and student output logits)
* MLM loss (between tokens labels and student output logits)
* Cosine embedding loss between mean of two consecutive hidden states of the teacher and one hidden state of the student
The model was trained for about 100 hrs. on 8 nVIDIA Tesla P100-SXM2.0 16Gb.
To evaluate improvements in the inference speed, we ran teacher and student models on random sequences with seq\_len=512, batch\_size = 16 (for throughput) and batch\_size=1 (for latency).
All tests were performed on Intel(R) Xeon(R) CPU E5-2698 v4 @ 2.20GHz and nVIDIA Tesla P100-SXM2.0 16Gb.
If you found the model useful for your research, we are kindly ask to cite this paper:
[1]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation (LREC 2016)
[2]: Shavrina T., Shapovalova O. (2017) TO THE METHODOLOGY OF CORPUS CONSTRUCTION FOR MACHINE LEARNING: «TAIGA» SYNTAX TREE CORPUS AND PARSER. in proc. of “CORPORA2017”, international conference , Saint-Petersbourg, 2017.
[3]: Sanh, V., Debut, L., Chaumond, J., & Wolf, T. (2019). DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
[4]: <URL
| [] | [
"TAGS\n#transformers #pytorch #distilbert #ru #arxiv-2205.02340 #endpoints_compatible #region-us \n"
] | [
34
] | [
"TAGS\n#transformers #pytorch #distilbert #ru #arxiv-2205.02340 #endpoints_compatible #region-us \n"
] |