File size: 717 Bytes
842eabb
929fb30
842eabb
929fb30
842eabb
 
 
929fb30
842eabb
 
475d585
929fb30
842eabb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0d3e0a6
842eabb
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
language:
- ne
thumbnail:
tags:
- roberta
- nepali-laguage-model
license: mit
datasets:
- cc100
widget:
- text: तिमीलाई कस्तो <mask>?
---

# nepbert

## Model description

Roberta trained from scratch on the Nepali CC-100 dataset with 12 million sentences.

## Intended uses & limitations

#### How to use

```python
from transformers import pipeline

pipe = pipeline(
    "fill-mask",
    model="amitness/nepbert",
    tokenizer="amitness/nepbert"
)
print(pipe(u"तिमीलाई कस्तो <mask>?"))
```

## Training data

The data was taken from the nepali language subset of CC-100 dataset.

## Training procedure
The model was trained on Google Colab using `1x Tesla V100`.