Instructions to use kd13/RoPERT-MLM-mini with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use kd13/RoPERT-MLM-mini with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("fill-mask", model="kd13/RoPERT-MLM-mini", trust_remote_code=True)# Load model directly from transformers import AutoModelForMaskedLM model = AutoModelForMaskedLM.from_pretrained("kd13/RoPERT-MLM-mini", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
Update README.md
Browse files
README.md
CHANGED
|
@@ -2,29 +2,3 @@
|
|
| 2 |
library_name: transformers
|
| 3 |
pipeline_tag: fill-mask
|
| 4 |
---
|
| 5 |
-
|
| 6 |
-
# MyBERT (RoPE + Pre-LN, ~21M params)
|
| 7 |
-
|
| 8 |
-
Custom BERT-style encoder trained with MLM on packed BookCorpus.
|
| 9 |
-
Trust remote code is required because the model uses RoPE.
|
| 10 |
-
|
| 11 |
-
```python
|
| 12 |
-
from transformers import AutoTokenizer, AutoModelForMaskedLM
|
| 13 |
-
import torch, torch.nn.functional as F
|
| 14 |
-
|
| 15 |
-
tok = AutoTokenizer.from_pretrained("USERNAME/REPO")
|
| 16 |
-
mdl = AutoModelForMaskedLM.from_pretrained("USERNAME/REPO", trust_remote_code=True).eval()
|
| 17 |
-
|
| 18 |
-
text = f"the capital of france is {tok.mask_token}."
|
| 19 |
-
enc = tok(text, return_tensors="pt")
|
| 20 |
-
with torch.no_grad():
|
| 21 |
-
logits = mdl(**enc).logits
|
| 22 |
-
mask_pos = (enc["input_ids"][0] == tok.mask_token_id).nonzero()[0, 0]
|
| 23 |
-
probs = F.softmax(logits[0, mask_pos], dim=-1)
|
| 24 |
-
for p, i in zip(*[t.tolist() for t in probs.topk(5)]):
|
| 25 |
-
print(f"{p:.4f} {tok.decode([i])!r}")
|
| 26 |
-
```
|
| 27 |
-
|
| 28 |
-
> **Note:** This is a small model trained for limited compute. It does not have
|
| 29 |
-
> strong factual knowledge and is best used as a base for fine-tuning on a
|
| 30 |
-
> downstream task.
|
|
|
|
| 2 |
library_name: transformers
|
| 3 |
pipeline_tag: fill-mask
|
| 4 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|