kd13 commited on
Commit
e819675
·
verified ·
1 Parent(s): 263d80b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -26
README.md CHANGED
@@ -2,29 +2,3 @@
2
  library_name: transformers
3
  pipeline_tag: fill-mask
4
  ---
5
-
6
- # MyBERT (RoPE + Pre-LN, ~21M params)
7
-
8
- Custom BERT-style encoder trained with MLM on packed BookCorpus.
9
- Trust remote code is required because the model uses RoPE.
10
-
11
- ```python
12
- from transformers import AutoTokenizer, AutoModelForMaskedLM
13
- import torch, torch.nn.functional as F
14
-
15
- tok = AutoTokenizer.from_pretrained("USERNAME/REPO")
16
- mdl = AutoModelForMaskedLM.from_pretrained("USERNAME/REPO", trust_remote_code=True).eval()
17
-
18
- text = f"the capital of france is {tok.mask_token}."
19
- enc = tok(text, return_tensors="pt")
20
- with torch.no_grad():
21
- logits = mdl(**enc).logits
22
- mask_pos = (enc["input_ids"][0] == tok.mask_token_id).nonzero()[0, 0]
23
- probs = F.softmax(logits[0, mask_pos], dim=-1)
24
- for p, i in zip(*[t.tolist() for t in probs.topk(5)]):
25
- print(f"{p:.4f} {tok.decode([i])!r}")
26
- ```
27
-
28
- > **Note:** This is a small model trained for limited compute. It does not have
29
- > strong factual knowledge and is best used as a base for fine-tuning on a
30
- > downstream task.
 
2
  library_name: transformers
3
  pipeline_tag: fill-mask
4
  ---