xmanii commited on
Commit
aae175e
·
verified ·
1 Parent(s): f5b1353

Model save

Browse files
README.md ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ model_name: mauxi-v0-3B
4
+ tags:
5
+ - generated_from_trainer
6
+ - trl
7
+ - sft
8
+ licence: license
9
+ ---
10
+
11
+ # Model Card for mauxi-v0-3B
12
+
13
+ This model is a fine-tuned version of [None](https://huggingface.co/None).
14
+ It has been trained using [TRL](https://github.com/huggingface/trl).
15
+
16
+ ## Quick start
17
+
18
+ ```python
19
+ from transformers import pipeline
20
+
21
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="xmanii/mauxi-v0-3B", device="cuda")
23
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
+ print(output["generated_text"])
25
+ ```
26
+
27
+ ## Training procedure
28
+
29
+
30
+
31
+ This model was trained with SFT.
32
+
33
+ ### Framework versions
34
+
35
+ - TRL: 0.12.1
36
+ - Transformers: 4.46.3
37
+ - Pytorch: 2.5.1
38
+ - Datasets: 3.1.0
39
+ - Tokenizers: 0.20.3
40
+
41
+ ## Citations
42
+
43
+
44
+
45
+ Cite TRL as:
46
+
47
+ ```bibtex
48
+ @misc{vonwerra2022trl,
49
+ title = {{TRL: Transformer Reinforcement Learning}},
50
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
51
+ year = 2020,
52
+ journal = {GitHub repository},
53
+ publisher = {GitHub},
54
+ howpublished = {\url{https://github.com/huggingface/trl}}
55
+ }
56
+ ```
generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.05,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.46.3"
14
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6201de9094fc143655fab8455dc35bf81c8df869f98856d72ae66921e126e441
3
  size 4957560304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e87869808b24e6a3add89293a281b5f102d2ae8849c3dee8a965a7b7243fcf
3
  size 4957560304
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:514c363da0fb1493a87f981345debc4d5f330c4661dfb9890c0a2253b4f9847c
3
  size 1214366696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6a9d4663f427fe45b2e377babd74658add6ea9c8ab96931d6e94d2f83761cf7
3
  size 1214366696
runs/Dec04_19-29-23_modal/events.out.tfevents.1733340565.modal.2.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20c11da36f8ae9d0d9ade10c10eeebe869f2982276ca9b272b7da039bf1a82f8
3
- size 9641
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4143e17c61d5f815dd173fc086020c3d8158b36221bd269f0247112eb0e5c694
3
+ size 10206