damienbenveniste commited on
Commit
8fff520
1 Parent(s): bcaab63

Push model using huggingface_hub.

Browse files
README.md CHANGED
@@ -25,7 +25,7 @@ You can then generate text as follows:
25
  ```python
26
  from transformers import pipeline
27
 
28
- generator = pipeline("text-generation", model="damienbenveniste//var/folders/qj/lfvfq6590q5fn7hnwx6c29k80000gn/T/tmppue5zznd/damienbenveniste/mistral-ppo")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
@@ -35,8 +35,8 @@ If you want to use the model for training or to obtain the outputs from the valu
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
- tokenizer = AutoTokenizer.from_pretrained("damienbenveniste//var/folders/qj/lfvfq6590q5fn7hnwx6c29k80000gn/T/tmppue5zznd/damienbenveniste/mistral-ppo")
39
- model = AutoModelForCausalLMWithValueHead.from_pretrained("damienbenveniste//var/folders/qj/lfvfq6590q5fn7hnwx6c29k80000gn/T/tmppue5zznd/damienbenveniste/mistral-ppo")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
 
25
  ```python
26
  from transformers import pipeline
27
 
28
+ generator = pipeline("text-generation", model="damienbenveniste//private/var/folders/dy/k5ycdcns28s2cxl8hc76v2mr0000gn/T/tmpl2a_v6sk/damienbenveniste/mistral-ppo")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
 
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
+ tokenizer = AutoTokenizer.from_pretrained("damienbenveniste//private/var/folders/dy/k5ycdcns28s2cxl8hc76v2mr0000gn/T/tmpl2a_v6sk/damienbenveniste/mistral-ppo")
39
+ model = AutoModelForCausalLMWithValueHead.from_pretrained("damienbenveniste//private/var/folders/dy/k5ycdcns28s2cxl8hc76v2mr0000gn/T/tmpl2a_v6sk/damienbenveniste/mistral-ppo")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
config.json CHANGED
@@ -3,8 +3,10 @@
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
 
8
  "hidden_act": "silu",
9
  "hidden_size": 768,
10
  "initializer_range": 0.02,
@@ -19,7 +21,7 @@
19
  "sliding_window": 768,
20
  "tie_word_embeddings": false,
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.35.2",
23
  "use_cache": true,
24
  "vocab_size": 32000
25
  }
 
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
6
+ "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
+ "head_dim": 48,
10
  "hidden_act": "silu",
11
  "hidden_size": 768,
12
  "initializer_range": 0.02,
 
21
  "sliding_window": 768,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.44.2",
25
  "use_cache": true,
26
  "vocab_size": 32000
27
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.35.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.44.2"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8de2bb878800087d0c7fbc318e19ee4fc1d428a587daa2bab14ccc6363b42ae
3
  size 338200972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ee50a22dabf1652e2365843c33fbb25cab5fa12acf25bc0c36147211f49b65
3
  size 338200972
tokenizer.json CHANGED
@@ -31,23 +31,13 @@
31
  "special": true
32
  }
33
  ],
34
- "normalizer": {
35
- "type": "Sequence",
36
- "normalizers": [
37
- {
38
- "type": "Prepend",
39
- "prepend": "▁"
40
- },
41
- {
42
- "type": "Replace",
43
- "pattern": {
44
- "String": " "
45
- },
46
- "content": "▁"
47
- }
48
- ]
49
  },
50
- "pre_tokenizer": null,
51
  "post_processor": {
52
  "type": "TemplateProcessing",
53
  "single": [
@@ -134,6 +124,7 @@
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
 
137
  "vocab": {
138
  "<unk>": 0,
139
  "<s>": 1,
 
31
  "special": true
32
  }
33
  ],
34
+ "normalizer": null,
35
+ "pre_tokenizer": {
36
+ "type": "Metaspace",
37
+ "replacement": "▁",
38
+ "prepend_scheme": "first",
39
+ "split": false
 
 
 
 
 
 
 
 
 
40
  },
 
41
  "post_processor": {
42
  "type": "TemplateProcessing",
43
  "single": [
 
124
  "end_of_word_suffix": null,
125
  "fuse_unk": true,
126
  "byte_fallback": true,
127
+ "ignore_merges": false,
128
  "vocab": {
129
  "<unk>": 0,
130
  "<s>": 1,
tokenizer_config.json CHANGED
@@ -1,4 +1,7 @@
1
  {
 
 
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<unk>",
@@ -29,8 +32,8 @@
29
  "bos_token": "<s>",
30
  "clean_up_tokenization_spaces": false,
31
  "eos_token": "</s>",
32
- "legacy": true,
33
- "max_length": 30000,
34
  "model_max_length": 1000000000000000019884624838656,
35
  "pad_to_multiple_of": null,
36
  "pad_token": "</s>",
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
32
  "bos_token": "<s>",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
+ "legacy": false,
36
+ "max_length": 512,
37
  "model_max_length": 1000000000000000019884624838656,
38
  "pad_to_multiple_of": null,
39
  "pad_token": "</s>",