ljcnju commited on
Commit
e576c46
1 Parent(s): 1e186ba

Training in progress, epoch 1

Browse files
Files changed (3) hide show
  1. config.json +1 -7
  2. model.safetensors +2 -2
  3. training_args.bin +3 -0
config.json CHANGED
@@ -14,15 +14,9 @@
14
  "eos_token_id": 2,
15
  "feed_forward_proj": "relu",
16
  "gradient_checkpointing": false,
17
- "id2label": {
18
- "0": "LABEL_0"
19
- },
20
  "initializer_factor": 1.0,
21
  "is_encoder_decoder": true,
22
  "is_gated_act": false,
23
- "label2id": {
24
- "LABEL_0": 0
25
- },
26
  "layer_norm_epsilon": 1e-06,
27
  "model_type": "t5",
28
  "n_positions": 512,
@@ -31,7 +25,7 @@
31
  "num_layers": 12,
32
  "output_past": true,
33
  "pad_token_id": 0,
34
- "problem_type": "regression",
35
  "relative_attention_max_distance": 128,
36
  "relative_attention_num_buckets": 32,
37
  "task_specific_params": {
 
14
  "eos_token_id": 2,
15
  "feed_forward_proj": "relu",
16
  "gradient_checkpointing": false,
 
 
 
17
  "initializer_factor": 1.0,
18
  "is_encoder_decoder": true,
19
  "is_gated_act": false,
 
 
 
20
  "layer_norm_epsilon": 1e-06,
21
  "model_type": "t5",
22
  "n_positions": 512,
 
25
  "num_layers": 12,
26
  "output_past": true,
27
  "pad_token_id": 0,
28
+ "problem_type": "single_label_classification",
29
  "relative_attention_max_distance": 128,
30
  "relative_attention_num_buckets": 32,
31
  "task_specific_params": {
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f142836eddf82de8aed6612981758a4b763ef4fac4e7d07dd7a9183243dd561b
3
- size 893927620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:753c8b2dda69290a87058d8046f542c66cde7550b985c3a78c34ac3aadafbd3f
3
+ size 893930696
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21c1bbc3a6f290d742ec025685d36c98987309745cef300addf37ebda83545cc
3
+ size 4728