gnumanth commited on
Commit
2dcf7da
1 Parent(s): 547c0a8

chore: more info

Browse files
Files changed (1) hide show
  1. README.md +14 -17
README.md CHANGED
@@ -1,35 +1,30 @@
1
  ---
2
- license: other
3
  base_model: facebook/opt-350m
4
  tags:
5
  - trl
6
  - sft
7
- - generated_from_trainer
8
  model-index:
9
  - name: tmp_trainer
10
  results: []
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- # tmp_trainer
17
 
18
- This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on an unknown dataset.
19
 
20
  ## Model description
21
 
22
- More information needed
23
-
24
- ## Intended uses & limitations
25
-
26
- More information needed
27
-
28
- ## Training and evaluation data
29
-
30
- More information needed
31
-
32
- ## Training procedure
33
 
34
  ### Training hyperparameters
35
 
@@ -44,11 +39,13 @@ The following hyperparameters were used during training:
44
 
45
  ### Training results
46
 
47
-
 
 
48
 
49
  ### Framework versions
50
 
51
  - Transformers 4.38.1
52
  - Pytorch 2.1.0+cu121
53
  - Datasets 2.17.1
54
- - Tokenizers 0.15.1
 
1
  ---
2
+ license: mit
3
  base_model: facebook/opt-350m
4
  tags:
5
  - trl
6
  - sft
7
+ - gnumanth/dadjokes-trained-otp
8
  model-index:
9
  - name: tmp_trainer
10
  results: []
11
+ datasets:
12
+ - gnumanth/dad-jokes
13
+ language:
14
+ - en
15
+ pipeline_tag: text-generation
16
  ---
17
 
18
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
  should probably proofread and complete it, then remove this comment. -->
20
 
21
+ #
22
 
23
+ This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on an [gnumanth/dad-jokes](https://huggingface.co/datasets/gnumanth/dad-jokes)dataset.
24
 
25
  ## Model description
26
 
27
+ SFT Trained simple model for fun!
 
 
 
 
 
 
 
 
 
 
28
 
29
  ### Training hyperparameters
30
 
 
39
 
40
  ### Training results
41
 
42
+ ```
43
+ TrainOutput(global_step=18, training_loss=2.2378472222222223, metrics={'train_runtime': 149.7511, 'train_samples_per_second': 0.881, 'train_steps_per_second': 0.12, 'total_flos': 9828797644800.0, 'train_loss': 2.2378472222222223, 'epoch': 3.0})
44
+ ```
45
 
46
  ### Framework versions
47
 
48
  - Transformers 4.38.1
49
  - Pytorch 2.1.0+cu121
50
  - Datasets 2.17.1
51
+ - Tokenizers 0.15.1