joshuaspear commited on
Commit
92a5a66
·
verified ·
1 Parent(s): 8ae05bd

Model save

Browse files
Files changed (1) hide show
  1. README.md +4 -9
README.md CHANGED
@@ -12,10 +12,7 @@ model-index:
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jsphd/week10_tutorial_llm/runs/wrln9zk3)
16
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jsphd/week10_tutorial_llm/runs/m0jvvs1u)
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jsphd/week10_tutorial_llm/runs/m0jvvs1u)
18
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jsphd/week10_tutorial_llm/runs/42z6s3gq)
19
  # bloom-560m-finetuned-health-qa
20
 
21
  This model is a fine-tuned version of [bigscience/bloom-560m](https://huggingface.co/bigscience/bloom-560m) on an unknown dataset.
@@ -37,12 +34,10 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 1e-05
41
- - train_batch_size: 2
42
- - eval_batch_size: 2
43
  - seed: 42
44
- - gradient_accumulation_steps: 2
45
- - total_train_batch_size: 4
46
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - num_epochs: 5
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jsphd/week10_tutorial_llm/runs/515khe3l)
 
 
 
16
  # bloom-560m-finetuned-health-qa
17
 
18
  This model is a fine-tuned version of [bigscience/bloom-560m](https://huggingface.co/bigscience/bloom-560m) on an unknown dataset.
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
+ - learning_rate: 3e-05
38
+ - train_batch_size: 1
39
+ - eval_batch_size: 1
40
  - seed: 42
 
 
41
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
42
  - lr_scheduler_type: linear
43
  - num_epochs: 5