yahma commited on
Commit
b198b99
·
1 Parent(s): 54700d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -3,12 +3,12 @@ license: mit
3
  datasets:
4
  - yahma/alpaca-cleaned
5
  ---
6
- This repo contains a low-rank adapter for LLaMA-7b fit on the Cleaned Alpaca dataset.
7
 
8
  This version of the weights was trained with the following hyperparameters:
9
 
10
- Cleaned dataset: Snapshot April 2, 2023
11
- Epochs: 3
12
  Validation set size: 1500
13
  Batch size: 128
14
  Micro batch size: 8
@@ -22,7 +22,7 @@ That is:
22
  python finetune.py \
23
  --base_model='decapoda-research/llama-7b-hf' \
24
  --data_path 'yahma/alpaca-cleaned' \
25
- --num_epochs=3 \
26
  --cutoff_len=512 \
27
  --output_dir='./lora-alpaca' \
28
  --lora_target_modules='[q_proj,k_proj, v_proj, o_proj]' \
 
3
  datasets:
4
  - yahma/alpaca-cleaned
5
  ---
6
+ This repo contains a low-rank adapter for LLaMA-7b fit on the Cleaned Alpaca dataset (with the new GPT-4 training data).
7
 
8
  This version of the weights was trained with the following hyperparameters:
9
 
10
+ Cleaned dataset: Snapshot April 8, 2023
11
+ Epochs: 6 (Checkpoint with lowest eval loss at 3.6 epochs uploaded here)
12
  Validation set size: 1500
13
  Batch size: 128
14
  Micro batch size: 8
 
22
  python finetune.py \
23
  --base_model='decapoda-research/llama-7b-hf' \
24
  --data_path 'yahma/alpaca-cleaned' \
25
+ --num_epochs=6 \
26
  --cutoff_len=512 \
27
  --output_dir='./lora-alpaca' \
28
  --lora_target_modules='[q_proj,k_proj, v_proj, o_proj]' \