Tatiana239 commited on
Commit
4e022f7
·
1 Parent(s): 98d27b3

End of training

Browse files
Files changed (1) hide show
  1. README.md +3 -25
README.md CHANGED
@@ -30,33 +30,11 @@ More information needed
30
 
31
  ## Training procedure
32
 
33
-
34
- The following `bitsandbytes` quantization config was used during training:
35
- - quant_method: gptq
36
- - bits: 4
37
- - tokenizer: None
38
- - dataset: None
39
- - group_size: 128
40
- - damp_percent: 0.1
41
- - desc_act: True
42
- - sym: True
43
- - true_sequential: True
44
- - use_cuda_fp16: False
45
- - model_seqlen: 4096
46
- - block_name_to_quantize: model.layers
47
- - module_name_preceding_first_block: ['model.embed_tokens']
48
- - batch_size: 1
49
- - pad_token_id: None
50
- - use_exllama: False
51
- - max_input_length: None
52
- - exllama_config: {'version': <ExllamaVersion.ONE: 1>}
53
- - cache_block_outputs: True
54
-
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
  - learning_rate: 0.0002
59
- - train_batch_size: 8
60
  - eval_batch_size: 8
61
  - seed: 42
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -70,8 +48,8 @@ The following hyperparameters were used during training:
70
 
71
  ### Framework versions
72
 
73
- - PEFT 0.7.0
74
- - Transformers 4.36.0
75
  - Pytorch 2.1.0+cu118
76
  - Datasets 2.15.0
77
  - Tokenizers 0.15.0
 
30
 
31
  ## Training procedure
32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
+ - train_batch_size: 4
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
48
 
49
  ### Framework versions
50
 
51
+ - PEFT 0.7.1
52
+ - Transformers 4.36.1
53
  - Pytorch 2.1.0+cu118
54
  - Datasets 2.15.0
55
  - Tokenizers 0.15.0