deuswoof commited on
Commit
48efb4f
·
1 Parent(s): 4be164b

Training in progress, step 10

Browse files
25_10_23_config_test_3.csv ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop_token,classification_of_valuems,stemming,lemmatization
2
+ 1,no variations,False,2,100,0.8,False,False,False,False
3
+ 2,lemmatization set True,False,2,100,0.8,False,False,False,True
4
+ 3,stemming set True,False,2,100,0.8,False,False,True,False
5
+ 4,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
6
+ 5,lemmatization set True stemming set True classification_of_valuems set True,False,2,100,0.8,False,True,True,True
7
+ 6,lemmatization set True classification_of_valuems set True,False,2,100,0.8,False,True,False,True
8
+ 7,lemmatization set True stemming set True classification_of_valuems set True,False,2,100,0.8,False,True,True,True
25_10_23_results_real_1.csv CHANGED
@@ -31,4 +31,4 @@ classification_of_valuems set True",20.55314040184021,0,0.0214299271266734,0.032
31
  stemming set True
32
  classification_of_valuems set True",29.53177309036255,0,0.0198469239279936,0.0259817116141352,0.0213430019398334,0.0289868657485789,0.0394450508626284,0.0308283988125781,0.0387057818116391,0.0563613255452064,0.0412743876629069,0.0023191133516241,0.0035262639417618,0.0026788789101704,0.0059363771603539,0.009667223107856,0.0068992474882522,0.0108012824018626,0.0183868709697907,0.0124010172661231,0.25,0.25,0.25
33
  27,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
34
- 28,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
 
31
  stemming set True
32
  classification_of_valuems set True",29.53177309036255,0,0.0198469239279936,0.0259817116141352,0.0213430019398334,0.0289868657485789,0.0394450508626284,0.0308283988125781,0.0387057818116391,0.0563613255452064,0.0412743876629069,0.0023191133516241,0.0035262639417618,0.0026788789101704,0.0059363771603539,0.009667223107856,0.0068992474882522,0.0108012824018626,0.0183868709697907,0.0124010172661231,0.25,0.25,0.25
33
  27,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
34
+ 28,1513.510999686485,no variations,17.442886114120483,0,0.2304200812619715,0.2802716481928804,0.2380130314409673,0.2751350231944529,0.3156470566432773,0.2752981663018682,0.3214566715059123,0.3551174721289881,0.3166235943285683,0.0878739353769647,0.0865994538526911,0.0838676015484247,0.1141785741241325,0.1100943210071629,0.1068031729472273,0.1421830338382975,0.1347031023991125,0.1326758730828171,0.5,0.5,0.5
25_10_23_results_real_2.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
+ 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
README.md CHANGED
@@ -712,6 +712,18 @@ The following `bitsandbytes` quantization config was used during training:
712
  - bnb_4bit_use_double_quant: True
713
  - bnb_4bit_compute_dtype: bfloat16
714
 
 
 
 
 
 
 
 
 
 
 
 
 
715
  The following `bitsandbytes` quantization config was used during training:
716
  - quant_method: bitsandbytes
717
  - load_in_8bit: False
@@ -784,5 +796,6 @@ The following `bitsandbytes` quantization config was used during training:
784
  - PEFT 0.5.0
785
  - PEFT 0.5.0
786
  - PEFT 0.5.0
 
787
 
788
  - PEFT 0.5.0
 
712
  - bnb_4bit_use_double_quant: True
713
  - bnb_4bit_compute_dtype: bfloat16
714
 
715
+ The following `bitsandbytes` quantization config was used during training:
716
+ - quant_method: bitsandbytes
717
+ - load_in_8bit: False
718
+ - load_in_4bit: True
719
+ - llm_int8_threshold: 6.0
720
+ - llm_int8_skip_modules: None
721
+ - llm_int8_enable_fp32_cpu_offload: False
722
+ - llm_int8_has_fp16_weight: False
723
+ - bnb_4bit_quant_type: nf4
724
+ - bnb_4bit_use_double_quant: True
725
+ - bnb_4bit_compute_dtype: bfloat16
726
+
727
  The following `bitsandbytes` quantization config was used during training:
728
  - quant_method: bitsandbytes
729
  - load_in_8bit: False
 
796
  - PEFT 0.5.0
797
  - PEFT 0.5.0
798
  - PEFT 0.5.0
799
+ - PEFT 0.5.0
800
 
801
  - PEFT 0.5.0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48458d12acaf19186458cebd42f16601f33238443872049d6b8cfb9b17cef0c4
3
  size 100733709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:694e28bef1367cbb6cebfbc6a53af7f494b6057481dc041d166e0ab56089874f
3
  size 100733709
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b3bf012bf5299471e31d6294748efd8a946e7bb9131064ff26c1d9057ff553b
3
  size 100690288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaa5b373a447766c8ae809d0e4bc892dbab990c48180f40a4f9fec29f94fd1ae
3
  size 100690288
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8056abb5bc119ab9859e0e9fc5ebf30b109fc7b9c6310ce476ab175e3f1b7914
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1e60b9bc95167a4476f9789e703c72923edffbe10e599d7636464152a3b36ea
3
  size 4283