01GangaPutraBheeshma commited on
Commit
b58d68b
1 Parent(s): 27ff0cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -3
README.md CHANGED
@@ -41,8 +41,38 @@ test_tokenizer_UT = AutoTokenizer.from_pretrained("01GangaPutraBheeshma/colab_co
41
  # Documentation
42
 
43
  This model was fine-tuned using LoRA because I wanted the model's weights to be efficient in solving other types of Python problems(Ones that were not included in the training data).
44
- Setting lora_alpha to 16 suggests that I chose a relatively strong regularization. The specific value of this hyperparameter often requires experimentation and tuning to find the optimal balance between preventing overfitting and allowing the model to capture important patterns in the data.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
- The lora_dropout rate is 0.1, which dropped out 10% of the neurons randomly during training. This helps to prevent overfitting by introducing a level of randomness and redundancy in the network.
47
- 'r' in LoRa represents a rank which helps to decide the level of representation of the model in terms of number of dimensions or features. This proved to be advantageous for tasks like fine-tuning, where reducing the complexity of the model while preserving information is paramount.
48
 
 
41
  # Documentation
42
 
43
  This model was fine-tuned using LoRA because I wanted the model's weights to be efficient in solving other types of Python problems(Ones that were not included in the training data).
44
+ Setting lora_alpha to 16 suggests a relatively strong regularization. The specific value of this hyperparameter often requires experimentation and tuning to find the optimal balance between preventing overfitting and allowing the model to capture important patterns in the data.
45
+
46
+ The lora_dropout rate is 0.1, which dropped out 10% of the neurons randomly during training. This helped to prevent overfitting by introducing a level of randomness and redundancy in the network.
47
+ 'r' in LoRa represents a rank that helps to decide the level of representation of the model in terms of a number of dimensions or features. This proved to be advantageous for tasks like fine-tuning, where we witness a reduction in the complexity of the model while preserving information is our paramount goal.
48
+
49
+ I am using bitsAndBytesConfig by loading the main model in 4 bits, as I wanted something to be trained quickly and be efficient rather than being super precise with its results. This tradeoff was needed due to the cluster that I am involved in working with.
50
+ There is a use of double quantization for the 4-bit representation. Quantization is a process of mapping a range of values to a smaller set of discrete values. "Double quantization" here implies an additional refinement or quantization step, possibly to improve the precision of the representation within the constraints of 4-bit storage.
51
+
52
+ The Datatype involved during the computational steps involving 4-bit representation is "float16". Using floating-point numbers allows for more precision in mathematical operations in comparision to integer representations.
53
+
54
+
55
+ ### Lora Config
56
+ '''
57
+ peft_config = LoraConfig(
58
+ lora_alpha=16,
59
+ lora_dropout=0.1,
60
+ r=64,
61
+ bias="none",
62
+ task_type="CAUSAL_LM",
63
+ )
64
+ '''
65
+
66
+ ### BitsAndBytesConfig
67
+ ```
68
+ bnb_config = BitsAndBytesConfig(
69
+ load_in_4bit=True,
70
+ bnb_4bit_use_double_quant=True,
71
+ bnb_4bit_quant_type="nf4",
72
+ bnb_4bit_compute_dtype="float16"
73
+ )
74
+
75
+ ```
76
+
77
 
 
 
78