SwastikM commited on
Commit
5501946
1 Parent(s): 0122800

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -0
README.md CHANGED
@@ -67,6 +67,22 @@ code = tokenizer.decode(outputs[0], skip_special_tokens=True)
67
  print(code)
68
  ```
69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
 
71
  ## Training Details
72
 
 
67
  print(code)
68
  ```
69
 
70
+ ### Size Comparison
71
+
72
+ ```
73
+ The table shows comparison VRAM requirements for loading and training
74
+ of FP16 Base Model and 4bit GPTQ quantized model with PEFT.
75
+ The value for base model referenced from [Model Memory Calculator](https://huggingface.co/docs/accelerate/main/en/usage_guides/model_size_estimator)
76
+ from HuggingFace
77
+ ```
78
+
79
+
80
+
81
+ | Model | Total Size | Training Using Adam |
82
+ | ------------------------|-------------| --------------------|
83
+ | Base Model | 12.37 GB | 49.48 GP |
84
+ | 4bitQuantized+PEFT | 3.90 GB | 11 GB |
85
+
86
 
87
  ## Training Details
88