acook46 commited on
Commit
712e4c9
·
verified ·
1 Parent(s): e061feb

Added PEFT instructions

Browse files
Files changed (1) hide show
  1. README.md +13 -0
README.md CHANGED
@@ -73,6 +73,19 @@ The training data consists of 100,000 Python functions and their docstrings extr
73
  - **Clarity:** Measures readability using simple, unambiguous language. Calculated using the Flesch-Kincaid readability score.
74
 
75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
  #### Hardware
77
 
78
  Fine-tuning was performed using an Intel 12900K CPU, an Nvidia RTX-3090 GPU, and 64 GB RAM. Total fine-tuning time was 48 GPU hours.
 
73
  - **Clarity:** Measures readability using simple, unambiguous language. Calculated using the Flesch-Kincaid readability score.
74
 
75
 
76
+ ## Model Inference
77
+
78
+ For running inference, PEFT must be used to load the fine-tuned model:
79
+
80
+ ```
81
+ from transformers import AutoModelForCausalLM, AutoTokenizer
82
+ from peft import PeftModel, PeftConfig
83
+
84
+ config = PeftConfig.from_pretrained(self.model_id)
85
+ model = AutoModelForCausalLM.from_pretrained("google/codegemma-2b", device_map = self.device)
86
+ fine_tuned_model = PeftModel.from_pretrained(model, "documint/CodeGemma2B-fine-tuned", device_map = self.device)
87
+ ```
88
+
89
  #### Hardware
90
 
91
  Fine-tuning was performed using an Intel 12900K CPU, an Nvidia RTX-3090 GPU, and 64 GB RAM. Total fine-tuning time was 48 GPU hours.