xianchaowu
commited on
Commit
•
1ae6855
1
Parent(s):
cd0f47e
13b model
Browse files
usage.py
CHANGED
@@ -13,7 +13,7 @@ import torch
|
|
13 |
|
14 |
#import ipdb; ipdb.set_trace()
|
15 |
cache_dir="/workspace/asr/peft/qlora"
|
16 |
-
# TODO set this cache_dir to the path where you stored (or, want to store) llama2-
|
17 |
|
18 |
lazylora_dir=os.getcwd() # the path that contains 'adapter_config.json' and 'adapter_model.bin'
|
19 |
|
@@ -41,7 +41,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
41 |
)
|
42 |
#model.print_trainable_parameters()
|
43 |
print(sum(p.numel() for p in model.parameters()))
|
44 |
-
# 6,671,979,520 -> half-size of
|
45 |
|
46 |
model = PeftModel.from_pretrained(model, lazylora_dir)
|
47 |
print('after adding lazy lora parameters:')
|
|
|
13 |
|
14 |
#import ipdb; ipdb.set_trace()
|
15 |
cache_dir="/workspace/asr/peft/qlora"
|
16 |
+
# TODO set this cache_dir to the path where you stored (or, want to store) llama2-13b-hf model
|
17 |
|
18 |
lazylora_dir=os.getcwd() # the path that contains 'adapter_config.json' and 'adapter_model.bin'
|
19 |
|
|
|
41 |
)
|
42 |
#model.print_trainable_parameters()
|
43 |
print(sum(p.numel() for p in model.parameters()))
|
44 |
+
# 6,671,979,520 -> half-size of 13B due to 4-bit loading
|
45 |
|
46 |
model = PeftModel.from_pretrained(model, lazylora_dir)
|
47 |
print('after adding lazy lora parameters:')
|