atsuki-yamaguchi commited on
Commit
9dd898f
1 Parent(s): 4e1c59b

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +33 -32
README.md CHANGED
@@ -1,38 +1,39 @@
1
  ---
2
- library_name: peft
 
3
  ---
4
- ## Training procedure
 
5
 
 
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - _load_in_8bit: True
10
- - _load_in_4bit: False
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: fp4
16
- - bnb_4bit_use_double_quant: False
17
- - bnb_4bit_compute_dtype: float32
18
- - load_in_4bit: False
19
- - load_in_8bit: True
20
 
21
- The following `bitsandbytes` quantization config was used during training:
22
- - quant_method: bitsandbytes
23
- - _load_in_8bit: True
24
- - _load_in_4bit: False
25
- - llm_int8_threshold: 6.0
26
- - llm_int8_skip_modules: None
27
- - llm_int8_enable_fp32_cpu_offload: False
28
- - llm_int8_has_fp16_weight: False
29
- - bnb_4bit_quant_type: fp4
30
- - bnb_4bit_use_double_quant: False
31
- - bnb_4bit_compute_dtype: float32
32
- - load_in_4bit: False
33
- - load_in_8bit: True
34
- ### Framework versions
 
 
 
 
35
 
36
- - PEFT 0.5.0
37
-
38
- - PEFT 0.5.0
 
1
  ---
2
+ license: mit
3
+ language: ja
4
  ---
5
+ Mistral-7B Japanese [LAPT + Random]
6
+ ===
7
 
8
+ ## How to use
9
+ ```python
10
+ from peft import AutoPeftModelForCausalLM
11
 
12
+ model = AutoPeftModelForCausalLM.from_pretrained(
13
+ "atsuki-yamaguchi/Mistral-7B-v0.1-random-ja"
14
+ )
15
+ tokenizer = AutoTokenizer.from_pretrained(
16
+ "atsuki-yamaguchi/Mistral-7B-v0.1-random-ja"
17
+ )
 
 
 
 
 
 
 
18
 
19
+ # w/ GPU
20
+ model = AutoPeftModelForCausalLM.from_pretrained(
21
+ "atsuki-yamaguchi/Mistral-7B-v0.1-random-ja",
22
+ device_map="auto",
23
+ load_in_8bit=True,
24
+ )
25
+ ```
26
+ ## Citation
27
+ ```
28
+ @article{yamaguchi2024empirical,
29
+ title={An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient Generative {LLM} Inference},
30
+ author={Atsuki Yamaguchi and Aline Villavicencio and Nikolaos Aletras},
31
+ journal={ArXiv},
32
+ year={2024},
33
+ volume={abs/2402.10712},
34
+ url={https://arxiv.org/abs/2402.10712}
35
+ }
36
+ ```
37
 
38
+ ## Link
39
+ For more details, please visit https://github.com/gucci-j/llm-cva