atsuki-yamaguchi commited on
Commit
5c6aaa4
1 Parent(s): d107b22

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +34 -17
README.md CHANGED
@@ -1,23 +1,40 @@
1
  ---
2
- library_name: peft
 
3
  ---
4
- ## Training procedure
 
5
 
 
 
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - _load_in_8bit: True
10
- - _load_in_4bit: False
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: fp4
16
- - bnb_4bit_use_double_quant: False
17
- - bnb_4bit_compute_dtype: float32
18
- - load_in_4bit: False
19
- - load_in_8bit: True
20
- ### Framework versions
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
- - PEFT 0.5.0
 
 
1
  ---
2
+ license: mit
3
+ language: ar
4
  ---
5
+ Mistral-7B Arabic [LAPT + Heuristics (Untied)]
6
+ ===
7
 
8
+ ## How to use
9
+ ```python
10
+ from peft import AutoPeftModelForCausalLM
11
+ from transformers import AutoTokenizer
12
 
13
+ model = AutoPeftModelForCausalLM.from_pretrained(
14
+ "atsuki-yamaguchi/Mistral-7B-v0.1-heuristics-untied-ar"
15
+ )
16
+ tokenizer = AutoTokenizer.from_pretrained(
17
+ "aubmindlab/aragpt2-base"
18
+ )
 
 
 
 
 
 
 
 
19
 
20
+ # w/ GPU
21
+ model = AutoPeftModelForCausalLM.from_pretrained(
22
+ "atsuki-yamaguchi/Mistral-7B-v0.1-heuristics-untied-ar",
23
+ device_map="auto",
24
+ load_in_8bit=True,
25
+ )
26
+ ```
27
+ ## Citation
28
+ ```
29
+ @article{yamaguchi2024empirical,
30
+ title={An Empirical Study on Cross-lingual Vocabulary Adaptation for Efficient Generative {LLM} Inference},
31
+ author={Atsuki Yamaguchi and Aline Villavicencio and Nikolaos Aletras},
32
+ journal={ArXiv},
33
+ year={2024},
34
+ volume={abs/2402.10712},
35
+ url={https://arxiv.org/abs/2402.10712}
36
+ }
37
+ ```
38
 
39
+ ## Link
40
+ For more details, please visit https://github.com/gucci-j/llm-cva