Rexe commited on
Commit
a0028d6
1 Parent(s): e21852f

Finetuning

Browse files
Files changed (3) hide show
  1. README.md +1 -22
  2. adapter_config.json +2 -0
  3. adapter_model.bin +1 -1
README.md CHANGED
@@ -1,27 +1,6 @@
1
  ---
2
  library_name: peft
3
- license: cc-by-nc-2.0
4
- language:
5
- - fr
6
- - en
7
- tags:
8
- - pytorch
9
- - llama
10
- - code
11
  ---
12
-
13
- ## Aria 7B V3
14
-
15
- We decided to build a V3 of Aria 7B based on Mistral instruct instead of LLAMA 2. The base model has been quantized with Qlora to reduce the model size and trained on a high quality french dataset.
16
-
17
- ## Base Model : Mistral-7B-Instruct-v0.1
18
-
19
- ## Technical issues Fixed & Limits of base model
20
-
21
- We noticed that the base model had a common issue of mixing french and english when the request was done in french in some cases,not all of them. This issue was more visible for
22
- prompts over 1000 tokens. By training the base model on our dataset, we fixed this issue and allow the model to reply in the same specific language used for the question to answer.
23
- This pain-point is a valuable upgrade for corporate users in non-english areas willing to deploy a model with an increased quality and accuracy in french language.
24
-
25
  ## Training procedure
26
 
27
 
@@ -39,4 +18,4 @@ The following `bitsandbytes` quantization config was used during training:
39
  ### Framework versions
40
 
41
 
42
- - PEFT 0.5.0
 
1
  ---
2
  library_name: peft
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ## Training procedure
5
 
6
 
 
18
  ### Framework versions
19
 
20
 
21
+ - PEFT 0.6.0.dev0
adapter_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "auto_mapping": null,
3
  "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
4
  "bias": "none",
@@ -12,6 +13,7 @@
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 16,
 
15
  "revision": null,
16
  "target_modules": [
17
  "q_proj",
 
1
  {
2
+ "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
  "bias": "none",
 
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 16,
16
+ "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
  "q_proj",
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dba267b1eaaf009b0c3341630d3a7b28fd734cc30b87ec45d6a72f09421f62f
3
  size 27308941
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d39ab534317fcef11f07327bbd422a5d23e196697a85bccfdb51abd000b5070
3
  size 27308941