fionazhang commited on
Commit
b4335bd
1 Parent(s): 162b38e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -17
README.md CHANGED
@@ -10,44 +10,78 @@ model-index:
10
  - name: fine-tune-mistral-environment-merge
11
  results: []
12
  ---
13
-
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
-
17
  # fine-tune-mistral-environment
18
 
19
- This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 2.0377
22
 
23
  ## Model description
24
 
25
- More information needed
26
 
27
- ## Intended uses & limitations
28
 
29
- More information needed
30
 
31
  ## Training and evaluation data
32
 
33
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Training procedure
36
 
37
- ### Training hyperparameters
38
 
39
- The following hyperparameters were used during training:
40
- - learning_rate: 5e-05
41
- - train_batch_size: 3
42
- - eval_batch_size: 3
43
- - seed: 42
 
 
 
 
 
 
 
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
 
 
 
 
 
45
  - lr_scheduler_type: constant
46
- - lr_scheduler_warmup_ratio: 0.03
47
- - num_epochs: 2
48
 
49
  ### Training results
50
 
 
 
 
 
 
 
 
51
 
52
 
53
  ### Framework versions
 
10
  - name: fine-tune-mistral-environment-merge
11
  results: []
12
  ---
 
 
 
 
13
  # fine-tune-mistral-environment
14
 
15
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1).
16
  It achieves the following results on the evaluation set:
17
  - Loss: 2.0377
18
 
19
  ## Model description
20
 
21
+ This model is fine-tuned to specialize in generating content related to the environment and sustainability domain. The training involved Soft Fine-Tuning, Parameter Efficient Fine-Tuning (PEFT), and Low-Rank Adaptation (LoRA) techniques to optimize model performance. The motivation behind this research is to explore the feasibility and effectiveness of Semantically Sufficient Private Large Language Models (LLMs) for secure, domain-specific knowledge extraction in the context of environment and sustainability.
22
 
23
+ ## Intended uses
24
 
25
+ The model is intended for information retrieval and knowledge extraction tasks within the domain of environment and sustainability.
26
 
27
  ## Training and evaluation data
28
 
29
+ The training data consists of domain-specific text collected from Wikipedia pages related to environmental topics.
30
+ This model was trained using the Short dataset. [Model trained with the Long dataset](https://huggingface.co/fionazhang/mistral-finetune-long).
31
+
32
+ | **Dataset** | **URLs** | **Number of Rows** | **Number of Words** | **Number of Sentences** |
33
+ |-------------|----------|--------------------|----------------------|--------------------------|
34
+ | Short | 11 | 577 | 51,526 | 2,150 |
35
+ | Long | 23 | 1,431 | 124,682 | 5,209 |
36
+
37
+ **Table 1:** Summary of Dataset Information
38
+
39
+
40
+ ### Environment and Sustainability
41
+
42
+ This model is tailored for the environment and sustainability domain, with a focus on assisting researchers and enterprises, particularly in alignment with the work of the Commonwealth Scientific and Industrial Research Organisation (CSIRO).
43
+
44
+ ### Data Collection Process
45
+
46
+ The training data was collected through a Python program that extracted and cleaned text content from specific Wikipedia pages related to environmental topics. The program utilized various libraries, such as `requests`, `BeautifulSoup`, and `nltk`, for efficient web scraping, HTML parsing, and natural language processing.
47
 
48
  ## Training procedure
49
 
50
+ ## Fine-tuning
51
 
52
+ The fine-tuning process involved Soft Fine-Tuning, PEFT, and LoRA techniques. Soft Fine-Tuning utilized continuous-valued probabilities as labels, suitable for generation models. PEFT focused on updating a small subset of parameters during fine-tuning to prevent catastrophic forgetting. LoRA, a lightweight training technique, reduced the number of trainable parameters for faster and memory-efficient training.
53
+
54
+ #### Low-Rank Adaptation (LoRA) Parameters
55
+
56
+ - lora_alpha: 16
57
+ - lora_dropout: 0.1
58
+ - r: 8
59
+
60
+ #### Training Parameters
61
+
62
+ - num_train_epochs: 2
63
+ - per_device_train_batch_size: 3
64
+ - per_device_eval_batch_size: 3
65
+ - gradient_accumulation_steps: 1
66
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
67
+ - learning_rate: 5e-05
68
+ - weight_decay: 0.001
69
+ - max_grad_norm: 0.3
70
+ - max_steps: -1
71
+ - warmup_ratio: 0.03
72
+ - group_by_length: True
73
  - lr_scheduler_type: constant
74
+ - seed: 42
 
75
 
76
  ### Training results
77
 
78
+ #### Training Loss
79
+
80
+
81
+ ![Training Loss - Short]
82
+ *Figure 1: Training loss curve of model fionazhang/fine-tune-short (logging step = 10)*
83
+
84
+ In the training process, the observed training losses exhibit jittery yet overall decreasing trends. The final evaluation loss reaches a satisfactory value of 1.8378, indicating successful learning and adaptation to the nuances of the provided data.
85
 
86
 
87
  ### Framework versions