chew commited on
Commit
0953aa7
·
1 Parent(s): 303ab2d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -13
README.md CHANGED
@@ -1,20 +1,43 @@
1
  ---
2
  library_name: peft
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
- ## Training procedure
5
 
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: nf4
15
- - bnb_4bit_use_double_quant: False
16
- - bnb_4bit_compute_dtype: float16
17
- ### Framework versions
18
 
 
 
 
 
 
19
 
20
- - PEFT 0.5.0.dev0
 
 
 
 
 
 
 
 
 
1
  ---
2
  library_name: peft
3
+ datasets:
4
+ - FinchResearch/TexTrend-llama2
5
+ language:
6
+ - en
7
+ - de
8
+ - es
9
+ metrics:
10
+ - accuracy
11
+ tags:
12
+ - language
13
+ - llama
14
+ - question-answering
15
+ - instruct
16
  ---
17
+ # Model Card: Seal
18
 
19
+ ## Overview
20
+ The "Seal" model is a novel language model built on top of Meta's LLAMA-2 architecture. This model has undergone a unique training process, combining fine-tuning techniques, model weight merging, and the application of adapters, resulting in an innovative adaptation while retaining learned information from fine-tuned models. The "Seal" model's development was made possible through the incorporation of the Open Platypus methodology, which played a critical role in its creation.
21
 
22
+ ## Model Details
23
+ - Model Name: Seal
24
+ - Architecture: Meta's LLAMA-2
25
+ - Training Approach: Fine-tuning with the LORA framework, model weight merging, adapter-based adaptation
26
+ - Development Methodology: Open Platypus
27
+ - Contributors: Your Name and Your Team
 
 
 
 
 
28
 
29
+ ## Training Process
30
+ The "Seal" model was trained through a multi-stage process aimed at maximizing its performance and adaptability:
31
+ 1. **Fine-Tuning:** The base model (Meta's LLAMA-2) was fine-tuned using the TextTrend Corpus dataset. This initial phase helped the model learn language patterns and semantic understanding from diverse real-time text data.
32
+ 2. **Model Weight Merging:** We merged the fine-tuned model weights with pre-trained adapters, effectively integrating the knowledge acquired during fine-tuning with the broader linguistic context of the adapters.
33
+ 3. **Adapter-Based Adaptation:** Adapters were utilized to modify and enhance specific linguistic capabilities without losing the knowledge gained from the fine-tuned model. This approach allowed for targeted improvements while maintaining the general language understanding.
34
 
35
+ ## Usage and Applications
36
+ The "Seal" model is designed to excel in various natural language processing tasks, including text generation, sentiment analysis, named entity recognition, and more. Its unique training process and incorporation of the Open Platypus methodology make it particularly well-suited for tasks that require a blend of real-time language trends and established linguistic patterns.
37
+
38
+ ## Limitations
39
+ - While the "Seal" model demonstrates enhanced linguistic capabilities, it may still exhibit biases or limitations present in the training data.
40
+ - The effectiveness of the model may vary depending on the specific task and data distribution.
41
+
42
+ ## License
43
+ The "Seal" model is released under a permissive license, encouraging its widespread use and experimentation. Refer to the accompanying license documentation for specific details.