abhi9ab commited on
Commit
3ce6b4a
·
verified ·
1 Parent(s): 7539c18

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit
3
  tags:
4
  - text-generation-inference
5
  - transformers
@@ -10,14 +10,94 @@ tags:
10
  license: apache-2.0
11
  language:
12
  - en
 
 
13
  ---
14
 
15
  # Uploaded model
16
 
17
  - **Developed by:** abhi9ab
18
  - **License:** apache-2.0
19
- - **Finetuned from model :** unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit
20
 
21
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
22
 
23
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: unsloth/DeepSeek-R1-Distill-Llama-8B
3
  tags:
4
  - text-generation-inference
5
  - transformers
 
10
  license: apache-2.0
11
  language:
12
  - en
13
+ datasets:
14
+ - Josephgflowers/Finance-Instruct-500k
15
  ---
16
 
17
  # Uploaded model
18
 
19
  - **Developed by:** abhi9ab
20
  - **License:** apache-2.0
21
+ - **Finetuned from model :** unsloth/DeepSeek-R1-Distill-Llama-8B
22
 
23
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
24
 
25
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
26
+
27
+ ---
28
+
29
+ # Model Card
30
+
31
+ The goal of this model is to enhance the base model's performance on financial tasks by fine-tuning it on a specialized financial dataset. Using LoRA, this model has been optimized for low-rank adaptation, allowing efficient fine-tuning with fewer resources.
32
+
33
+ ---
34
+
35
+ ## Model Details
36
+
37
+ - Base Model: unsloth/DeepSeek-R1-Distill-Llama-8B
38
+ - Model Type: Language Model (Distilled)
39
+ - Fine-Tuning Technique: LoRA (Low-Rank Adaptation)
40
+ - Fine-Tuned Model: DeepSeek-R1-Distill-Llama-8B-finance-v1
41
+ - Dataset: [Josephgflowers/Finance-Instruct-500k](https://huggingface.co/datasets/Josephgflowers/Finance-Instruct-500k) (reduced to 5k JSONL entries)
42
+ - Platform: Free-tier Kaggle Notebook
43
+ - Library: Hugging Face Transformers, Unsloth and Pytorch
44
+
45
+ This model is a fine-tuned version of the [unsloth/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B), utilizing LoRA for efficient parameter adaptation. It has been specifically tuned on a reduced version (5k) of the [Josephgflowers/Finance-Instruct-500k](https://huggingface.co/datasets/Josephgflowers/Finance-Instruct-500k) dataset to enhance performance in finance-related tasks.
46
+
47
+ ---
48
+
49
+ ## Intended Use
50
+
51
+ The model is intended for tasks related to financial question answering, generation, and instructions that require domain-specific knowledge in finance. It can also be used in other natural language understanding and generation tasks that benefit from fine-tuning on a finance-specific dataset.
52
+
53
+ ---
54
+
55
+ ## Dataset
56
+
57
+ The model was fine-tuned on a subset of the Finance-Instruct-500k dataset from Hugging Face, specifically reduced to 5,000 JSONL entries for the fine-tuning process. This dataset contains financial questions and answers, providing a rich set of examples for training the model.
58
+
59
+ ---
60
+
61
+ ## Training Data
62
+
63
+ - Dataset Name: [Josephgflowers/Finance-Instruct-500k](https://huggingface.co/datasets/Josephgflowers/Finance-Instruct-500k)
64
+ - Data Size: 5k samples (subset from original dataset)
65
+ - Domain: Finance
66
+ - Task: Instruction-based fine-tuning for financial information retrieval and generation.
67
+
68
+ ---
69
+
70
+ ## Notes
71
+
72
+ - This fine-tuning was performed on the free-tier of Kaggle Notebook, so training time and available resources are limited.
73
+ - Ensure that your runtime in Colab/Kaggle is set to a GPU environment to speed up the training process.
74
+ - The reduced 5k dataset is a smaller sample for experimentation. You can scale this up depending on your needs and available resources.
75
+
76
+ ---
77
+
78
+ ## Performance
79
+
80
+ The model performs well in financial instruction tasks, delivering accurate responses based on the reduced dataset. Performance can be further evaluated through specific finance-related benchmarks.
81
+
82
+ ---
83
+
84
+ ## Usage
85
+
86
+ ```bash
87
+ from transformers import AutoTokenizer, AutoModelForCausalLM
88
+
89
+ tokenizer = AutoTokenizer.from_pretrained("abhi9ab/DeepSeek-R1-Distill-Llama-8B-finance-v1")
90
+ model = AutoModelForCausalLM.from_pretrained("abhi9ab/DeepSeek-R1-Distill-Llama-8B-finance-v1")
91
+
92
+ inputs = tokenizer("Example finance-related query", return_tensors="pt")
93
+ outputs = model.generate(inputs['input_ids'])
94
+ ```
95
+
96
+ ---
97
+
98
+ ## Acknowledgement
99
+
100
+ - Josephgflowers for the dataset.
101
+ - Hugging Face Transformers library for model implementation and Unsloth for LoRA-based fine-tuning.
102
+
103
+ ---