bitext-innovations
commited on
Commit
•
066acb1
1
Parent(s):
4cf2e21
Update README.md
Browse files
README.md
CHANGED
@@ -3,36 +3,44 @@ base_model: ai21labs/AI21-Jamba-1.5-Mini
|
|
3 |
datasets:
|
4 |
- bitext/Bitext-combined-banking-wealth_management-mortgage_loans
|
5 |
license: other
|
|
|
|
|
|
|
|
|
6 |
tags:
|
7 |
- alignment-handbook
|
8 |
- trl
|
9 |
- sft
|
|
|
10 |
- generated_from_trainer
|
|
|
|
|
11 |
model-index:
|
12 |
- name: Bitext-Jamba-1.5-Mini-banking-customer-support
|
13 |
results: []
|
|
|
|
|
|
|
|
|
|
|
14 |
---
|
15 |
|
16 |
-
|
17 |
-
should probably proofread and complete it, then remove this comment. -->
|
18 |
|
19 |
-
|
20 |
|
21 |
-
This model is
|
22 |
-
|
23 |
-
-
|
24 |
|
25 |
-
##
|
26 |
|
27 |
-
|
|
|
28 |
|
29 |
-
##
|
30 |
|
31 |
-
|
32 |
-
|
33 |
-
## Training and evaluation data
|
34 |
-
|
35 |
-
More information needed
|
36 |
|
37 |
## Training procedure
|
38 |
|
@@ -65,4 +73,21 @@ The following hyperparameters were used during training:
|
|
65 |
- Transformers 4.45.0.dev0
|
66 |
- Pytorch 2.1.0+cu118
|
67 |
- Datasets 3.1.0
|
68 |
-
- Tokenizers 0.19.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
datasets:
|
4 |
- bitext/Bitext-combined-banking-wealth_management-mortgage_loans
|
5 |
license: other
|
6 |
+
license_name: jamba-open-model-license
|
7 |
+
license_link: https://www.ai21.com/licenses/jamba-open-model-license
|
8 |
+
language:
|
9 |
+
- en
|
10 |
tags:
|
11 |
- alignment-handbook
|
12 |
- trl
|
13 |
- sft
|
14 |
+
- jamba
|
15 |
- generated_from_trainer
|
16 |
+
- text-generation-inference
|
17 |
+
library_name: transformers
|
18 |
model-index:
|
19 |
- name: Bitext-Jamba-1.5-Mini-banking-customer-support
|
20 |
results: []
|
21 |
+
pipeline_tag: text-generation
|
22 |
+
widget:
|
23 |
+
- messages:
|
24 |
+
- role: user
|
25 |
+
content: I want open an account
|
26 |
---
|
27 |
|
28 |
+
# Bitext-Jamba-1.5-Mini-Banking-Customer-Support
|
|
|
29 |
|
30 |
+
## Model Description
|
31 |
|
32 |
+
This model is version of [ai21labs/AI21-Jamba-1.5-Mini](https://huggingface.co/ai21labs/AI21-Jamba-1.5-Mini) fine-tuned on the [Bitext Banking Customer Support Dataset](bitext/Bitext-combined-banking-wealth_management-mortgage_loans) dataset, which is specifically tailored for the Banking domain. It is optimized to answer questions and assist users with various banking transactions. It has been trained using hybrid synthetic data generated using our NLP/NLG technology and our automated Data Labeling (DAL) tools.
|
33 |
+
|
34 |
+
The goal of this model is to show that a generic verticalized model makes customization for a final use case much easier. For example, if you are "ACME Bank", you can create your own customized model by using this fine-tuned model and doing an additional fine-tuning using a small amount of your own data. An overview of this approach can be found at: [From General-Purpose LLMs to Verticalized Enterprise Models](https://www.bitext.com/blog/general-purpose-models-verticalized-enterprise-genai/)
|
35 |
|
36 |
+
## Intended Use
|
37 |
|
38 |
+
- **Recommended applications**: This model is designed to be used as the first step in Bitext’s two-step approach to LLM fine-tuning for the creation of chatbots, virtual assistants and copilots for the Banking domain, providing customers with fast and accurate answers about their banking needs.
|
39 |
+
- **Out-of-scope**: This model is not suited for non-banking related questions and should not be used for providing health, legal, or critical safety advice.
|
40 |
|
41 |
+
## Training Data
|
42 |
|
43 |
+
The model was fine-tuned on a dataset comprising various banking-related intents, including transactions like balance checks, money transfers, loan applications, and more, totaling 89 intents each represented by approximately 1000 examples. This comprehensive training helps the model address a broad spectrum of banking-related questions effectively. The dataset follows the same structured approach as our dataset published on Hugging Face as [bitext/Bitext-customer-support-llm-chatbot-training-dataset](https://huggingface.co/datasets/bitext/Bitext-customer-support-llm-chatbot-training-dataset), but with a focus on banking.
|
|
|
|
|
|
|
|
|
44 |
|
45 |
## Training procedure
|
46 |
|
|
|
73 |
- Transformers 4.45.0.dev0
|
74 |
- Pytorch 2.1.0+cu118
|
75 |
- Datasets 3.1.0
|
76 |
+
- Tokenizers 0.19.1
|
77 |
+
|
78 |
+
## Limitations and Bias
|
79 |
+
|
80 |
+
- The model is trained for banking-specific contexts but may underperform in unrelated areas.
|
81 |
+
- Potential biases in the training data could affect the neutrality of the responses; users are encouraged to evaluate responses critically.
|
82 |
+
|
83 |
+
## Ethical Considerations
|
84 |
+
|
85 |
+
It is important to use this technology thoughtfully, ensuring it does not substitute for human judgment where necessary, especially in sensitive financial situations.
|
86 |
+
|
87 |
+
## Acknowledgments
|
88 |
+
|
89 |
+
This model was developed and trained by Bitext using proprietary data and technology.
|
90 |
+
|
91 |
+
## License
|
92 |
+
|
93 |
+
This model, "Bitext-Jamba-1.5-Mini-Banking-Customer-Support", is licensed under the [Jamba Open Model License](https://www.ai21.com/licenses/jamba-open-model-license), a permissive license allowing full research use and commercial use under the license terms. If you need to license the model for your needs, [talk to us](https://www.bitext.com/contact-us/).
|