doberst commited on
Commit
78162d0
1 Parent(s): 034e4f3

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -20
README.md CHANGED
@@ -6,47 +6,57 @@ license: apache-2.0
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
- bling-red-pajamas-3b-0.1 part of the BLING ("Best Little Instruction-following No-GPU-required") model series, RAG-instruct trained on top of a RedPajama-INCITE-Base-3B-v1 base model.
10
 
11
- BLING models are fine-tuned with distilled high-quality custom instruct datasets, targeted at a specific subset of instruct tasks with
12
- the objective of providing a high-quality Instruct model that is 'inference-ready' on a CPU laptop even
13
- without using any advanced quantization optimizations.
14
 
15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ### Model Description
17
 
18
  <!-- Provide a longer summary of what this model is. -->
19
 
20
  - **Developed by:** llmware
21
- - **Model type:** GPTNeoX instruct-trained decoder
22
  - **Language(s) (NLP):** English
23
  - **License:** Apache 2.0
24
- - **Finetuned from model:** togethercomputer/RedPajama-INCITE-Base-3B-v1
25
 
26
  ## Uses
27
 
28
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
29
 
30
- The intended use of BLING models is two-fold:
 
 
31
 
32
- 1. Provide high-quality Instruct models that can run on a laptop for local testing. We have found it extremely useful when building a
33
- proof-of-concept, or working with sensitive enterprise data that must be closely guarded, especially in RAG use cases.
34
 
35
- 2. Push the state of the art for smaller Instruct-following models in the sub-7B parameter range, especially 1B-3B, as single-purpose
36
- automation tools for specific tasks through targeted fine-tuning datasets and focused "instruction" tasks.
37
 
38
 
39
  ### Direct Use
40
 
41
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
42
 
43
- BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
44
- legal and regulatory industries with complex information sources. Rather than try to be "all things to all people," BLING models try to focus on a narrower set of Instructions more suitable to a ~1-3B parameter GPT model.
45
-
46
- BLING is ideal for rapid prototyping, testing, and the ability to perform an end-to-end workflow locally on a laptop without
47
- having to send sensitive information over an Internet-based API.
48
 
49
- The first BLING models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
50
  without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
51
 
52
 
@@ -62,8 +72,8 @@ Any model can provide inaccurate or incomplete information, and should be used i
62
  The fastest way to get started with BLING is through direct import in transformers:
63
 
64
  from transformers import AutoTokenizer, AutoModelForCausalLM
65
- tokenizer = AutoTokenizer.from_pretrained("bling-red-pajamas-3b-0.1")
66
- model = AutoModelForCausalLM.from_pretrained("bling-red-pajamas-3b-0.1")
67
 
68
 
69
  The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
@@ -85,7 +95,7 @@ my_prompt = {{text_passage}} + "\n" + {{question/instruction}}
85
 
86
  Darren Oberst & llmware team
87
 
88
- Please reach out anytime if you are interested in this project and would like to participate and work with us!
89
 
90
 
91
 
 
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
+ dragon-llama-7b-0.1 part of the dRAGon ("Delivering RAG On Private Cloud") model series, RAG-instruct trained on top of a LLama-2 base model.
10
 
11
+ DRAGON models are fine-tuned with high-quality custom instruct datasets, designed for production quality use in RAG scenarios.
 
 
12
 
13
 
14
+ ### Benchmark Tests
15
+
16
+ Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
17
+ Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
18
+
19
+ --**Accuracy Score**: **99.0** correct out of 100
20
+ --Not Found Classification: 95.0%
21
+ --Boolean: 82.5%
22
+ --Math/Logic: 70.0%
23
+ --Complex Questions (1-5): 4 (Low-Medium)
24
+ --Summarization Quality (1-5): 4 (Coherent, extractive)
25
+ --Hallucinations: No hallucinations observed in test runs.
26
+
27
+ For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
28
+
29
  ### Model Description
30
 
31
  <!-- Provide a longer summary of what this model is. -->
32
 
33
  - **Developed by:** llmware
34
+ - **Model type:** LLama-2
35
  - **Language(s) (NLP):** English
36
  - **License:** Apache 2.0
37
+ - **Finetuned from model:** Llama-2-7B-Base
38
 
39
  ## Uses
40
 
41
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
 
43
+ The intended use of DRAGON models is two-fold:
44
+
45
+ 1. Provide high-quality RAG-Instruct models designed for fact-based, no "hallucination" question-answering in connection with an enterprise RAG workflow.
46
 
47
+ 2. DRAGON models are fine-tuned on top of leading base foundation models, generally in the 6-7B+ range, and purposefully rolled-out across multiple base models to provide choices and "drop-in" replacements for RAG specific use cases.
 
48
 
49
+ 3. DRAGON models were trained on the same principles as the BLING models, so generally, it should be easy to "upgrade" from a BLING model in testing to a DRAGON model in production.
 
50
 
51
 
52
  ### Direct Use
53
 
54
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
55
 
56
+ DRAGON is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
57
+ legal and regulatory industries with complex information sources.
 
 
 
58
 
59
+ DRAGON models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
60
  without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
61
 
62
 
 
72
  The fastest way to get started with BLING is through direct import in transformers:
73
 
74
  from transformers import AutoTokenizer, AutoModelForCausalLM
75
+ tokenizer = AutoTokenizer.from_pretrained("dragon-llama-7b-0.1")
76
+ model = AutoModelForCausalLM.from_pretrained("dragon-llama-7b-0.1")
77
 
78
 
79
  The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
 
95
 
96
  Darren Oberst & llmware team
97
 
98
+ Please reach out anytime if you are interested in this project!
99
 
100
 
101