doberst commited on
Commit
b7d275e
·
verified ·
1 Parent(s): 2c7e033

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -16
README.md CHANGED
@@ -3,13 +3,13 @@ license: apache-2.0
3
  inference: false
4
  ---
5
 
6
- # DRAGON-QWEN-7B
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- dragon-qwen-7b is part of the dRAGon ("Delivering RAG On ...") model series, RAG-instruct trained on top of a Qwen2 7b base model.
11
 
12
- DRAGON models have been fine-tuned with the specific objective of fact-based question-answering over complex business and legal documents with an emphasis on reducing hallucinations and providing short, clear answers for workflow automation.
13
 
14
 
15
  ### Benchmark Tests
@@ -17,19 +17,17 @@ DRAGON models have been fine-tuned with the specific objective of fact-based que
17
  Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
18
  1 Test Run with sample=False & temperature=0.0 (deterministic output) - 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
19
 
20
- --**Accuracy Score**: **99.0** correct out of 100
21
- --Not Found Classification: 85.0%
22
- --Boolean: 100.0%
23
- --Math/Logic: 92.5%
24
- --Complex Questions (1-5): 5 (Best in Class)
25
  --Summarization Quality (1-5): 3 (Average)
26
  --Hallucinations: No hallucinations observed in test runs.
27
 
28
  For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
29
 
30
- Please note that these test results were achieved using the 4_K_M quantized version of this model - [dragon-qwen-7b-gguf](https://www.huggingface.co/llmware/dragon-qwen-7b-gguf).
31
-
32
- Note: compare results with [dragon-mistral-0.3-gguf](https://www.huggingface.co/llmware/dragon-mistral-0.3-gguf).
33
 
34
 
35
  ### Model Description
@@ -40,17 +38,17 @@ Note: compare results with [dragon-mistral-0.3-gguf](https://www.huggingface.co/
40
  - **Model type:** Qwen
41
  - **Language(s) (NLP):** English
42
  - **License:** Apache 2.0
43
- - **Finetuned from model:** Qwen2-7b-base
44
 
45
 
46
  ### Direct Use
47
 
48
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
49
 
50
- DRAGON is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
51
  legal and regulatory industries with complex information sources.
52
 
53
- DRAGON models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
54
  without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
55
 
56
 
@@ -66,8 +64,8 @@ Any model can provide inaccurate or incomplete information, and should be used i
66
  The fastest way to get started with dRAGon is through direct import in transformers:
67
 
68
  from transformers import AutoTokenizer, AutoModelForCausalLM
69
- tokenizer = AutoTokenizer.from_pretrained("dragon-qwen-7b")
70
- model = AutoModelForCausalLM.from_pretrained("dragon-qwen-7b")
71
 
72
  Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
73
 
 
3
  inference: false
4
  ---
5
 
6
+ # BLING-QWEN-1.5B
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ bling-qwen-1.5b is part of the BLING model series, RAG-instruct trained on top of a Qwen2 1.5b base model.
11
 
12
+ BLING models have been fine-tuned with the specific objective of fact-based question-answering over complex business and legal documents with an emphasis on reducing hallucinations and providing short, clear answers for workflow automation.
13
 
14
 
15
  ### Benchmark Tests
 
17
  Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
18
  1 Test Run with sample=False & temperature=0.0 (deterministic output) - 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
19
 
20
+ --**Accuracy Score**: **93.5** correct out of 100
21
+ --Not Found Classification: 75.0%
22
+ --Boolean: 87.5%
23
+ --Math/Logic: 70%
24
+ --Complex Questions (1-5): 3 (Best in Class)
25
  --Summarization Quality (1-5): 3 (Average)
26
  --Hallucinations: No hallucinations observed in test runs.
27
 
28
  For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
29
 
30
+ Please note that these test results were achieved using the 4_K_M quantized version of this model - [bling-qwen-mini-tool](https://www.huggingface.co/llmware/bling-qwen-mini-tool).
 
 
31
 
32
 
33
  ### Model Description
 
38
  - **Model type:** Qwen
39
  - **Language(s) (NLP):** English
40
  - **License:** Apache 2.0
41
+ - **Finetuned from model:** Qwen2-1.5b-base
42
 
43
 
44
  ### Direct Use
45
 
46
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
 
48
+ BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
49
  legal and regulatory industries with complex information sources.
50
 
51
+ BLING models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
52
  without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
53
 
54
 
 
64
  The fastest way to get started with dRAGon is through direct import in transformers:
65
 
66
  from transformers import AutoTokenizer, AutoModelForCausalLM
67
+ tokenizer = AutoTokenizer.from_pretrained("llmware/bling-qwen-1.5b")
68
+ model = AutoModelForCausalLM.from_pretrained("llmware/bling-qwen-1.5b")
69
 
70
  Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
71