doberst commited on
Commit
608eb87
·
verified ·
1 Parent(s): 1c3bc25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -40
README.md CHANGED
@@ -3,29 +3,13 @@ license: apache-2.0
3
  inference: false
4
  ---
5
 
6
- # dragon-phi-3-answer-tool
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- dragon-phi-3-answer-tool is part of the DRAGON ("Delivering RAG On ...") model series, RAG-instruct trained on top of a Microsoft Phi-3 base model.
11
 
12
- DRAGON models are fine-tuned with high-quality custom instruct datasets, designed for production use in RAG scenarios.
13
-
14
-
15
- ### Benchmark Tests
16
-
17
- Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
18
- Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
19
-
20
- --**Accuracy Score**: **100.0** correct out of 100
21
- --Not Found Classification: 95.0%
22
- --Boolean: 97.5%
23
- --Math/Logic: 80.0%
24
- --Complex Questions (1-5): 4 (Above Average - multiple-choice, causal)
25
- --Summarization Quality (1-5): 4 (Above Average)
26
- --Hallucinations: No hallucinations observed in test runs.
27
-
28
- For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
29
 
30
  ### Model Description
31
 
@@ -37,27 +21,6 @@ For test run results (and good indicator of target use cases), please see the fi
37
  - **License:** Apache 2.0
38
  - **Finetuned from model:** Microsoft Phi-3
39
 
40
- ## Uses
41
-
42
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
-
44
- The intended use of BLING models is two-fold:
45
-
46
- 1. Provide high-quality RAG-Instruct models designed for fact-based, no "hallucination" question-answering in connection with an enterprise RAG workflow.
47
-
48
- 2. BLING models are fine-tuned on top of leading base foundation models, generally in the 1-3B+ range, and purposefully rolled-out across multiple base models to provide choices and "drop-in" replacements for RAG specific use cases.
49
-
50
-
51
- ### Direct Use
52
-
53
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
54
-
55
- BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
56
- legal and regulatory industries with complex information sources.
57
-
58
- BLING models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
59
- without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
60
-
61
 
62
  ## Bias, Risks, and Limitations
63
 
 
3
  inference: false
4
  ---
5
 
6
+ # slim-q-gen-phi-3
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ slim-q-gen-phi-3 is a specialized function-calling model, finetuned on top of a phi-3-mini-4k base, to generate a python dictionary consisting of a single key "question".
11
 
12
+ For most inference use cases, we would recommend using the quantized 'tool' version of this model - [slim-q-gen-phi-3-tool](https://huggingface.co/llmware/slim-q-gen-phi-3-tool)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ### Model Description
15
 
 
21
  - **License:** Apache 2.0
22
  - **Finetuned from model:** Microsoft Phi-3
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
 
25
  ## Bias, Risks, and Limitations
26