doberst commited on
Commit
34643e0
1 Parent(s): 9f5d8bb

Upload 3 files

Browse files
Files changed (3) hide show
  1. README.md +122 -0
  2. config.json +23 -0
  3. tokenizer.model +3 -0
README.md CHANGED
@@ -1,3 +1,125 @@
1
  ---
2
  license: llama2
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama2
3
  ---
4
+
5
+ # Model Card for Model ID
6
+
7
+ <!-- Provide a quick summary of what the model is/does. -->
8
+
9
+ dragon-llama-7b-v0 part of the dRAGon ("Delivering RAG On ...") model series, RAG-instruct trained on top of a LLama-2 base model.
10
+
11
+ DRAGON models are fine-tuned with high-quality custom instruct datasets, designed for production quality use in RAG scenarios.
12
+
13
+
14
+ ### Benchmark Tests
15
+
16
+ Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
17
+ Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
18
+
19
+ --**Accuracy Score**: **99.0** correct out of 100
20
+ --Not Found Classification: 95.0%
21
+ --Boolean: 82.5%
22
+ --Math/Logic: 70.0%
23
+ --Complex Questions (1-5): 4 (Low-Medium)
24
+ --Summarization Quality (1-5): 4 (Coherent, extractive)
25
+ --Hallucinations: No hallucinations observed in test runs.
26
+
27
+ For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
28
+
29
+ ### Model Description
30
+
31
+ <!-- Provide a longer summary of what this model is. -->
32
+
33
+ - **Developed by:** llmware
34
+ - **Model type:** LLama-2
35
+ - **Language(s) (NLP):** English
36
+ - **License:** LLama 2 Community License Agreement
37
+ - **Finetuned from model:** Llama-2-7B-Base
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ The intended use of DRAGON models is two-fold:
44
+
45
+ 1. Provide high-quality RAG-Instruct models designed for fact-based, no "hallucination" question-answering in connection with an enterprise RAG workflow.
46
+
47
+ 2. DRAGON models are fine-tuned on top of leading base foundation models, generally in the 6-7B+ range, and purposefully rolled-out across multiple base models to provide choices and "drop-in" replacements for RAG specific use cases.
48
+
49
+ 3. DRAGON models were trained on the same principles as the BLING models, so generally, it should be easy to "upgrade" from a BLING model in testing to a DRAGON model in production.
50
+
51
+
52
+ ### Direct Use
53
+
54
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
55
+
56
+ DRAGON is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
57
+ legal and regulatory industries with complex information sources.
58
+
59
+ DRAGON models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
60
+ without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
61
+
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ Any model can provide inaccurate or incomplete information, and should be used in conjunction with appropriate safeguards and fact-checking mechanisms.
68
+
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ The fastest way to get started with dRAGon is through direct import in transformers:
73
+
74
+ from transformers import AutoTokenizer, AutoModelForCausalLM
75
+ tokenizer = AutoTokenizer.from_pretrained("dragon-llama-7b-v0")
76
+ model = AutoModelForCausalLM.from_pretrained("dragon-llama-7b-v0")
77
+
78
+ Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
79
+
80
+ The dRAGon model was fine-tuned with a simple "\<human> and \<bot>" wrapper, so to get the best results, wrap inference entries as:
81
+
82
+ full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
83
+
84
+ The BLING model was fine-tuned with closed-context samples, which assume generally that the prompt consists of two sub-parts:
85
+
86
+ 1. Text Passage Context, and
87
+ 2. Specific question or instruction based on the text passage
88
+
89
+ To get the best results, package "my_prompt" as follows:
90
+
91
+ my_prompt = {{text_passage}} + "\n" + {{question/instruction}}
92
+
93
+
94
+ If you are using a HuggingFace generation script:
95
+
96
+ # prepare prompt packaging used in fine-tuning process
97
+ new_prompt = "<human>: " + entries["context"] + "\n" + entries["query"] + "\n" + "<bot>:"
98
+
99
+ inputs = tokenizer(new_prompt, return_tensors="pt")
100
+ start_of_output = len(inputs.input_ids[0])
101
+
102
+ # temperature: set at 0.3 for consistency of output
103
+ # max_new_tokens: set at 100 - may prematurely stop a few of the summaries
104
+
105
+ outputs = model.generate(
106
+ inputs.input_ids.to(device),
107
+ eos_token_id=tokenizer.eos_token_id,
108
+ pad_token_id=tokenizer.eos_token_id,
109
+ do_sample=True,
110
+ temperature=0.3,
111
+ max_new_tokens=100,
112
+ )
113
+
114
+ output_only = tokenizer.decode(outputs[0][start_of_output:],skip_special_tokens=True)
115
+
116
+ # note: due to artifact of the fine-tuning, use this post-processing with HF generation
117
+
118
+ eot = output_only.find("<|endoftext|>")
119
+ if eot > -1:
120
+ output_only = output_only[:eot]
121
+
122
+
123
+ ## Model Card Contact
124
+
125
+ Darren Oberst & llmware team
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_class": "llama2",
3
+ "model_size": "7b",
4
+ "architectures": [
5
+ "Llama2ForCausalLM"
6
+ ],
7
+ "dim": 4096,
8
+ "n_layers": 32,
9
+ "n_heads": 32,
10
+ "n_kv_heads": null,
11
+ "vocab_size": 32000,
12
+ "multiple_of": 256,
13
+ "ffn_dim_multiplier": null,
14
+ "norm_eps": 1e-5,
15
+ "max_batch_size": 32,
16
+ "max_seq_len": 2048,
17
+ "bos_token_id":1,
18
+ "eos_token_id":2,
19
+ "pad_token_id":-1,
20
+ "torch_dtype": "float16",
21
+ "pretraining_base": "llama2-7b-2t-tokens",
22
+ "model_repo_folder_path": "llama2-7b-base"
23
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723