doberst commited on
Commit
5c1d51a
1 Parent(s): 090e7bf

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -35
README.md CHANGED
@@ -8,22 +8,19 @@ license: apache-2.0
8
 
9
  BLING-1.4b-0.1 is the first model release in the BLING ("Best Little Instruction-following No-GPU-required") model series.
10
 
11
- BLING models are designed as custom instruct-following laptop-effective GPT decoder-based models (~1B-2.7B parameters). BLING models are currently built on top of Pythia (GPTNeox architecture) base models and other Apache 2.0-licensed GPT-compatible models with primary focus on 'little' models in the range of 1B, 1.3-1.4B, and 2.7B parameters. (Note: in our testing, we have seen relatively limited success with instruct-following models below <1B parameters.)
12
-
13
- BLING models are fine-tuned with distilled high-quality custom instruct datasets, targeted at a specific subset of instruct tasks with the objective of providing a high-quality Instruct model that can be run entirely without a GPU server, with good quality instruct-following capability that can be loaded and run locally on a laptop.
14
-
15
- ## Model Details
16
 
17
  ### Model Description
18
 
19
  <!-- Provide a longer summary of what this model is. -->
20
 
21
  - **Developed by:** llmware
22
- - **Shared by [optional]:** Darren Oberst
23
  - **Model type:** GPTNeoX instruct-trained decoder
24
  - **Language(s) (NLP):** English
25
  - **License:** Apache 2.0
26
- - **Finetuned from model [optional]:** EleutherAI/Pythia-1b-deduped
27
 
28
  ## Uses
29
 
@@ -31,10 +28,11 @@ BLING models are fine-tuned with distilled high-quality custom instruct datasets
31
 
32
  The intended use of BLING models is two-fold:
33
 
34
- 1. Provide a high-quality Instruct models that can run on a laptop for local testing. We have found it extremely useful when building a
35
  proof-of-concept, or working with sensitive enterprise data that must be closely guarded, especially in RAG use cases.
36
 
37
- 2. Push the state of the art for smaller Instruct-following models in the 1B - 7B range through improved fine-tuning datasets and targeted "instruction" tasks.
 
38
 
39
 
40
  ### Direct Use
@@ -42,23 +40,14 @@ The intended use of BLING models is two-fold:
42
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
 
44
  BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
45
- legal and regulatory industries. BLING is intended to be an experimental series of little instruct models targeted as specific
46
- RAG automation tasks with complex information sources. Rather than try to be "all things to all people," BLING models try to focus
47
- on a narrower set of Instructions more suitable to a ~1B parameter GPT model.
48
 
49
  BLING is ideal for rapid prototyping, testing, and the ability to perform an end-to-end workflow locally on a laptop without
50
  having to send sensitive information over an Internet-based API.
51
 
52
- The first BLING models have been trained on question-answering, key-value extraction, and basic summarization as the core instruction types.
53
-
54
-
55
- [More Information Needed]
56
-
57
- ### Downstream Use [optional]
58
-
59
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
60
 
61
- [More Information Needed]
62
 
63
  ### Out-of-Scope Use
64
 
@@ -69,31 +58,27 @@ The first BLING models have been trained on question-answering, key-value extrac
69
  2. BLING is not optimal for most production applications, other than simple and highly specific use cases.
70
 
71
 
72
- [More Information Needed]
73
-
74
  ## Bias, Risks, and Limitations
75
 
76
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
77
 
78
- BLING has not been designed for end consumer-oriented applications, and there has been any focus in training on important safeguards to
79
- mitigate potential bias and safety. We would strongly discourage any use of BLING for any 'chatbot' use case.
80
-
81
- [More Information Needed]
82
 
83
 
84
  ## How to Get Started with the Model
85
 
86
  The fastest way to get started with BLING is through direct import in transformers:
87
 
88
- from transformers import AutoTokenizer, AutoModelForCausalLM
89
- tokenizer = AutoTokenizer.from_pretrained("llmware/bling-1.4b-0.1")
90
- model = AutoModelForCausalLM.from_pretrained("llmware/bling-1.4b-0.1")
 
91
 
92
- The BLING model was fine-tuned with a simple "<human> and <bot> wrapper", so to get the best results, wrap inference entries as:
93
 
94
- full_prompt = "<human>: " + my_prompt + "\n" + "<bot>: "
95
 
96
- The BLING model was fine-tuned with closed-context samples, which assume generally that the prompt consists of sub-parts:
97
 
98
  1. Text Passage Context, and
99
  2. Specific question or instruction based on the text passage
@@ -105,13 +90,23 @@ my_prompt = {{text_passage}} + "\n" + {{question/instruction}}
105
 
106
  ## Citation [optional]
107
 
108
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
 
 
 
 
 
 
 
 
 
 
109
 
110
  ## Model Card Contact
111
 
112
  Darren Oberst & llmware team
113
 
114
- Please reach out anytime if you are interested in this research program and would like to participate and work with us!
115
 
116
 
117
 
 
8
 
9
  BLING-1.4b-0.1 is the first model release in the BLING ("Best Little Instruction-following No-GPU-required") model series.
10
 
11
+ BLING models are fine-tuned with distilled high-quality custom instruct datasets, targeted at a specific subset of instruct tasks with
12
+ the objective of providing a high-quality Instruct model that is 'inference-ready' on a CPU laptop even
13
+ without using any advanced quantization optimizations.
 
 
14
 
15
  ### Model Description
16
 
17
  <!-- Provide a longer summary of what this model is. -->
18
 
19
  - **Developed by:** llmware
 
20
  - **Model type:** GPTNeoX instruct-trained decoder
21
  - **Language(s) (NLP):** English
22
  - **License:** Apache 2.0
23
+ - **Finetuned from model [optional]:** EleutherAI/Pythia-1.4b-v0
24
 
25
  ## Uses
26
 
 
28
 
29
  The intended use of BLING models is two-fold:
30
 
31
+ 1. Provide high-quality Instruct models that can run on a laptop for local testing. We have found it extremely useful when building a
32
  proof-of-concept, or working with sensitive enterprise data that must be closely guarded, especially in RAG use cases.
33
 
34
+ 2. Push the state of the art for smaller Instruct-following models in the sub-7B parameter range, especially 1B-3B, as single-purpose
35
+ automation tools for specific tasks through targeted fine-tuning datasets and focused "instruction" tasks.
36
 
37
 
38
  ### Direct Use
 
40
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
41
 
42
  BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
43
+ legal and regulatory industries with complex information sources. Rather than try to be "all things to all people," BLING models try to focus on a narrower set of Instructions more suitable to a ~1B parameter GPT model.
 
 
44
 
45
  BLING is ideal for rapid prototyping, testing, and the ability to perform an end-to-end workflow locally on a laptop without
46
  having to send sensitive information over an Internet-based API.
47
 
48
+ The first BLING models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
49
+ without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
 
 
 
 
 
 
50
 
 
51
 
52
  ### Out-of-Scope Use
53
 
 
58
  2. BLING is not optimal for most production applications, other than simple and highly specific use cases.
59
 
60
 
 
 
61
  ## Bias, Risks, and Limitations
62
 
63
  <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
 
65
+ BLING has not been designed for end consumer-oriented applications, and there has not been any focus in training on safeguards to mitigate potential bias. We would strongly discourage any use of BLING for any 'chatbot' use case.
 
 
 
66
 
67
 
68
  ## How to Get Started with the Model
69
 
70
  The fastest way to get started with BLING is through direct import in transformers:
71
 
72
+ from transformers import AutoTokenizer, AutoModelForCausalLM
73
+ tokenizer = AutoTokenizer.from_pretrained("llmware/bling-1.4b-0.1")
74
+ model = AutoModelForCausalLM.from_pretrained("llmware/bling-1.4b-0.1")
75
+
76
 
77
+ The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
78
 
79
+ full_prompt = "\<human>\: " + my_prompt + "\n" + "\<bot>\: "
80
 
81
+ The BLING model was fine-tuned with closed-context samples, which assume generally that the prompt consists of two sub-parts:
82
 
83
  1. Text Passage Context, and
84
  2. Specific question or instruction based on the text passage
 
90
 
91
  ## Citation [optional]
92
 
93
+ BLING models are built on top of EleutherAI/Pythia base - please see citation for Pythia below:
94
+
95
+ @misc{biderman2023pythia,
96
+ title={Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling},
97
+ author={Stella Biderman and Hailey Schoelkopf and Quentin Anthony and Herbie Bradley and Kyle O'Brien and Eric Hallahan and Mohammad Aflah Khan and Shivanshu Purohit and USVSN Sai Prashanth and Edward Raff and Aviya Skowron and Lintang Sutawika and Oskar van der Wal},
98
+ year={2023},
99
+ eprint={2304.01373},
100
+ archivePrefix={arXiv},
101
+ primaryClass={cs.CL}
102
+ }
103
+
104
 
105
  ## Model Card Contact
106
 
107
  Darren Oberst & llmware team
108
 
109
+ Please reach out anytime if you are interested in this project and would like to participate and work with us!
110
 
111
 
112