doberst commited on
Commit
8c6bee9
1 Parent(s): 9d60662

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -11
README.md CHANGED
@@ -7,22 +7,22 @@ inference: false
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- **slim-sentiment** is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") model series, consisting of small, specialized decoder-based models, fine-tuned for function-calling.
11
 
12
- slim-sentiment has been fine-tuned for **sentiment analysis** function calls, generating output consisting of a python dictionary corresponding to specified keys, e.g.:
13
 
14
- &nbsp;&nbsp;&nbsp;&nbsp;`{"sentiment": ["positive"]}`
15
 
16
 
17
  SLIM models are designed to provide a flexible natural language generative model that can be used as part of a multi-step, multi-model LLM-based automation workflow.
18
 
19
- Each slim model has a 'quantized tool' version, e.g., [**'slim-sentiment-tool'**](https://huggingface.co/llmware/slim-sentiment-tool).
20
 
21
 
22
  ## Prompt format:
23
 
24
  `function = "classify"`
25
- `params = "sentiment"`
26
  `prompt = "<human> " + {text} + "\n" + `
27
  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp; &nbsp; &nbsp; &nbsp;`"<{function}> " + {params} + "</{function}>" + "\n<bot>:"`
28
 
@@ -30,13 +30,13 @@ Each slim model has a 'quantized tool' version, e.g., [**'slim-sentiment-tool'*
30
  <details>
31
  <summary>Transformers Script </summary>
32
 
33
- model = AutoModelForCausalLM.from_pretrained("llmware/slim-sentiment")
34
- tokenizer = AutoTokenizer.from_pretrained("llmware/slim-sentiment")
35
 
36
  function = "classify"
37
- params = "sentiment"
38
 
39
- text = "The stock market declined yesterday as investors worried increasingly about the slowing economy."
40
 
41
  prompt = "<human>: " + text + "\n" + f"<{function}> {params} </{function}>\n<bot>:"
42
 
@@ -73,8 +73,8 @@ Each slim model has a 'quantized tool' version, e.g., [**'slim-sentiment-tool'*
73
  <summary>Using as Function Call in LLMWare</summary>
74
 
75
  from llmware.models import ModelCatalog
76
- slim_model = ModelCatalog().load_model("llmware/slim-sentiment")
77
- response = slim_model.function_call(text,params=["sentiment"], function="classify")
78
 
79
  print("llmware - llm_response: ", response)
80
 
 
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ **slim-intent** is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") model series, consisting of small, specialized decoder-based models, fine-tuned for function-calling.
11
 
12
+ slim-intent has been fine-tuned for **intent analysis** function calls, generating output consisting of a python dictionary corresponding to specified keys, e.g.:
13
 
14
+ &nbsp;&nbsp;&nbsp;&nbsp;`{"intent": ["complaint"]}`
15
 
16
 
17
  SLIM models are designed to provide a flexible natural language generative model that can be used as part of a multi-step, multi-model LLM-based automation workflow.
18
 
19
+ Each slim model has a 'quantized tool' version, e.g., [**'slim-intent-tool'**](https://huggingface.co/llmware/slim-intent-tool).
20
 
21
 
22
  ## Prompt format:
23
 
24
  `function = "classify"`
25
+ `params = "intent"`
26
  `prompt = "<human> " + {text} + "\n" + `
27
  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp; &nbsp; &nbsp; &nbsp;`"<{function}> " + {params} + "</{function}>" + "\n<bot>:"`
28
 
 
30
  <details>
31
  <summary>Transformers Script </summary>
32
 
33
+ model = AutoModelForCausalLM.from_pretrained("llmware/slim-intent")
34
+ tokenizer = AutoTokenizer.from_pretrained("llmware/slim-intent")
35
 
36
  function = "classify"
37
+ params = "intent"
38
 
39
+ text = "I am really impressed with the quality of the product and the service that I have received so far."
40
 
41
  prompt = "<human>: " + text + "\n" + f"<{function}> {params} </{function}>\n<bot>:"
42
 
 
73
  <summary>Using as Function Call in LLMWare</summary>
74
 
75
  from llmware.models import ModelCatalog
76
+ slim_model = ModelCatalog().load_model("llmware/slim-intent")
77
+ response = slim_model.function_call(text,params=["intent"], function="classify")
78
 
79
  print("llmware - llm_response: ", response)
80