doberst commited on
Commit
0fe91b7
1 Parent(s): d06c209

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -31
README.md CHANGED
@@ -6,53 +6,43 @@ license: apache-2.0
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
- **slim-ner-tool** is part of the SLIM ("Structured Language Instruction Model") model series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
10
 
11
- slim-ner-tool is a 4_K_M quantized GGUF version of slim-ner, providing a small, fast inference implementation.
12
 
13
- Load in your favorite GGUF inference engine (see details in config.json to set up the prompt template), or try with llmware as follows:
14
 
15
- from llmware.models import ModelCatalog
16
 
17
- # to load the model and make a basic inference
18
- ner_tool = ModelCatalog().load_model("slim-ner-tool")
19
- response = ner_tool.function_call(text_sample)
20
-
21
- # this one line will download the model and run a series of tests
22
- ModelCatalog().test_run("slim-ner-tool", verbose=True)
23
-
24
 
25
- Slim models can also be loaded even more simply as part of a multi-model, multi-step LLMfx calls:
26
-
27
- from llmware.agents import LLMfx
28
-
29
- llm_fx = LLMfx()
30
- llm_fx.load_tool("ner")
31
- response = llm_fx.named_entity_extraction(text)
32
 
 
33
 
34
- ### Model Description
 
 
 
 
35
 
36
- <!-- Provide a longer summary of what this model is. -->
 
37
 
38
- - **Developed by:** llmware
39
- - **Model type:** GGUF
40
- - **Language(s) (NLP):** English
41
- - **License:** Apache 2.0
42
- - **Quantized from model:** llmware/slim-sentiment (finetuned tiny llama)
43
 
44
- ## Uses
45
 
46
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
 
48
- SLIM models provide a fast, flexible, intuitive way to integrate classifiers and structured function calls into RAG and LLM application workflows.
 
 
49
 
50
- Model instructions, details and test samples have been packaged into the config.json file in the repository, along with the GGUF file.
51
 
 
52
 
53
 
54
  ## Model Card Contact
55
 
56
- Darren Oberst & llmware team
57
-
58
 
 
 
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
 
9
 
10
+ **slim-sentiment-tool** is a 4_K_M quantized GGUF version of slim-sentiment, providing a small, fast inference implementation, optimized for multi-model concurrent deployment.
11
 
12
+ [**slim-sentiment**](https://huggingface.co/llmware/slim-sentiment) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
13
 
14
+ To pull the model via API:
15
 
16
+ from huggingface_hub import snapshot_download
17
+ snapshot_download("llmware/slim-sentiment-tool", local_dir="/path/on/your/machine/", local_dir_use_symlinks=False)
 
 
 
 
 
18
 
 
 
 
 
 
 
 
19
 
20
+ Load in your favorite GGUF inference engine, or try with llmware as follows:
21
 
22
+ from llmware.models import ModelCatalog
23
+
24
+ # to load the model and make a basic inference
25
+ model = ModelCatalog().load_model("slim-sentiment-tool")
26
+ response = model.function_call(text_sample)
27
 
28
+ # this one line will download the model and run a series of tests
29
+ ModelCatalog().tool_test_run("slim-sentiment-tool", verbose=True)
30
 
 
 
 
 
 
31
 
32
+ Slim models can also be loaded even more simply as part of a multi-model, multi-step LLMfx calls:
33
 
34
+ from llmware.agents import LLMfx
35
 
36
+ llm_fx = LLMfx()
37
+ llm_fx.load_tool("sentiment")
38
+ response = llm_fx.sentiment(text)
39
 
 
40
 
41
+ Note: please review [**config.json**](https://huggingface.co/llmware/slim-sentiment-tool/blob/main/config.json) in the repository for prompt wrapping information, details on the model, and full test set.
42
 
43
 
44
  ## Model Card Contact
45
 
46
+ Darren Oberst & llmware team
 
47
 
48
+ [Any questions? Join us on Discord](https://discord.gg/MhZn5Nc39h)