HoneyBadger2989 commited on
Commit
58be25c
·
verified ·
1 Parent(s): 3a674ee

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ llama-3-groq-8b-tool-use.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ llama-3-groq-8b-tool-use.bf16.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/Meta-Llama-3-8B
3
+ language:
4
+ - en
5
+ license: llama3
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - facebook
9
+ - meta
10
+ - pytorch
11
+ - llama
12
+ - llama-3
13
+ - groq
14
+ - tool-use
15
+ - function-calling
16
+ - autoquant
17
+ - gguf
18
+ ---
19
+
20
+ # Llama-3-Groq-8B-Tool-Use
21
+
22
+ This is the 8B parameter version of the Llama 3 Groq Tool Use model, specifically designed for advanced tool use and function calling tasks.
23
+
24
+ ## Model Details
25
+
26
+ - **Model Type:** Causal language model fine-tuned for tool use
27
+ - **Language(s):** English
28
+ - **License:** Meta Llama 3 Community License
29
+ - **Model Architecture:** Optimized transformer
30
+ - **Training Approach:** Full fine-tuning and Direct Preference Optimization (DPO) on Llama 3 8B base model
31
+ - **Input:** Text
32
+ - **Output:** Text, with enhanced capabilities for tool use and function calling
33
+
34
+ ## Performance
35
+
36
+ - **Berkeley Function Calling Leaderboard (BFCL) Score:** 89.06% overall accuracy
37
+ - This score represents the best performance among all open-source 8B LLMs on the BFCL
38
+
39
+ ## Usage and Limitations
40
+
41
+ This model is designed for research and development in tool use and function calling scenarios. It excels at tasks involving API interactions, structured data manipulation, and complex tool use. However, users should note:
42
+
43
+ - For general knowledge or open-ended tasks, a general-purpose language model may be more suitable
44
+ - The model may still produce inaccurate or biased content in some cases
45
+ - Users are responsible for implementing appropriate safety measures for their specific use case
46
+
47
+ Note the model is quite sensitive to the `temperature` and `top_p` sampling configuration. Start at `temperature=0.5, top_p=0.65` and move up or down as needed.
48
+
49
+ Text prompt example:
50
+
51
+ We'd like to give a special shoutout to [@NousResearch](https://x.com/NousResearch) for pushing open source tool use forward with their public & open exploration of tool use in LLMs.
52
+
53
+ ```
54
+ <|start_header_id|>system<|end_header_id|>
55
+
56
+ You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. For each function call return a json object with function name and arguments within <tool_call></tool_call> XML tags as follows:
57
+ <tool_call>
58
+ {"name": <function-name>,"arguments": <args-dict>}
59
+ </tool_call>
60
+
61
+ Here are the available tools:
62
+ <tools> {
63
+ "name": "get_current_weather",
64
+ "description": "Get the current weather in a given location",
65
+ "parameters": {
66
+ "properties": {
67
+ "location": {
68
+ "description": "The city and state, e.g. San Francisco, CA",
69
+ "type": "string"
70
+ },
71
+ "unit": {
72
+ "enum": [
73
+ "celsius",
74
+ "fahrenheit"
75
+ ],
76
+ "type": "string"
77
+ }
78
+ },
79
+ "required": [
80
+ "location"
81
+ ],
82
+ "type": "object"
83
+ }
84
+ } </tools><|eot_id|><|start_header_id|>user<|end_header_id|>
85
+
86
+ What is the weather like in San Francisco?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
87
+
88
+ <tool_call>
89
+ {"id":"call_deok","name":"get_current_weather","arguments":{"location":"San Francisco","unit":"celsius"}}
90
+ </tool_call><|eot_id|><|start_header_id|>tool<|end_header_id|>
91
+
92
+ <tool_response>
93
+ {"id":"call_deok","result":{"temperature":"72","unit":"celsius"}}
94
+ </tool_response><|eot_id|><|start_header_id|>assistant<|end_header_id|>
95
+
96
+ ```
97
+
98
+ ## Ethical Considerations
99
+
100
+ While fine-tuned for tool use, this model inherits the ethical considerations of the base Llama 3 model. Use responsibly and implement additional safeguards as needed for your application.
101
+
102
+ ## Availability
103
+
104
+ The model is available through:
105
+ - [Groq API console](https://console.groq.com)
106
+ - [Hugging Face](https://huggingface.co/Groq/Llama-3-Groq-8B-Tool-Use)
107
+
108
+ For full details on responsible use, ethical considerations, and latest benchmarks, please refer to the [official Llama 3 documentation](https://llama.meta.com/) and the Groq model card.
llama-3-groq-8b-tool-use.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:949e991e774c676de8f1997c99d854177455fb92e5e640fd62e9cd61513a749f
3
+ size 3179160320
llama-3-groq-8b-tool-use.bf16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89c0a100b46126e8bf5119b027eb9bc141cdcdee42f584ab81613168fd0eb51d
3
+ size 16068990144