Spaces:
Running
Running
Add support for Meta Llama 3 70B in HuggingChat (#1029)
Browse files
.env.template
CHANGED
|
@@ -27,6 +27,31 @@ MODELS=`[
|
|
| 27 |
}
|
| 28 |
]
|
| 29 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
{
|
| 31 |
"name" : "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
| 32 |
"tokenizer": "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
|
@@ -51,7 +76,7 @@ MODELS=`[
|
|
| 51 |
"prompt": "How do I make a delicious lemon cheesecake?"
|
| 52 |
}
|
| 53 |
]
|
| 54 |
-
|
| 55 |
{
|
| 56 |
"name" : "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 57 |
"description" : "The latest MoE model from Mistral AI! 8x7B and outperforms Llama 2 70B in most benchmarks.",
|
|
|
|
| 27 |
}
|
| 28 |
]
|
| 29 |
},
|
| 30 |
+
{
|
| 31 |
+
"name" : "meta-llama/Meta-Llama-3-70B-Instruct",
|
| 32 |
+
"description": "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.",
|
| 33 |
+
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
|
| 34 |
+
"modelUrl": "https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct",
|
| 35 |
+
"websiteUrl": "https://llama.meta.com/llama3/",
|
| 36 |
+
"tokenizer" : "meta-llama/Meta-Llama-3-70B-Instruct",
|
| 37 |
+
"promptExamples" : [
|
| 38 |
+
{
|
| 39 |
+
"title": "Write an email from bullet list",
|
| 40 |
+
"prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
|
| 41 |
+
}, {
|
| 42 |
+
"title": "Code a snake game",
|
| 43 |
+
"prompt": "Code a basic snake game in python, give explanations for each step."
|
| 44 |
+
}, {
|
| 45 |
+
"title": "Assist in a task",
|
| 46 |
+
"prompt": "How do I make a delicious lemon cheesecake?"
|
| 47 |
+
}
|
| 48 |
+
],
|
| 49 |
+
"parameters": {
|
| 50 |
+
"stop": ["<|eot_id|>"],
|
| 51 |
+
"truncate": 6144,
|
| 52 |
+
"max_new_tokens": 2047
|
| 53 |
+
}
|
| 54 |
+
},
|
| 55 |
{
|
| 56 |
"name" : "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
| 57 |
"tokenizer": "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
|
|
|
| 76 |
"prompt": "How do I make a delicious lemon cheesecake?"
|
| 77 |
}
|
| 78 |
]
|
| 79 |
+
},
|
| 80 |
{
|
| 81 |
"name" : "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 82 |
"description" : "The latest MoE model from Mistral AI! 8x7B and outperforms Llama 2 70B in most benchmarks.",
|
src/lib/components/ModelCardMetadata.svelte
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
<script lang="ts">
|
| 2 |
import CarbonEarth from "~icons/carbon/earth";
|
| 3 |
import CarbonArrowUpRight from "~icons/carbon/arrow-up-right";
|
|
|
|
| 4 |
import type { Model } from "$lib/types/Model";
|
| 5 |
|
| 6 |
export let model: Pick<Model, "name" | "datasetName" | "websiteUrl" | "modelUrl" | "datasetUrl">;
|
|
@@ -41,8 +42,13 @@
|
|
| 41 |
class="ml-auto flex items-center hover:underline"
|
| 42 |
rel="noreferrer"
|
| 43 |
>
|
| 44 |
-
|
| 45 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
</a>
|
| 47 |
{/if}
|
| 48 |
</div>
|
|
|
|
| 1 |
<script lang="ts">
|
| 2 |
import CarbonEarth from "~icons/carbon/earth";
|
| 3 |
import CarbonArrowUpRight from "~icons/carbon/arrow-up-right";
|
| 4 |
+
import BIMeta from "~icons/bi/meta";
|
| 5 |
import type { Model } from "$lib/types/Model";
|
| 6 |
|
| 7 |
export let model: Pick<Model, "name" | "datasetName" | "websiteUrl" | "modelUrl" | "datasetUrl">;
|
|
|
|
| 42 |
class="ml-auto flex items-center hover:underline"
|
| 43 |
rel="noreferrer"
|
| 44 |
>
|
| 45 |
+
{#if model.name === "meta-llama/Meta-Llama-3-70B-Instruct"}
|
| 46 |
+
<BIMeta class="mr-1.5 shrink-0 text-xs text-gray-400" />
|
| 47 |
+
Built with Meta Llama 3
|
| 48 |
+
{:else}
|
| 49 |
+
<CarbonEarth class="mr-1.5 shrink-0 text-xs text-gray-400" />
|
| 50 |
+
Website
|
| 51 |
+
{/if}
|
| 52 |
</a>
|
| 53 |
{/if}
|
| 54 |
</div>
|
src/routes/conversation/[id]/+server.ts
CHANGED
|
@@ -447,7 +447,8 @@ export async function POST({ request, locals, params, getClientAddress }) {
|
|
| 447 |
messageToWriteTo.content += output.token.text;
|
| 448 |
}
|
| 449 |
} else {
|
| 450 |
-
messageToWriteTo.interrupted =
|
|
|
|
| 451 |
// add output.generated text to the last message
|
| 452 |
// strip end tokens from the output.generated_text
|
| 453 |
const text = (model.parameters.stop ?? []).reduce((acc: string, curr: string) => {
|
|
|
|
| 447 |
messageToWriteTo.content += output.token.text;
|
| 448 |
}
|
| 449 |
} else {
|
| 450 |
+
messageToWriteTo.interrupted =
|
| 451 |
+
!output.token.special && !model.parameters.stop?.includes(output.token.text);
|
| 452 |
// add output.generated text to the last message
|
| 453 |
// strip end tokens from the output.generated_text
|
| 454 |
const text = (model.parameters.stop ?? []).reduce((acc: string, curr: string) => {
|