alway respond user instruction

#1
by telehan - opened

thanks for your gguf model, but it seems always respond user words
test with ollama with following configuration

$ ollama show --modelfile mistral:7b-ins-fcjson-v0.2-q8
# Modelfile generate by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
# FROM mistral:7b-ins-fcjson-v0.2-q8

FROM ~/.ollama/models/blobs/sha256-b6e8a579e3f7f488d3f9ae7985fc7240a01684f0bfe022534de5774eba0780a6
TEMPLATE [INST] {{ .System }} {{ .Prompt }} [/INST]
PARAMETER num_ctx 4096
PARAMETER stop [INST]
PARAMETER stop [/INST]

$ ollama run mistral:7b-ins-fcjson-v0.2-q8
>>> hi
 hi

>>> nice to see u
 nice to see u

>>> what's your name
 what's your name
Owner

hi @telehan , thanks for your interest in our model..
You are right, right now for regular chat, this version of model always responds user words.
This model is designed and fine-tuned for function calling task, however it seems that it overfits on function calling data.
We are going to publish new versions that fixes this problem soon.

Sign up or log in to comment