Model outputs erroneous

#6
by Kurapika993 - opened

Working poorly to follow instructions which is unexpected for a 65b parameter model. Due you have any suggested prompting template ?

As it's an Alpaca model, try this:

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
Prompt goes here

### Response:

I have two other 65B models you could try as well. For example gpt4-alpaca-lora_mlp is also trained on Alpaca, but using GPT4 responses instead of GPT3.5. So it's generally higher quality.

You're right that a 65B should be pretty good at instructions. Even base Llama 65B can respond well to instructions when given a suitable prompt. However as this has been fine tuned, it may be that it only responds well to the exact prompt template; so if you weren't trying the above, it may work better if you do.

Kurapika993 changed discussion status to closed

Sign up or log in to comment