Is this model cloned from Photolens

#1
by YanaS - opened

Hey, I see you have liked Photolens/llama-2-7b-langchain-chat model and I wondered is your model here is cloned from his'? I as because I tried to use his model but with a standard prompt template, I got terrible answers and I assume there is some change in the template for chat.
For example, my prompt template looks like this:
[INST]<<SYS>>
You are a helpful assistant, etc.
<</SYS>>

    Context: {context}
    User: {question}

    [/INST]

It works perfectly with TheBloke's model but with this one, I get results formatted like this:

{"action": "Final Answer", "action_input": <text>}
```  ัั–ั‡ะฝั 18, 2023, 15:45 (UTC)

<INST> <text>[/INST] 
```json
{"action": "Final Answer", "action_input": <text>}
```  janvier 18, 2023, 15:47 (UTC)

<INST> <text>}
```  enero 18, 2023, 15:47 (UTC)

<INST> <text> [/INST] ```json
{"action": "Final Answer", "action_input": <text>}
```  janvier 18, 2023, 15:47 (UTC)
...

and it goes on like this for 100 rows.

I would be glad to know if you have used it and how is your prompt formatted, so that you get normal answer.

Sign up or log in to comment