huggi / src /lib /buildPrompt.ts
nsarrazin's picture
nsarrazin HF staff
Backend refactor for streaming endpoints (#444)
e943a05 unverified
raw
history blame
1.46 kB
import type { BackendModel } from "./server/models";
import type { Message } from "./types/Message";
import { format } from "date-fns";
import type { WebSearch } from "./types/WebSearch";
/**
* Convert [{user: "assistant", content: "hi"}, {user: "user", content: "hello"}] to:
*
* <|assistant|>hi<|endoftext|><|prompter|>hello<|endoftext|><|assistant|>
*/
interface buildPromptOptions {
messages: Pick<Message, "from" | "content">[];
model: BackendModel;
locals?: App.Locals;
webSearch?: WebSearch;
preprompt?: string;
}
export async function buildPrompt({
messages,
model,
webSearch,
preprompt,
}: buildPromptOptions): Promise<string> {
if (webSearch && webSearch.context) {
const messagesWithoutLastUsrMsg = messages.slice(0, -1);
const lastUserMsg = messages.slice(-1)[0];
const currentDate = format(new Date(), "MMMM d, yyyy");
messages = [
...messagesWithoutLastUsrMsg,
{
from: "user",
content: `Please answer my question "${lastUserMsg.content}" using the supplied context below (paragraphs from various websites). For the context, today is ${currentDate}:
=====================
${webSearch.context}
=====================
So my question is "${lastUserMsg.content}"`,
},
];
}
return (
model
.chatPromptRender({ messages, preprompt })
// Not super precise, but it's truncated in the model's backend anyway
.split(" ")
.slice(-(model.parameters?.truncate ?? 0))
.join(" ")
);
}