jbilcke-hf HF staff commited on
Commit
9f66c7d
1 Parent(s): 3220193
src/app/queries/predictNextPanels.ts CHANGED
@@ -61,19 +61,21 @@ export const predictNextPanels = async ({
61
  try {
62
  // console.log(`calling predict(${query}, ${nbTotalPanels})`)
63
  result = `${await predict(query, nbPanelsToGenerate) || ""}`.trim()
 
64
  if (!result.length) {
65
- throw new Error("empty result!")
66
  }
67
  } catch (err) {
68
  // console.log(`prediction of the story failed, trying again..`)
69
  try {
70
- result = `${await predict(query+".", nbPanelsToGenerate) || ""}`.trim()
 
71
  if (!result.length) {
72
- throw new Error("empty result!")
73
  }
74
  } catch (err) {
75
- console.error(`prediction of the story failed again 💩`)
76
- throw new Error(`failed to generate the story ${err}`)
77
  }
78
  }
79
 
 
61
  try {
62
  // console.log(`calling predict(${query}, ${nbTotalPanels})`)
63
  result = `${await predict(query, nbPanelsToGenerate) || ""}`.trim()
64
+ console.log("LLM result (1st trial):", result)
65
  if (!result.length) {
66
+ throw new Error("empty result on 1st trial!")
67
  }
68
  } catch (err) {
69
  // console.log(`prediction of the story failed, trying again..`)
70
  try {
71
+ result = `${await predict(query + " \n ", nbPanelsToGenerate) || ""}`.trim()
72
+ console.log("LLM result (2nd trial):", result)
73
  if (!result.length) {
74
+ throw new Error("empty result on 2nd trial!")
75
  }
76
  } catch (err) {
77
+ console.error(`prediction of the story failed twice 💩`)
78
+ throw new Error(`failed to generate the story twice 💩 ${err}`)
79
  }
80
  }
81
 
src/app/queries/predictWithHuggingFace.ts CHANGED
@@ -14,7 +14,7 @@ export async function predict(inputs: string, nbPanels: number): Promise<string>
14
 
15
  // we don't require a lot of token for our task
16
  // but to be safe, let's count ~110 tokens per panel
17
- const nbMaxNewTokens = nbPanels * 110
18
 
19
  switch (llmEngine) {
20
  case "INFERENCE_ENDPOINT":
 
14
 
15
  // we don't require a lot of token for our task
16
  // but to be safe, let's count ~110 tokens per panel
17
+ const nbMaxNewTokens = nbPanels * 130 // 110 isn't enough anymore for long dialogues
18
 
19
  switch (llmEngine) {
20
  case "INFERENCE_ENDPOINT":