"use client" import { useEffect, useState, useTransition } from "react" import { cn } from "@/lib/utils" import { TopMenu } from "./interface/top-menu" import { fonts } from "@/lib/fonts" import { useStore } from "./store" import { Zoom } from "./interface/zoom" import { BottomBar } from "./interface/bottom-bar" import { Page } from "./interface/page" import { GeneratedPanel } from "@/types" import { joinWords } from "@/lib/joinWords" import { getStoryContinuation } from "./queries/getStoryContinuation" export default function Main() { const [_isPending, startTransition] = useTransition() const isGeneratingStory = useStore(state => state.isGeneratingStory) const setGeneratingStory = useStore(state => state.setGeneratingStory) const font = useStore(state => state.font) const preset = useStore(state => state.preset) const prompt = useStore(state => state.prompt) const nbPages = useStore(state => state.nbPages) const nbTotalPanels = useStore(state => state.nbTotalPanels) const setPanels = useStore(state => state.setPanels) const setCaptions = useStore(state => state.setCaptions) const zoomLevel = useStore(state => state.zoomLevel) const [waitABitMore, setWaitABitMore] = useState(false) // react to prompt changes useEffect(() => { if (!prompt) { return } startTransition(async () => { setWaitABitMore(false) setGeneratingStory(true) // I don't think we are going to need a rate limiter on the LLM part anymore const enableRateLimiter = false // `${process.env.NEXT_PUBLIC_ENABLE_RATE_LIMITER}` === "true" const [stylePrompt, userStoryPrompt] = prompt.split("||").map(x => x.trim()) // we have to limit the size of the prompt, otherwise the rest of the style won't be followed let limitedStylePrompt = stylePrompt.trim().slice(0, 77).trim() if (limitedStylePrompt.length !== stylePrompt.length) { console.log("Sorry folks, the style prompt was cut to:", limitedStylePrompt) } // new experimental prompt: let's drop the user prompt, and only use the style const lightPanelPromptPrefix = joinWords(preset.imagePrompt(limitedStylePrompt)) // this prompt will be used if the LLM generation failed const degradedPanelPromptPrefix = joinWords([ ...preset.imagePrompt(limitedStylePrompt), // we re-inject the story, then userStoryPrompt ]) let existingPanels: GeneratedPanel[] = [] const newPanelsPrompts: string[] = [] const newCaptions: string[] = [] const nbPanelsToGenerate = 2 for ( let currentPanel = 0; currentPanel < nbTotalPanels; currentPanel += nbPanelsToGenerate ) { try { const candidatePanels = await getStoryContinuation({ preset, stylePrompt, userStoryPrompt, nbPanelsToGenerate, nbTotalPanels, existingPanels, }) console.log("LLM generated some new panels:", candidatePanels) existingPanels.push(...candidatePanels) console.log(`Converting the ${nbPanelsToGenerate} new panels into image prompts..`) const startAt = currentPanel const endAt = currentPanel + nbPanelsToGenerate for (let p = startAt; p < endAt; p++) { newCaptions.push(existingPanels[p]?.caption.trim() || "...") const newPanel = joinWords([ // what we do here is that ideally we give full control to the LLM for prompting, // unless there was a catastrophic failure, in that case we preserve the original prompt existingPanels[p]?.instructions ? lightPanelPromptPrefix : degradedPanelPromptPrefix, existingPanels[p]?.instructions ]) newPanelsPrompts.push(newPanel) console.log(`Image prompt for panel ${p} => "${newPanel}"`) } // update the frontend // console.log("updating the frontend..") setCaptions(newCaptions) setPanels(newPanelsPrompts) setGeneratingStory(false) } catch (err) { console.log("failed to generate the story, aborting here") setGeneratingStory(false) break } if (currentPanel > (nbTotalPanels / 2)) { console.log("good, we are half way there, hold tight!") // setWaitABitMore(true) } } /* setTimeout(() => { setGeneratingStory(false) setWaitABitMore(false) }, enableRateLimiter ? 12000 : 0) */ }) }, [prompt, preset?.label, nbTotalPanels]) // important: we need to react to preset changes too return (