@echo off
SETLOCAL
REM Starts a direct chat with the GLM-4 9B model.
REM NOTE: Direct chat with this model may produce suboptimal results.
REM Using the server mode with a compatible client is recommended.
SET MODEL_PATH=gemma_cli\glm-4-9b-chat-Q4_K_M.gguf
SET NGL=15
echo [INFO] Loading GLM-4 9B for chat...
echo [INFO] Offloading %NGL% layers to GPU...
build\bin\Release\llama-cli.exe -m "%MODEL_PATH%" -ngl %NGL% -t 12 -c 4096 --temp 0.7 --color --interactive-first
pause > nul
ENDLOCAL