#!/bin/sh # For mlock support ulimit -l unlimited python3 -B main.py #python3 -m llama_cpp.server --model ./model/gguf-model.bin --host 0.0.0.0 --port 2600