#!/bin/sh # For mlock support ulimit -l unlimited python3 -B -m llama_cpp.server --model model/ggml-model-q4_0.bin