orca_mini_v3_13B-GGML / start_server.sh
limcheekin's picture
feat: updated to 13B ggmlv3.q6_K model
a5da896
raw
history blame
146 Bytes
#!/bin/sh
# For mlock support
ulimit -l unlimited
python3 -B -m llama_cpp.server --model model/ggmlv3-model.bin --n_threads 2 --embedding False