# Start Ollama server in the background | |
ollama serve --port 7860 & | |
# Wait for the server to be ready | |
while ! nc -z localhost 7860; do | |
echo "Waiting for Ollama server to start..." | |
sleep 1 | |
done | |
# Pull the model | |
echo "Pulling the model..." | |
ollama pull steamdj/llama3.1-cpu-only | |
# Keep the container running | |
wait | |