import subprocess | |
import sys | |
# Uninstall the existing llama-cpp-python package | |
# uninstall_command = ["pip", "uninstall", "-y", "llama-cpp-python"] | |
# subprocess.run(uninstall_command, shell=True) | |
# Install llama-cpp-python with CUDA support | |
install_command = ["pip" + " install" + " --upgrade" + " --force-reinstall" + " llama-cpp-python"+ " --no-cache-dir"] | |
install_command_cmake_args = ["CMAKE_ARGS=-DLLAMA_CUBLAS=on" + " FORCE_CMAKE=1"] | |
install_command_cmake_args.extend(install_command) | |
subprocess.run(install_command, shell=True) | |
# Start the Hugging Face Space | |
uvicorn_command = ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"] | |
subprocess.run(uvicorn_command, shell=True) | |
subprocess.run(install_command, shell=True) | |