Upload install_requirements.sh
Browse files- install_requirements.sh +18 -0
install_requirements.sh
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#!/bin/bash
|
2 |
+
export PYTHONNOUSERSITE=1
|
3 |
+
git submodule update --init --recursive
|
4 |
+
if [[ $1 = "cuda" || $1 = "CUDA" ]]; then
|
5 |
+
wget -qO- https://micromamba.snakepit.net/api/micromamba/linux-64/latest | tar -xvj bin/micromamba
|
6 |
+
bin/micromamba create -f environments/huggingface.yml -r runtime -n koboldai -y
|
7 |
+
# Weird micromamba bug causes it to fail the first time, running it twice just to be safe, the second time is much faster
|
8 |
+
bin/micromamba create -f environments/huggingface.yml -r runtime -n koboldai -y
|
9 |
+
exit
|
10 |
+
fi
|
11 |
+
if [[ $1 = "rocm" || $1 = "ROCM" ]]; then
|
12 |
+
wget -qO- https://micromamba.snakepit.net/api/micromamba/linux-64/latest | tar -xvj bin/micromamba
|
13 |
+
bin/micromamba create -f environments/rocm.yml -r runtime -n koboldai-rocm -y
|
14 |
+
# Weird micromamba bug causes it to fail the first time, running it twice just to be safe, the second time is much faster
|
15 |
+
bin/micromamba create -f environments/rocm.yml -r runtime -n koboldai-rocm -y
|
16 |
+
exit
|
17 |
+
fi
|
18 |
+
echo Please specify either CUDA or ROCM
|