updated README.md on how to make changes
#17
by
absy
- opened
README.md
CHANGED
@@ -123,6 +123,22 @@ If you want to have a chat-style conversation, replace the `-p <PROMPT>` argumen
|
|
123 |
|
124 |
Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
|
125 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
126 |
<!-- footer start -->
|
127 |
## Discord
|
128 |
|
|
|
123 |
|
124 |
Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md).
|
125 |
|
126 |
+
## How to make changes.
|
127 |
+
|
128 |
+
Downloading GGML files from repositories can be done in a more efficient way than using git clone. Here's how:
|
129 |
+
|
130 |
+
1. Avoid Git Clone: It's not recommended to use git clone for GGML repositories as it downloads all files, resulting in unnecessary data storage.
|
131 |
+
2. Use wget or Download Client: Instead, use the wget command or any download client to directly grab the specific file you need. This prevents excessive data usage and storage.
|
132 |
+
3. Python Code Option: If you prefer Python, you can use the hf_hub_download function from huggingface_hub. Here's an example:
|
133 |
+
|
134 |
+
```
|
135 |
+
from huggingface_hub import hf_hub_download
|
136 |
+
hf_hub_download(repo_id='TheBloke/Llama-2-7B-Chat-GGML', filename='llama-2-7b-chat.ggmlv3.q4_K_M.bin')
|
137 |
+
#This code snippet downloads the specified GGML file directly.
|
138 |
+
```
|
139 |
+
|
140 |
+
4. Inference from Python: If you plan to perform inference from Python, the ctransformers Python library is a convenient option. It handles automatic file downloads for GGML inference.
|
141 |
+
|
142 |
<!-- footer start -->
|
143 |
## Discord
|
144 |
|