GSridhar1982
commited on
Commit
•
c9dd695
1
Parent(s):
408ac44
Updated model name
Browse files
app.py
CHANGED
@@ -4,7 +4,7 @@ import subprocess
|
|
4 |
from llama_cpp import Llama
|
5 |
|
6 |
# Initialize the model
|
7 |
-
model = Llama(model_path="
|
8 |
|
9 |
def generate_response(prompt):
|
10 |
response = model.create_chat_completion(messages=[{"role": "user", "content": prompt}])
|
@@ -14,7 +14,7 @@ def generate_response(prompt):
|
|
14 |
def predict(text):
|
15 |
# Call llama.cpp with the input text
|
16 |
result = subprocess.run(
|
17 |
-
["./llama.cpp/main", "-m", "
|
18 |
capture_output=True,
|
19 |
text=True
|
20 |
)
|
|
|
4 |
from llama_cpp import Llama
|
5 |
|
6 |
# Initialize the model
|
7 |
+
model = Llama(model_path="QA_llama31_unsloth.Q4_K_M.gguf")
|
8 |
|
9 |
def generate_response(prompt):
|
10 |
response = model.create_chat_completion(messages=[{"role": "user", "content": prompt}])
|
|
|
14 |
def predict(text):
|
15 |
# Call llama.cpp with the input text
|
16 |
result = subprocess.run(
|
17 |
+
["./llama.cpp/main", "-m", "QA_llama31_unsloth.Q4_K_M.gguf", "-p", text],
|
18 |
capture_output=True,
|
19 |
text=True
|
20 |
)
|