GSridhar1982 commited on
Commit
c9dd695
1 Parent(s): 408ac44

Updated model name

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -4,7 +4,7 @@ import subprocess
4
  from llama_cpp import Llama
5
 
6
  # Initialize the model
7
- model = Llama(model_path="QA_Llama31_Quantized_GGUF")
8
 
9
  def generate_response(prompt):
10
  response = model.create_chat_completion(messages=[{"role": "user", "content": prompt}])
@@ -14,7 +14,7 @@ def generate_response(prompt):
14
  def predict(text):
15
  # Call llama.cpp with the input text
16
  result = subprocess.run(
17
- ["./llama.cpp/main", "-m", "QA_Llama31_Quantized_GGUF", "-p", text],
18
  capture_output=True,
19
  text=True
20
  )
 
4
  from llama_cpp import Llama
5
 
6
  # Initialize the model
7
+ model = Llama(model_path="QA_llama31_unsloth.Q4_K_M.gguf")
8
 
9
  def generate_response(prompt):
10
  response = model.create_chat_completion(messages=[{"role": "user", "content": prompt}])
 
14
  def predict(text):
15
  # Call llama.cpp with the input text
16
  result = subprocess.run(
17
+ ["./llama.cpp/main", "-m", "QA_llama31_unsloth.Q4_K_M.gguf", "-p", text],
18
  capture_output=True,
19
  text=True
20
  )