Ethan Chang commited on
Commit
8ea200f
·
1 Parent(s): 6a3dbe6

Removed minor print statement, logging is possible but unimplemented for the current solution

Browse files
code/modules/chat/chat_model_loader.py CHANGED
@@ -32,7 +32,6 @@ class ChatModelLoader:
32
  elif self.config["llm_params"]["llm_loader"] == "local_llm":
33
  n_batch = 512 # Should be between 1 and n_ctx, consider the amount of VRAM in your GPU.
34
  model_path = self._verify_model_cache(self.config["llm_params"]["local_llm_params"]["model"])
35
- print(model_path)
36
  llm = LlamaCpp(
37
  model_path=model_path,
38
  n_batch=n_batch,
 
32
  elif self.config["llm_params"]["llm_loader"] == "local_llm":
33
  n_batch = 512 # Should be between 1 and n_ctx, consider the amount of VRAM in your GPU.
34
  model_path = self._verify_model_cache(self.config["llm_params"]["local_llm_params"]["model"])
 
35
  llm = LlamaCpp(
36
  model_path=model_path,
37
  n_batch=n_batch,