Warlord-K commited on
Commit
7732090
·
1 Parent(s): 8846d1e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -34,7 +34,7 @@ torch_device = "cuda" if torch.cuda.is_available() else "cpu"
34
  print("Running on device:", torch_device)
35
  print("CPU threads:", torch.get_num_threads())
36
 
37
- model_id = "princeton-nlp/Sheared-LLaMA-1.3B"
38
  biencoder = SentenceTransformer("intfloat/e5-large-v2", device=torch_device)
39
  cross_encoder = CrossEncoder("cross-encoder/ms-marco-MiniLM-L-12-v2", max_length=512, device=torch_device)
40
 
@@ -250,6 +250,7 @@ def generate(
250
  query_embedding = create_query_embedding(condensed_query)
251
  relevant_chunks = find_nearest_neighbors(query_embedding)
252
  reranked_relevant_chunks = rerank_chunks_with_cross_encoder(condensed_query, relevant_chunks)
 
253
  qa_prompt = create_qa_prompt(condensed_query, reranked_relevant_chunks)
254
  print(f"{qa_prompt=}")
255
  generator = get_completion(
 
34
  print("Running on device:", torch_device)
35
  print("CPU threads:", torch.get_num_threads())
36
 
37
+ model_id = "princeton-nlp/Sheared-LLaMA-2.7B"
38
  biencoder = SentenceTransformer("intfloat/e5-large-v2", device=torch_device)
39
  cross_encoder = CrossEncoder("cross-encoder/ms-marco-MiniLM-L-12-v2", max_length=512, device=torch_device)
40
 
 
250
  query_embedding = create_query_embedding(condensed_query)
251
  relevant_chunks = find_nearest_neighbors(query_embedding)
252
  reranked_relevant_chunks = rerank_chunks_with_cross_encoder(condensed_query, relevant_chunks)
253
+ print(reranked_relevant_chunks)
254
  qa_prompt = create_qa_prompt(condensed_query, reranked_relevant_chunks)
255
  print(f"{qa_prompt=}")
256
  generator = get_completion(