Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -9,7 +9,8 @@ from langchain.chains.question_answering import load_qa_chain
|
|
9 |
from langchain.prompts import PromptTemplate
|
10 |
import os
|
11 |
import json
|
12 |
-
|
|
|
13 |
|
14 |
####CREDIT#####
|
15 |
#Credit to author (Sri Laxmi) of original code reference: SriLaxmi1993
|
@@ -19,7 +20,13 @@ from transformers import AutoTokenizer
|
|
19 |
os.system("pip install -r requirements.txt")
|
20 |
|
21 |
#some model
|
22 |
-
tokenizer = AutoTokenizer.from_pretrained("bofenghuang/vigogne-2-7b-chat")
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
|
24 |
st.set_page_config(page_title="Gemini RAG", layout="wide")
|
25 |
|
|
|
9 |
from langchain.prompts import PromptTemplate
|
10 |
import os
|
11 |
import json
|
12 |
+
|
13 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig, TextStreamer
|
14 |
|
15 |
####CREDIT#####
|
16 |
#Credit to author (Sri Laxmi) of original code reference: SriLaxmi1993
|
|
|
20 |
os.system("pip install -r requirements.txt")
|
21 |
|
22 |
#some model
|
23 |
+
#tokenizer = AutoTokenizer.from_pretrained("bofenghuang/vigogne-2-7b-chat")
|
24 |
+
model_name_or_path = "bofenghuang/vigogne-2-7b-chat"
|
25 |
+
revision = "v2.0"
|
26 |
+
|
27 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, revision=revision, padding_side="right", use_fast=False)
|
28 |
+
model = AutoModelForCausalLM.from_pretrained(model_name_or_path, revision=revision, torch_dtype=torch.float16, device_map="auto")
|
29 |
+
|
30 |
|
31 |
st.set_page_config(page_title="Gemini RAG", layout="wide")
|
32 |
|