Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Update app.py
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ from auditqa.sample_questions import QUESTIONS
|
|
11 |
from auditqa.reports import files, report_list, new_files, new_report_list
|
12 |
from auditqa.process_chunks import load_chunks, getconfig, get_local_qdrant, load_new_chunks
|
13 |
from auditqa.retriever import get_context
|
14 |
-
from auditqa.reader import nvidia_client, dedicated_endpoint
|
15 |
from auditqa.utils import make_html_source, parse_output_llm_with_sources, save_logs, get_message_template, get_client_location, get_client_ip, get_platform_info
|
16 |
from dotenv import load_dotenv
|
17 |
load_dotenv()
|
@@ -242,7 +242,7 @@ async def chat(query,history,sources,reports,subtype,year, client_ip=None, sessi
|
|
242 |
# yield update
|
243 |
|
244 |
else:
|
245 |
-
chat_model =
|
246 |
async def process_stream():
|
247 |
nonlocal answer_yet
|
248 |
try:
|
|
|
11 |
from auditqa.reports import files, report_list, new_files, new_report_list
|
12 |
from auditqa.process_chunks import load_chunks, getconfig, get_local_qdrant, load_new_chunks
|
13 |
from auditqa.retriever import get_context
|
14 |
+
from auditqa.reader import nvidia_client, dedicated_endpoint, serverless_api
|
15 |
from auditqa.utils import make_html_source, parse_output_llm_with_sources, save_logs, get_message_template, get_client_location, get_client_ip, get_platform_info
|
16 |
from dotenv import load_dotenv
|
17 |
load_dotenv()
|
|
|
242 |
# yield update
|
243 |
|
244 |
else:
|
245 |
+
chat_model = serverless_api() # TESTING: ADAPTED FOR HF INFERENCE API (needs to be reverted for production version)
|
246 |
async def process_stream():
|
247 |
nonlocal answer_yet
|
248 |
try:
|