Spaces:
Running
on
Zero
Running
on
Zero
Commit
·
ca208c8
1
Parent(s):
9d9ede7
Add debug print for full predict question in prediction function
Browse files
app.py
CHANGED
@@ -195,6 +195,7 @@ def predict(message,
|
|
195 |
question = '<image>\n'+message
|
196 |
else:
|
197 |
question = message
|
|
|
198 |
response, conv_history = model.chat(tokenizer, pixel_values, question, generation_config, history=history, return_history=True)
|
199 |
return response, conv_history
|
200 |
|
|
|
195 |
question = '<image>\n'+message
|
196 |
else:
|
197 |
question = message
|
198 |
+
print(f"FULL predict question: {question}")
|
199 |
response, conv_history = model.chat(tokenizer, pixel_values, question, generation_config, history=history, return_history=True)
|
200 |
return response, conv_history
|
201 |
|