Update app.py
Browse files
app.py
CHANGED
@@ -3,7 +3,6 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
3 |
|
4 |
model_name = "Llama-2-7b-finetuned-with-QLoRa"
|
5 |
|
6 |
-
# Load model and tokenizer
|
7 |
@st.cache_resource
|
8 |
def load_model_and_tokenizer(model_name):
|
9 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
@@ -34,7 +33,7 @@ def main():
|
|
34 |
# Generate button
|
35 |
if st.sidebar.button("Generate Response"):
|
36 |
with st.spinner("Generating response..."):
|
37 |
-
response = generate_response(topic)
|
38 |
st.subheader(f"Generated response on '{topic}':")
|
39 |
st.write(response)
|
40 |
|
|
|
3 |
|
4 |
model_name = "Llama-2-7b-finetuned-with-QLoRa"
|
5 |
|
|
|
6 |
@st.cache_resource
|
7 |
def load_model_and_tokenizer(model_name):
|
8 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
|
|
33 |
# Generate button
|
34 |
if st.sidebar.button("Generate Response"):
|
35 |
with st.spinner("Generating response..."):
|
36 |
+
response = generate_response(f"[INST] {topic} [/INST]" )
|
37 |
st.subheader(f"Generated response on '{topic}':")
|
38 |
st.write(response)
|
39 |
|