phucdt89s commited on
Commit
a17b1be
·
verified ·
1 Parent(s): 66cd024

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +21 -21
app.py CHANGED
@@ -1,21 +1,21 @@
1
- import torch
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
- import gradio as gr
4
-
5
- # Load model and tokenizer
6
- MODEL_NAME = "vilm/vinallama-7b-chat"
7
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
8
- model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype=torch.float16, device_map="auto")
9
-
10
- # Define chatbot function
11
- def chatbot(input_text):
12
- inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
13
- outputs = model.generate(inputs["input_ids"], max_length=200, do_sample=True, temperature=0.7)
14
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
15
- return response
16
-
17
- # Create Gradio interface
18
- interface = gr.Interface(fn=chatbot, inputs="text", outputs="text", title="Vinamallama 7B Chatbot")
19
-
20
- # Launch app
21
- interface.launch()
 
1
+ import torch
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer
3
+ import gradio as gr
4
+
5
+ # Load model and tokenizer
6
+ MODEL_NAME = "vilm/vinallama-7b-chat"
7
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
8
+ model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
9
+
10
+ # Define chatbot function
11
+ def chatbot(input_text):
12
+ inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
13
+ outputs = model.generate(inputs["input_ids"], max_length=200, do_sample=True, temperature=0.7)
14
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
15
+ return response
16
+
17
+ # Create Gradio interface
18
+ interface = gr.Interface(fn=chatbot, inputs="text", outputs="text", title="Vinamallama 7B Chatbot")
19
+
20
+ # Launch app
21
+ interface.launch()