MakiAi commited on
Commit
ead5e82
·
verified ·
1 Parent(s): c06b9fa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -24
app.py CHANGED
@@ -1,30 +1,12 @@
1
- import gradio as gr
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
- import torch
4
-
5
- # Base modelとadapterの設定
6
- base_model_id = "unsloth/Llama-3.2-3B" # または適切なbase model
7
- adapter_model_id = "MakiAi/Llama-3.2-3B-Instruct-bnb-4bit-OKU_wiki_llama3.1_8b_inst_Reflexive_chunk200_overlap700-10epochs"
8
 
9
- # モデルとトークナイザーの読み込み
10
- model = AutoModelForCausalLM.from_pretrained(
11
- adapter_model_id,
12
- torch_dtype=torch.float16,
13
- device_map="auto"
14
- )
15
- tokenizer = AutoTokenizer.from_pretrained(adapter_model_id)
16
 
17
- # Gradioインターフェースの作成
18
- def generate_text(prompt):
19
- inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
20
- outputs = model.generate(**inputs, max_new_tokens=100)
21
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
22
 
23
- demo = gr.Interface(
24
- fn=generate_text,
25
- inputs="text",
26
- outputs="text"
27
- )
28
 
29
  if __name__ == "__main__":
30
  demo.launch()
 
 
 
 
 
 
 
 
1
 
2
+ import gradio as gr
3
+ import spaces
4
+ import transformers_gradio
 
 
 
 
5
 
6
+ #demo = gr.load(name="MakiAi/Llama-3-2-3B-Instruct-bnb-4bit-OKU-v1-10epochs-adapter", src=transformers_gradio.registry)
7
+ demo = gr.load(name="unsloth/Llama-3.2-1B", src=transformers_gradio.registry)
 
 
 
8
 
9
+ demo.fn = spaces.GPU()(demo.fn)
 
 
 
 
10
 
11
  if __name__ == "__main__":
12
  demo.launch()