Spaces:
Sleeping
Sleeping
fix
Browse files- app.py +0 -11
- requirements.txt +1 -2
app.py
CHANGED
@@ -3,9 +3,6 @@ import gradio as gr
|
|
3 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
4 |
import torch
|
5 |
|
6 |
-
# Gradio client kullanımı için gerekli kod
|
7 |
-
from gradio_client import Client
|
8 |
-
|
9 |
MODEL_NAME = "osmankoc/llama-2-7b-zoa"
|
10 |
|
11 |
# Model ve tokenizer'ı önceden yükle
|
@@ -38,11 +35,3 @@ demo = gr.Interface(
|
|
38 |
demo.add_api_route("/predict", generate)
|
39 |
|
40 |
demo.launch()
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
client = Client("osmankoc/llama2-zoa-api")
|
45 |
-
result = client.predict(
|
46 |
-
prompt="Hello!!",
|
47 |
-
api_name="/predict"
|
48 |
-
)
|
|
|
3 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
4 |
import torch
|
5 |
|
|
|
|
|
|
|
6 |
MODEL_NAME = "osmankoc/llama-2-7b-zoa"
|
7 |
|
8 |
# Model ve tokenizer'ı önceden yükle
|
|
|
35 |
demo.add_api_route("/predict", generate)
|
36 |
|
37 |
demo.launch()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
requirements.txt
CHANGED
@@ -2,5 +2,4 @@ fastapi
|
|
2 |
uvicorn
|
3 |
torch
|
4 |
transformers
|
5 |
-
accelerate
|
6 |
-
gradio_client
|
|
|
2 |
uvicorn
|
3 |
torch
|
4 |
transformers
|
5 |
+
accelerate
|
|