Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,20 +1,3 @@
|
|
1 |
-
# import requests
|
2 |
-
# headers = {"Authorization": f"Bearer {API_TOKEN}"}
|
3 |
-
# API_URL = "https://api-inference.huggingface.co/models/kepinsam/ind-to-bbc-nmt-v5"
|
4 |
-
# def query(payload):
|
5 |
-
# response = requests.post(API_URL, headers=headers, json=payload)
|
6 |
-
# return response.json()
|
7 |
-
# data = query({"inputs": ""})
|
8 |
-
|
9 |
-
# ---
|
10 |
-
|
11 |
-
# # Use a pipeline as a high-level helper
|
12 |
-
# from transformers import pipeline
|
13 |
-
|
14 |
-
# pipe = pipeline("text2text-generation", model="kepinsam/ind-to-bbc-nmt-v5")
|
15 |
-
|
16 |
-
# ---
|
17 |
-
|
18 |
import requests
|
19 |
import streamlit as st
|
20 |
|
@@ -27,5 +10,5 @@ def query(payload):
|
|
27 |
return response.json()
|
28 |
|
29 |
output = query({
|
30 |
-
"inputs": "
|
31 |
})
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
import requests
|
2 |
import streamlit as st
|
3 |
|
|
|
10 |
return response.json()
|
11 |
|
12 |
output = query({
|
13 |
+
"inputs": "",
|
14 |
})
|