Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,7 +3,6 @@ from huggingface_hub import InferenceClient
|
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
4 |
import torch
|
5 |
import os
|
6 |
-
import pyttsx3 # Importing pyttsx3 for text-to-speech
|
7 |
|
8 |
# Replace 'your_huggingface_token' with your actual Hugging Face access token
|
9 |
access_token = os.getenv('token')
|
@@ -20,8 +19,6 @@ model.eval() # Set the model to evaluation mode
|
|
20 |
# Initialize the inference client (if needed for other API-based tasks)
|
21 |
client = InferenceClient(token=access_token)
|
22 |
|
23 |
-
# Initialize the text-to-speech engine
|
24 |
-
tts_engine = pyttsx3.init()
|
25 |
|
26 |
# Import required modules for E2-F5-TTS
|
27 |
from huggingface_hub import Client
|
|
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
4 |
import torch
|
5 |
import os
|
|
|
6 |
|
7 |
# Replace 'your_huggingface_token' with your actual Hugging Face access token
|
8 |
access_token = os.getenv('token')
|
|
|
19 |
# Initialize the inference client (if needed for other API-based tasks)
|
20 |
client = InferenceClient(token=access_token)
|
21 |
|
|
|
|
|
22 |
|
23 |
# Import required modules for E2-F5-TTS
|
24 |
from huggingface_hub import Client
|