import gradio as gr import numpy as np import soundfile as sf from datetime import datetime from time import time as ttime from my_utils import load_audio from transformers import pipeline from text.cleaner import clean_text from polyglot.detect import Detector from feature_extractor import cnhubert from timeit import default_timer as timer from text import cleaned_text_to_sequence from module.models import SynthesizerTrn from module.mel_processing import spectrogram_torch from transformers.pipelines.audio_utils import ffmpeg_read import os,re,sys,LangSegment,librosa,pdb,torch,pytz,random from transformers import AutoModelForMaskedLM, AutoTokenizer from AR.models.t2s_lightning_module import Text2SemanticLightningModule import logging logging.getLogger("markdown_it").setLevel(logging.ERROR) logging.getLogger("urllib3").setLevel(logging.ERROR) logging.getLogger("httpcore").setLevel(logging.ERROR) logging.getLogger("httpx").setLevel(logging.ERROR) logging.getLogger("asyncio").setLevel(logging.ERROR) logging.getLogger("charset_normalizer").setLevel(logging.ERROR) logging.getLogger("torchaudio._extension").setLevel(logging.ERROR) logging.getLogger("multipart").setLevel(logging.WARNING) from download import * download() from TTS_infer_pack.TTS import TTS, TTS_Config from TTS_infer_pack.text_segmentation_method import get_method if "_CUDA_VISIBLE_DEVICES" in os.environ: os.environ["CUDA_VISIBLE_DEVICES"] = os.environ["_CUDA_VISIBLE_DEVICES"] tz = pytz.timezone('Asia/Singapore') device = "cuda" if torch.cuda.is_available() else "cpu" def abs_path(dir): global_dir = os.path.dirname(os.path.abspath(sys.argv[0])) return(os.path.join(global_dir, dir)) gpt_path = abs_path("MODELS/22/22.ckpt") sovits_path=abs_path("MODELS/22/22.pth") cnhubert_base_path = os.environ.get("cnhubert_base_path", "pretrained_models/chinese-hubert-base") bert_path = os.environ.get("bert_path", "pretrained_models/chinese-roberta-wwm-ext-large") if not os.path.exists(cnhubert_base_path): cnhubert_base_path = "TencentGameMate/chinese-hubert-base" if not os.path.exists(bert_path): bert_path = "hfl/chinese-roberta-wwm-ext-large" cnhubert.cnhubert_base_path = cnhubert_base_path whisper_path = os.environ.get("whisper_path", "pretrained_models/whisper-tiny") if not os.path.exists(whisper_path): whisper_path = "openai/whisper-tiny" pipe = pipeline( task="automatic-speech-recognition", model=whisper_path, chunk_length_s=30, device=device,) is_half = eval( os.environ.get("is_half", "True" if torch.cuda.is_available() else "False") ) dict_language = { "中文1": "all_zh", "English": "en", "日文1": "all_ja", "中文": "zh", "日本語": "ja", "混合": "auto", } cut_method = { "Do not split/不切":"cut0", "Split into groups of 4 sentences/四句一切": "cut1", "Split every 50 characters/50字一切": "cut2", "Split at CN/JP periods (。)/按中日文句号切": "cut3", "Split at English periods (.)/按英文句号切": "cut4", "Split at punctuation marks/按标点切": "cut5", } tts_config = TTS_Config("GPT_SoVITS/configs/tts_infer.yaml") tts_config.device = device tts_config.is_half = is_half if gpt_path is not None: tts_config.t2s_weights_path = gpt_path if sovits_path is not None: tts_config.vits_weights_path = sovits_path if cnhubert_base_path is not None: tts_config.cnhuhbert_base_path = cnhubert_base_path if bert_path is not None: tts_config.bert_base_path = bert_path tts_pipline = TTS(tts_config) gpt_path = tts_config.t2s_weights_path sovits_path = tts_config.vits_weights_path def inference(text, text_lang, ref_audio_path, prompt_text, prompt_lang, top_k, top_p, temperature, text_split_method, batch_size, speed_factor, ref_text_free, split_bucket, volume ): if not duration(ref_audio_path): return None if text == '': wprint("Please input text to generate/请输入生成文字") return None text=trim_text(text,text_language) try: lang=dict_language[text_lang] inputs={ "text": text, "text_lang": lang, "ref_audio_path": ref_audio_path, "prompt_text": prompt_text if not ref_text_free else "", "prompt_lang": dict_language[prompt_lang], "top_k": top_k, "top_p": top_p, "temperature": temperature, "text_split_method": cut_method[text_split_method], "batch_size":int(batch_size), "speed_factor":float(speed_factor), "split_bucket":split_bucket, "volume":volume, "return_fragment":False, } yield next(tts_pipline.run(inputs)) except KeyError as e: wprint(f'Unsupported language type:{e}') return None #==========custom functions============ splits = {",", "。", "?", "!", ",", ".", "?", "!", "~", ":", ":", "—", "…", } def tprint(text): now=datetime.now(tz).strftime('%H:%M:%S') print(f'UTC+8 - {now} - {text}') def wprint(text): tprint(text) gr.Warning(text) def lang_detector(text): min_chars = 5 if len(text) < min_chars: return "Input text too short/输入文本太短" try: detector = Detector(text).language lang_info = str(detector) code = re.search(r"name: (\w+)", lang_info).group(1) if code == 'Japanese': return "日本語" elif code == 'Chinese': return "中文" elif code == 'English': return 'English' else: return code except Exception as e: return f"ERROR:{str(e)}" def trim_text(text,language): limit_cj = 120 #character limit_en = 60 #words search_limit_cj = limit_cj+30 search_limit_en = limit_en +30 text = text.replace('\n', '').strip() if language =='English': words = text.split() if len(words) <= limit_en: return text # English for i in range(limit_en, -1, -1): if any(punct in words[i] for punct in splits): return ' '.join(words[:i+1]) for i in range(limit_en, min(len(words), search_limit_en)): if any(punct in words[i] for punct in splits): return ' '.join(words[:i+1]) return ' '.join(words[:limit_en]) else: if len(text) <= limit_cj: return text for i in range(limit_cj, -1, -1): if text[i] in splits: return text[:i+1] for i in range(limit_cj, min(len(text), search_limit_cj)): if text[i] in splits: return text[:i+1] return text[:limit_cj] def duration(audio_file_path): if not audio_file_path: wprint("Failed to obtain uploaded audio/未找到音频文件") return False try: audio_duration = librosa.get_duration(filename=audio_file_path) if not 3 < audio_duration < 10: wprint("The audio length must be between 3~10 seconds/音频时长须在3~10秒之间") return False return True except FileNotFoundError: return False def update_model(choice): #global tts_config.vits_weights_path, tts_config.t2s_weights_path model_info = models[choice] gpt_path = abs_path(model_info["gpt_weight"]) sovits_path = abs_path(model_info["sovits_weight"]) tts_pipline.init_vits_weights(sovits_path) tts_pipline.init_t2s_weights(gpt_path) model_name = choice tone_info = model_info["tones"]["tone1"] tone_sample_path = abs_path(tone_info["sample"]) tprint(f'✅SELECT MODEL:{choice}') # 返回默认tone“tone1” return ( tone_info["example_voice_wav"], tone_info["example_voice_wav_words"], model_info["default_language"], model_info["default_language"], model_name, "tone1" , tone_sample_path ) def update_tone(model_choice, tone_choice): model_info = models[model_choice] tone_info = model_info["tones"][tone_choice] example_voice_wav = abs_path(tone_info["example_voice_wav"]) example_voice_wav_words = tone_info["example_voice_wav_words"] tone_sample_path = abs_path(tone_info["sample"]) return example_voice_wav, example_voice_wav_words,tone_sample_path def transcribe(voice): time1=timer() tprint('⚡Start Clone - transcribe') task="transcribe" if voice is None: wprint("No audio file submitted! Please upload or record an audio file before submitting your request.") R = pipe(voice, batch_size=8, generate_kwargs={"task": task}, return_timestamps=True,return_language=True) text=R['text'] lang=R['chunks'][0]['language'] if lang=='english': language='English' elif lang =='chinese': language='中文' elif lang=='japanese': language = '日本語' time2=timer() tprint(f'transcribe COMPLETE,{round(time2-time1,4)}s') tprint(f' \nTranscribe result:\n 🔣Language:{language} \n 🔣Text:{text}' ) return text,language def clone_voice(user_voice,user_text,user_lang): if not duration(user_voice): return None if user_text == '': wprint("Please enter text to generate/请输入生成文字") return None user_text=trim_text(user_text,user_lang) global gpt_path, sovits_path gpt_path = abs_path("pretrained_models/s1bert25hz-2kh-longer-epoch=68e-step=50232.ckpt") #tprint(f'Model loaded:{gpt_path}') sovits_path = abs_path("pretrained_models/s2G488k.pth") #tprint(f'Model loaded:{sovits_path}') try: prompt_text, prompt_lang = transcribe(user_voice) except UnboundLocalError as e: wprint(f"The language in the audio cannot be recognized :{str(e)}") return None tts_pipline.init_vits_weights(sovits_path) tts_pipline.init_t2s_weights(gpt_path) inputs={ "text": user_text, "text_lang": dict_language[user_lang], "ref_audio_path": user_voice, "prompt_text": prompt_text, "prompt_lang": dict_language[prompt_lang], "top_k": 5, "top_p": 1, "temperature": 1, "text_split_method": "cut1", "batch_size":20, "speed_factor":1.0, "split_bucket":True, "volume":1.0, "return_fragment":False, } yield next(tts_pipline.run(inputs)) with open('dummy') as f: dummy_txt = f.read().strip().splitlines() def dice(): return random.choice(dummy_txt), '🎲' from info import models models_by_language = { "English": [], "中文": [], "日本語": [] } for model_name, model_info in models.items(): language = model_info["default_language"] models_by_language[language].append((model_name, model_info)) ##########GRADIO########### with gr.Blocks(theme='Kasien/ali_theme_custom') as app: gr.HTML('''
If you like this space, please click the ❤️ at the top of the page..如喜欢,请点一下页面顶部的❤️
Need 3~10s audio.This involves voice-to-text conversion followed by text-to-voice conversion, so it takes longer time
需要3~10秒语音,这个会涉及语音转文字,之后再转语音,所以耗时比较久