Spaces:
Sleeping
Sleeping
File size: 5,631 Bytes
5548515 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 |
# Copyright (c) 2023 Amphion.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
######## Build Experiment Environment ###########
exp_dir=$(cd `dirname $0`; pwd)
work_dir=$(dirname $(dirname $(dirname $exp_dir)))
export WORK_DIR=$work_dir
export PYTHONPATH=$work_dir
export PYTHONIOENCODING=UTF-8
cd $work_dir/modules/monotonic_align
mkdir -p monotonic_align
python setup.py build_ext --inplace
cd $work_dir
######## Parse the Given Parameters from the Commond ###########
options=$(getopt -o c:n:s --long gpu:,config:,infer_expt_dir:,ar_model_ckpt_dir:,infer_output_dir:,infer_mode:,infer_test_list_file:,infer_text:,infer_text_prompt:,infer_audio_prompt:,model_train_stage:,name:,stage: -- "$@")
eval set -- "$options"
while true; do
case $1 in
# Experimental Configuration File
-c | --config) shift; exp_config=$1 ; shift ;;
# Experimental Name
-n | --name) shift; exp_name=$1 ; shift ;;
# Running Stage
-s | --stage) shift; running_stage=$1 ; shift ;;
# Visible GPU machines. The default value is "0".
--gpu) shift; gpu=$1 ; shift ;;
# [Only for Training] Model training stage.
--model_train_stage) shift; model_train_stage=$1 ; shift ;;
# [Only for Training] The stage1 ckpt dir. The value is like "[Your path to save logs and checkpoints]/[YourExptName]"
--ar_model_ckpt_dir) shift; ar_model_ckpt_dir=$1 ; shift ;;
# [Only for Inference] The experiment dir. The value is like "[Your path to save logs and checkpoints]/[YourExptName]"
--infer_expt_dir) shift; infer_expt_dir=$1 ; shift ;;
# [Only for Inference] The output dir to save inferred audios. Its default value is "$expt_dir/result"
--infer_output_dir) shift; infer_output_dir=$1 ; shift ;;
# [Only for Inference] The inference mode. It can be "batch" to generate speech by batch, or "single" to generage a single clip of speech.
--infer_mode) shift; infer_mode=$1 ; shift ;;
# [Only for Inference] The inference test list file. It is only used when the inference model is "batch".
--infer_test_list_file) shift; infer_test_list_file=$1 ; shift ;;
# [Only for Inference] The text to be synthesized from. It is only used when the inference model is "single".
--infer_text) shift; infer_text=$1 ; shift ;;
# [Only for Inference] The inference text prompt. It is only used when the inference model is "single".
--infer_text_prompt) shift; infer_text_prompt=$1 ; shift ;;
# [Only for Inference] The inference audio prompt. It is only used when the inference model is "single".
--infer_audio_prompt) shift; infer_audio_prompt=$1 ; shift ;;
--) shift ; break ;;
*) echo "Invalid option: $1" exit 1 ;;
esac
done
### Value check ###
if [ -z "$running_stage" ]; then
echo "[Error] Please specify the running stage"
exit 1
fi
if [ -z "$exp_config" ]; then
exp_config="${exp_dir}"/exp_config.json
fi
echo "Exprimental Configuration File: $exp_config"
if [ -z "$gpu" ]; then
gpu="0"
fi
######## Features Extraction ###########
if [ $running_stage -eq 1 ]; then
CUDA_VISIBLE_DEVICES=$gpu python "${work_dir}"/bins/tts/preprocess.py \
--config=$exp_config \
--num_workers=4
fi
######## Training ###########
if [ $running_stage -eq 2 ]; then
if [ -z "$exp_name" ]; then
echo "[Error] Please specify the experiments name"
exit 1
fi
if [ "$model_train_stage" = "2" ] && [ -z "$ar_model_ckpt_dir" ]; then
echo "[Error] Please specify the ckeckpoint path to the trained model in stage1."
exit 1
fi
if [ "$model_train_stage" = "1" ]; then
ar_model_ckpt_dir=None
fi
echo "Exprimental Name: $exp_name"
CUDA_VISIBLE_DEVICES=$gpu accelerate launch --main_process_port 29510 \
"${work_dir}"/bins/tts/train.py \
--config $exp_config \
--exp_name $exp_name \
--log_level debug \
--train_stage $model_train_stage \
--checkpoint_path $ar_model_ckpt_dir
fi
######## Inference ###########
if [ $running_stage -eq 3 ]; then
if [ -z "$infer_expt_dir" ]; then
echo "[Error] Please specify the experimental directionary. The value is like [Your path to save logs and checkpoints]/[YourExptName]"
exit 1
fi
if [ -z "$infer_output_dir" ]; then
infer_output_dir="$expt_dir/result"
fi
if [ -z "$infer_mode" ]; then
echo "[Error] Please specify the inference mode, e.g., "batch", "single""
exit 1
fi
if [ "$infer_mode" = "batch" ] && [ -z "$infer_test_list_file" ]; then
echo "[Error] Please specify the test list file used in inference when the inference mode is batch"
exit 1
fi
if [ "$infer_mode" = "single" ] && [ -z "$infer_text" ]; then
echo "[Error] Please specify the text to be synthesized when the inference mode is single"
exit 1
fi
if [ "$infer_mode" = "single" ]; then
echo 'Text: ' ${infer_text}
infer_test_list_file=None
elif [ "$infer_mode" = "batch" ]; then
infer_text=""
infer_text_prompt=""
infer_audio_prompt=""
fi
CUDA_VISIBLE_DEVICES=$gpu accelerate launch "$work_dir"/bins/tts/inference.py \
--config $exp_config \
--log_level debug \
--acoustics_dir $infer_expt_dir \
--output_dir $infer_output_dir \
--mode $infer_mode \
--text "$infer_text" \
--text_prompt "$infer_text_prompt" \
--audio_prompt $infer_audio_prompt\
--test_list_file $infer_test_list_file \
fi
|