同步操作将从 PaddlePaddle/PaddleSpeech 强制同步,此操作会覆盖自 Fork 仓库以来所做的任何修改,且无法恢复!!!
确定后同步将在后台操作,完成时将刷新页面,请耐心等待。
Download LJSpeech-1.1 from it's Official Website and extract it to ~/datasets
. Then the dataset is in the directory ~/datasets/LJSpeech-1.1
.
Assume the path to the dataset is ~/datasets/LJSpeech-1.1
.
Assume the path to the Tacotron2 generated mels is ../tts0/output/test
.
Run the command below to
./run.sh
You can choose a range of stages you want to run, or set stage
equal to stop-stage
to use only one stage, for example, running the following command will only preprocess the dataset.
./run.sh --stage 0 --stop-stage 0
./local/preprocess.sh ${preprocess_path}
./local/train.sh
calls ${BIN_DIR}/train.py
.
CUDA_VISIBLE_DEVICES=${gpus} ./local/train.sh ${preprocess_path} ${train_output_path}
The training script requires 4 command line arguments.
--data
is the path of the training dataset.--output
is the path of the output directory.--ngpu
is the number of gpus to use, if ngpu == 0, use cpu.If you want distributed training, set a larger --ngpu
(e.g. 4). Note that distributed training with cpu is not supported yet.
./local/synthesize.sh
calls ${BIN_DIR}/synthesize.py
, which can synthesize waveform from mels.
CUDA_VISIBLE_DEVICES=${gpus} ./local/synthesize.sh ${input_mel_path} ${train_output_path} ${ckpt_name}
Synthesize waveform.
--input
is a directory containing several mel spectrograms(log magnitude) in .npy
format.--output
directory, containing several .wav
files, each with the same name as the mel spectrogram does.--checkpoint_path
should be the path of the parameter file (.pdparams
) to load. Note that the extention name .pdparmas
is not included here.--ngpu
is the number of gpus to use, if ngpu == 0, use cpu.Pretrained Model with residual channel equals 128 can be downloaded here:
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。