In our paper we propose Daft-Exprt, a multi-speaker acoustic model advancing the state-of-the-art for cross-speaker prosody transfer on any text. This is one of the most challenging, and rarely directly addressed, task in speech synthesis, especially for highly expressive data. Daft-Exprt uses FiLM conditioning layers to strategically inject different prosodic information in all parts of the architecture. The model explicitly encodes traditional low-level prosody features such as pitch, loudness and duration, but also higher level prosodic information that helps generating convincing voices in highly expressive styles. Speaker identity and prosodic information are disentangled through an adversarial training strategy that enables accurate prosody transfer across speakers. Experimental results show that Daft-Exprt significantly outperforms strong baselines on inter-text cross-speaker prosody transfer tasks, while yielding naturalness comparable to state-of-the-art expressive models. Moreover, results indicate that the model discards speaker identity information from the prosody representation, and consistently generate speech with the desired voice. Visit our demo page for audio samples related to the paper experiments.
Full disclosure: The model provided in this repository is not the same as in the paper evaluation. The model of the paper was trained with proprietary data which prevents us to release it publicly.
We pre-train Daft-Exprt on a combination of LJ speech dataset and the emotional speech dataset (ESD) from Zhou et al.
Visit the releases of this repository to download the pre-trained model and to listen to prosody transfer examples using this same model.
Requirements:
- Ubuntu >= 20.04
- Python >= 3.8
- NVIDIA Driver >= 450.80.02
- CUDA Toolkit >= 11.1
- CuDNN >= v8.0.5
We recommend using conda for python environment management, for example download and install Miniconda.
Create your python environment and install dependencies using the Makefile:
conda create -n daft_exprt python=3.8 -y
conda activate daft_exprt
cd environment
make
All Linux/Conda/Pip dependencies will be installed by the Makefile, and the repository will be installed as a pip package in editable mode.
Requirements:
- NVIDIA Docker
- NVIDIA Driver >= 450.80.02
Build the Docker image using the associated Dockerfile:
docker build -f environment/Dockerfile -t daft_exprt .
This quick start guide will illustrate how to use the different scripts of this repository to:
- Format datasets
- Pre-process these datasets
- Train Daft-Exprt on the pre-processed data
- Generate a dataset for vocoder fine-tuning
- Use Daft-Exprt for TTS synthesis
All scripts are located in scripts directory.
Daft-Exprt source code is located in daft_exprt directory.
Config parameters used in the scripts are all instanciated in hparams.py.
As a quick start example, we consider using the 22kHz LJ speech dataset and the 16kHz emotional speech dataset (ESD) from Zhou et al.
This combines a total of 11 speakers. All speaker datasets must be in the same root directory. For example:
/data_dir
LJ_Speech
ESD
spk_1
...
spk_N
In this example, we use the docker image built in the previous section:
docker run -it --gpus all -v /path/to/data_dir:/workdir/data_dir -v path/to/repo_dir:/workdir/repo_dir IMAGE_ID
The source code expects the specific tree structure for each speaker data set:
/speaker_dir
metadata.csv
/wavs
wav_file_name_1.wav
...
wav_file_name_N.wav
metadata.csv must be formatted as follows:
wav_file_name_1|text_1
...
wav_file_name_N|text_N
Given each dataset has its own nomenclature, this project does not provide a ready-made universal script.
However, the script format_dataset.py already proposes the code to format LJ and ESD:
python format_dataset.py \
--data_set_dir /workdir/data_dir/LJ_Speech \
LJ
python format_dataset.py \
--data_set_dir /workdir/data_dir/ESD \
ESD \
--language english
In this section, the code will:
- Align data using MFA
- Extract features for training
- Create train and validation sets
- Extract features stats on the train set for speaker standardization
To pre-process all available formatted data (i.e. LJ and ESD in this example):
python training.py \
--experiment_name EXPERIMENT_NAME \
--data_set_dir /workdir/data_dir \
pre_process
This will pre-process data using the default hyper-parameters that are set for 22kHz audios.
All outputs related to the experiment will be stored in /workdir/repo_dir/trainings/EXPERIMENT_NAME
.
You can also target specific speakers for data pre-processing. For example, to consider only ESD speakers:
python training.py \
--experiment_name EXPERIMENT_NAME \
--speakers ESD/spk_1 ... ESD/spk_N \
--data_set_dir /workdir/data_dir \
pre_process
The pre-process function takes several arguments:
--features_dir
: absolute path where pre-processed data will be stored. Default to/workdir/repo_dir/datasets
--proportion_validation
: Proportion of examples that will be in the validation set. Default to0.1
% per speaker.--nb_jobs
: number of cores to use for python multi-processing. If set tomax
, all CPU cores are used. Default to6
.
Note that if it is the first time that you pre-process the data, this step will take several hours.
You can decrease computing time by increasing the --nb_jobs
parameter.
Once pre-processing is finished, launch training. To train on all pre-processed data:
python training.py \
--experiment_name EXPERIMENT_NAME \
--data_set_dir /workdir/data_dir \
train
Or if you targeted specific speakers during pre-processing (e.g. ESD speakers):
python training.py \
--experiment_name EXPERIMENT_NAME \
--speakers ESD/spk_1 ... ESD/spk_N \
--data_set_dir /workdir/data_dir \
train
All outputs related to the experiment will be stored in /workdir/repo_dir/trainings/EXPERIMENT_NAME
.
The train function takes several arguments:
--checkpoint
: absolute path of a Daft-Exprt checkpoint. Default to""
--no_multiprocessing_distributed
: disable PyTorch multi-processing distributed training. Default toFalse
--world_size
: number of nodes for distributed training. Default to1
.--rank
: node rank for distributed training. Default to0
.--master
: url used to set up distributed training. Default totcp://localhost:54321
.
These default values will launch a new training starting at iteration 0, using all available GPUs on the machine.
The code supposes that only 1 GPU is available on the machine.
Default batch size and gradient accumulation hyper-parameters are set to values to reproduce the batch size of 48 from the paper.
The code also supports tensorboard logging. To display logging outputs:
tensorboard --logdir_spec=EXPERIMENT_NAME:/workdir/repo_dir/trainings/EXPERIMENT_NAME/logs
Once training is finished, you can create a dataset for vocoder fine-tuning:
python training.py \
--experiment_name EXPERIMENT_NAME \
--data_set_dir /workdir/data_dir \
fine_tune \
--checkpoint CHECKPOINT_PATH
Or if you targeted specific speakers during pre-processing and training (e.g. ESD speakers):
python training.py \
--experiment_name EXPERIMENT_NAME \
--speakers ESD/spk_1 ... ESD/spk_N \
--data_set_dir /workdir/data_dir \
fine_tune \
--checkpoint CHECKPOINT_PATH
Fine-tuning dataset will be stored in /workdir/repo_dir/trainings/EXPERIMENT_NAME/fine_tuning_dataset
.
For an example on how to use Daft-Exprt for TTS synthesis, run the script synthesize.py.
python synthesize.py \
--output_dir OUTPUT_DIR \
--checkpoint CHECKPOINT
Default sentences and reference utterances are used in the script.
The script also offers the possibility to:
--batch_size
: process batch of sentences in parallel--real_time_factor
: estimate Daft-Exprt real time factor performance given the chosen batch size--control
: perform local prosody control
@article{Zaidi2021,
abstract = {},
journal = {arXiv},
arxivId = {2108.02271},
author = {Za{\"{i}}di, Julian and Seut{\'{e}}, Hugo and van Niekerk, Benjamin and Carbonneau, Marc-Andr{\'{e}}},
eprint = {2108.02271},
title = {{Daft-Exprt: Robust Prosody Transfer Across Speakers for Expressive Speech Synthesis}},
url = {https://arxiv.org/pdf/2108.02271.pdf},
year = {2021}
}
Any contribution to this repository is more than welcome!
If you have any feedback, please send it to julian.zaidi@ubisoft.com.
© [2021] Ubisoft Entertainment. All Rights Reserved