| | from encoder.preprocess import preprocess_librispeech, preprocess_voxceleb1, preprocess_voxceleb2 |
| | from utils.argutils import print_args |
| | from pathlib import Path |
| | import argparse |
| |
|
| |
|
| | if __name__ == "__main__": |
| | class MyFormatter(argparse.ArgumentDefaultsHelpFormatter, argparse.RawDescriptionHelpFormatter): |
| | pass |
| |
|
| | parser = argparse.ArgumentParser( |
| | description="Preprocesses audio files from datasets, encodes them as mel spectrograms and " |
| | "writes them to the disk. This will allow you to train the encoder. The " |
| | "datasets required are at least one of VoxCeleb1, VoxCeleb2 and LibriSpeech. " |
| | "Ideally, you should have all three. You should extract them as they are " |
| | "after having downloaded them and put them in a same directory, e.g.:\n" |
| | "-[datasets_root]\n" |
| | " -LibriSpeech\n" |
| | " -train-other-500\n" |
| | " -VoxCeleb1\n" |
| | " -wav\n" |
| | " -vox1_meta.csv\n" |
| | " -VoxCeleb2\n" |
| | " -dev", |
| | formatter_class=MyFormatter |
| | ) |
| | parser.add_argument("datasets_root", type=Path, help=\ |
| | "Path to the directory containing your LibriSpeech/TTS and VoxCeleb datasets.") |
| | parser.add_argument("-o", "--out_dir", type=Path, default=argparse.SUPPRESS, help=\ |
| | "Path to the output directory that will contain the mel spectrograms. If left out, " |
| | "defaults to <datasets_root>/SV2TTS/encoder/") |
| | parser.add_argument("-d", "--datasets", type=str, |
| | default="librispeech_other,voxceleb1,voxceleb2", help=\ |
| | "Comma-separated list of the name of the datasets you want to preprocess. Only the train " |
| | "set of these datasets will be used. Possible names: librispeech_other, voxceleb1, " |
| | "voxceleb2.") |
| | parser.add_argument("-s", "--skip_existing", action="store_true", help=\ |
| | "Whether to skip existing output files with the same name. Useful if this script was " |
| | "interrupted.") |
| | parser.add_argument("--no_trim", action="store_true", help=\ |
| | "Preprocess audio without trimming silences (not recommended).") |
| | args = parser.parse_args() |
| |
|
| | |
| | if not args.no_trim: |
| | try: |
| | import webrtcvad |
| | except: |
| | raise ModuleNotFoundError("Package 'webrtcvad' not found. This package enables " |
| | "noise removal and is recommended. Please install and try again. If installation fails, " |
| | "use --no_trim to disable this error message.") |
| | del args.no_trim |
| |
|
| | |
| | args.datasets = args.datasets.split(",") |
| | if not hasattr(args, "out_dir"): |
| | args.out_dir = args.datasets_root.joinpath("SV2TTS", "encoder") |
| | assert args.datasets_root.exists() |
| | args.out_dir.mkdir(exist_ok=True, parents=True) |
| |
|
| | |
| | print_args(args, parser) |
| | preprocess_func = { |
| | "librispeech_other": preprocess_librispeech, |
| | "voxceleb1": preprocess_voxceleb1, |
| | "voxceleb2": preprocess_voxceleb2, |
| | } |
| | args = vars(args) |
| | for dataset in args.pop("datasets"): |
| | print("Preprocessing %s" % dataset) |
| | preprocess_func[dataset](**args) |
| |
|