"loss_masking":true,// enable / disable loss masking against the sequence padding.
"enable_eos_bos_chars":false,// enable/disable beginning of sentence and end of sentence chars.
"stopnet":true,// Train stopnet predicting the end of synthesis.
"separate_stopnet":true,// Train stopnet seperately if 'stopnet==true'. It prevents stopnet loss to influence the rest of the model. It causes a better model, but it trains SLOWER.
"tb_model_param_stats":false,// true, plots param stats per layer on tensorboard. Might be memory consuming, but good for debugging.
"batch_size":32,// Batch size for training. Lower values than 32 might cause hard to learn attention.
"eval_batch_size":32,
"r":5,// Number of frames to predict for step.
"wd":0.000001,// Weight decay weight.
"checkpoint":true,// If true, it saves checkpoints per "save_step"
"save_step":1000,// Number of training steps expected to save traning stats and checkpoints.
"print_step":10,// Number of steps to log traning on console.
"batch_group_size":0,//Number of batches to shuffle after bucketing.
"run_eval":false,
"test_sentences_file":"de_sentences.txt",// set a file to load sentences to be used for testing. If it is null then we use default english sentences.
"test_delay_epochs":5,//Until attention is aligned, testing only wastes computation time.
"data_path":"/home/erogol/Data/m-ai-labs/de_DE/by_book/",// DATASET-RELATED: can overwritten from command argument
],// DATASET-RELATED: metafile for training dataloader.
"meta_file_val":null,// DATASET-RELATED: metafile for evaluation dataloader.
"dataset":"mailabs",// DATASET-RELATED: one of TTS.dataset.preprocessors depending on your target dataset. Use "tts_cache" for pre-computed dataset by extract_features.py
"min_seq_len":15,// DATASET-RELATED: minimum text length to use in training
"max_seq_len":200,// DATASET-RELATED: maximum text length
"output_path":"/media/erogol/data_ssd/Data/models/mozilla_models/",// DATASET-RELATED: output path for all training outputs.
"num_loader_workers":0,// number of training data loader processes. Don't set it too big. 4-8 are good values.
"num_val_loader_workers":4,// number of evaluation data loader processes.
"phoneme_cache_path":"phoneme_cache",// phoneme computation is slow, therefore, it caches results in the given folder.
"use_phonemes":true,// use phonemes instead of raw characters. It is suggested for better pronounciation.
"phoneme_language":"de",// depending on your target language, pick one from https://github.com/bootphon/phonemizer#languages