"location_attn":true,// ONLY TACOTRON2 - enable_disable location sensitive attention. It is enabled for TACOTRON by default.
"loss_masking":true,// enable / disable loss masking against the sequence padding.
"enable_eos_bos_chars":false,// enable/disable beginning of sentence and end of sentence chars.
"stopnet":true,// Train stopnet predicting the end of synthesis.
"separate_stopnet":true,// Train stopnet seperately if 'stopnet==true'. It prevents stopnet loss to influence the rest of the model. It causes a better model, but it trains SLOWER.
"tb_model_param_stats":false,// true, plots param stats per layer on tensorboard. Might be memory consuming, but good for debugging.
"batch_size":32,// Batch size for training. Lower values than 32 might cause hard to learn attention.
"eval_batch_size":32,
"r":5,// Number of frames to predict for step.
"wd":0.000001,// Weight decay weight.
"checkpoint":true,// If true, it saves checkpoints per "save_step"
"save_step":1000,// Number of training steps expected to save traning stats and checkpoints.
"print_step":10,// Number of steps to log traning on console.
"batch_group_size":0,//Number of batches to shuffle after bucketing.
"run_eval":false,
"test_sentences_file":"de_sentences.txt",// set a file to load sentences to be used for testing. If it is null then we use default english sentences.
"test_delay_epochs":5,//Until attention is aligned, testing only wastes computation time.
"data_path":"/home/erogol/Data/m-ai-labs/de_DE/by_book/",// DATASET-RELATED: can overwritten from command argument
"meta_file_val":null,// DATASET-RELATED: metafile for evaluation dataloader.
"dataset":"mailabs",// DATASET-RELATED: one of TTS.dataset.preprocessors depending on your target dataset. Use "tts_cache" for pre-computed dataset by extract_features.py
"min_seq_len":15,// DATASET-RELATED: minimum text length to use in training
"max_seq_len":200,// DATASET-RELATED: maximum text length