"grad_clip":1,// upper limit for gradients for clipping.
"epochs":1000,// total number of epochs to train.
"lr":0.0001,// Initial learning rate. If Noam decay is active, maximum learning rate.
"lr_decay":false,// if true, Noam learning rate decaying is applied through training.
"warmup_steps":4000,// Noam decay steps to increase the learning rate from 0 to "lr"
"windowing":false,// Enables attention windowing. Used only in eval mode.
"memory_size":5,// ONLY TACOTRON - memory queue size used to queue network predictions to feed autoregressive connection. Useful if r < 5.
"attention_norm":"sigmoid",// softmax or sigmoid. Suggested to use softmax for Tacotron2 and sigmoid for Tacotron.
"prenet_type":"original",// ONLY TACOTRON2 - "original" or "bn".
"prenet_dropout":true,// ONLY TACOTRON2 - enable/disable dropout at prenet.
"use_forward_attn":true,// ONLY TACOTRON2 - if it uses forward attention. In general, it aligns faster.
"forward_attn_mask":false,
"transition_agent":false,// ONLY TACOTRON2 - enable/disable transition agent of forward attention.
"location_attn":false,// ONLY TACOTRON2 - enable_disable location sensitive attention. It is enabled for TACOTRON by default.
"loss_masking":true,// enable / disable loss masking against the sequence padding.
"enable_eos_bos_chars":false,// enable/disable beginning of sentence and end of sentence chars.
"stopnet":true,// Train stopnet predicting the end of synthesis.
"separate_stopnet":true,// Train stopnet seperately if 'stopnet==true'. It prevents stopnet loss to influence the rest of the model. It causes a better model, but it trains SLOWER.
"tb_model_param_stats":false,// true, plots param stats per layer on tensorboard. Might be memory consuming, but good for debugging.
"test_delay_epochs":5,//Until attention is aligned, testing only wastes computation time.
"test_sentences_file":null,// set a file to load sentences to be used for testing. If it is null then we use default english sentences.
"data_path":"/home/erogol/Data/Libri-TTS/train-clean-360/",// DATASET-RELATED: can overwritten from command argument
"meta_file_train":null,// DATASET-RELATED: metafile for training dataloader.
"meta_file_val":null,// DATASET-RELATED: metafile for evaluation dataloader.
"dataset":"libri_tts",// DATASET-RELATED: one of TTS.dataset.preprocessors depending on your target dataset. Use "tts_cache" for pre-computed dataset by extract_features.py
"min_seq_len":6,// DATASET-RELATED: minimum text length to use in training
"max_seq_len":150,// DATASET-RELATED: maximum text length
"output_path":"/media/erogol/data_ssd/Models/libri_tts/",// DATASET-RELATED: output path for all training outputs.