mirror of https://github.com/coqui-ai/TTS.git
Fix & update WaveRNN vocoder model (#1749)
* Fixes KeyError bug. Adding logging to dashboard. * Make pep8 compliant * Make style compliant * Still fixing stylepull/1809/head
parent
1a065fa6ed
commit
5094499eba
|
@ -233,6 +233,7 @@ class Wavernn(BaseVocoder):
|
||||||
else:
|
else:
|
||||||
raise RuntimeError("Unknown model mode value - ", self.args.mode)
|
raise RuntimeError("Unknown model mode value - ", self.args.mode)
|
||||||
|
|
||||||
|
self.ap = AudioProcessor(**config.audio.to_dict())
|
||||||
self.aux_dims = self.args.res_out_dims // 4
|
self.aux_dims = self.args.res_out_dims // 4
|
||||||
|
|
||||||
if self.args.use_upsample_net:
|
if self.args.use_upsample_net:
|
||||||
|
@ -571,7 +572,7 @@ class Wavernn(BaseVocoder):
|
||||||
def test(
|
def test(
|
||||||
self, assets: Dict, test_loader: "DataLoader", output: Dict # pylint: disable=unused-argument
|
self, assets: Dict, test_loader: "DataLoader", output: Dict # pylint: disable=unused-argument
|
||||||
) -> Tuple[Dict, Dict]:
|
) -> Tuple[Dict, Dict]:
|
||||||
ap = assets["audio_processor"]
|
ap = self.ap
|
||||||
figures = {}
|
figures = {}
|
||||||
audios = {}
|
audios = {}
|
||||||
samples = test_loader.dataset.load_test_samples(1)
|
samples = test_loader.dataset.load_test_samples(1)
|
||||||
|
@ -587,8 +588,16 @@ class Wavernn(BaseVocoder):
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
audios.update({f"test_{idx}/audio": y_hat})
|
audios.update({f"test_{idx}/audio": y_hat})
|
||||||
|
# audios.update({f"real_{idx}/audio": y_hat})
|
||||||
return figures, audios
|
return figures, audios
|
||||||
|
|
||||||
|
def test_log(
|
||||||
|
self, outputs: Dict, logger: "Logger", assets: Dict, steps: int # pylint: disable=unused-argument
|
||||||
|
) -> Tuple[Dict, np.ndarray]:
|
||||||
|
figures, audios = outputs
|
||||||
|
logger.eval_figures(steps, figures)
|
||||||
|
logger.eval_audios(steps, audios, self.ap.sample_rate)
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def format_batch(batch: Dict) -> Dict:
|
def format_batch(batch: Dict) -> Dict:
|
||||||
waveform = batch[0]
|
waveform = batch[0]
|
||||||
|
@ -605,7 +614,7 @@ class Wavernn(BaseVocoder):
|
||||||
verbose: bool,
|
verbose: bool,
|
||||||
num_gpus: int,
|
num_gpus: int,
|
||||||
):
|
):
|
||||||
ap = assets["audio_processor"]
|
ap = self.ap
|
||||||
dataset = WaveRNNDataset(
|
dataset = WaveRNNDataset(
|
||||||
ap=ap,
|
ap=ap,
|
||||||
items=samples,
|
items=samples,
|
||||||
|
|
Loading…
Reference in New Issue