mirror of https://github.com/coqui-ai/TTS.git
Remove unused kwarg and added num_beams=1 as default
parent
a1c441f205
commit
9e92adc5ac
|
@ -562,7 +562,7 @@ class GPT(nn.Module):
|
|||
|
||||
def inference(self, cond_latents, text_inputs, **hf_generate_kwargs):
|
||||
self.compute_embeddings(cond_latents, text_inputs)
|
||||
return self.generate(cond_latents, text_inputs, input_tokens=None, **hf_generate_kwargs)
|
||||
return self.generate(cond_latents, text_inputs, **hf_generate_kwargs)
|
||||
|
||||
def compute_embeddings(
|
||||
self,
|
||||
|
|
|
@ -642,6 +642,7 @@ class Xtts(BaseTTS):
|
|||
diffusion_temperature=1.0,
|
||||
decoder_sampler="ddim",
|
||||
decoder="hifigan",
|
||||
num_beams=1,
|
||||
**hf_generate_kwargs,
|
||||
):
|
||||
text = text.strip().lower()
|
||||
|
@ -673,6 +674,7 @@ class Xtts(BaseTTS):
|
|||
top_k=top_k,
|
||||
temperature=temperature,
|
||||
num_return_sequences=self.gpt_batch_size,
|
||||
num_beams=num_beams,
|
||||
length_penalty=length_penalty,
|
||||
repetition_penalty=repetition_penalty,
|
||||
output_attentions=False,
|
||||
|
|
Loading…
Reference in New Issue