mirror of https://github.com/coqui-ai/TTS.git
Move `main_classes` to a separate folder
parent
5723eb4738
commit
3584a04552
|
@ -0,0 +1,25 @@
|
|||
# AudioProcessor
|
||||
|
||||
`TTS.utils.audio.AudioProcessor` is the core class for all the audio processing routines. It provides an API for
|
||||
|
||||
- Feature extraction.
|
||||
- Sound normalization.
|
||||
- Reading and writing audio files.
|
||||
- Sampling audio signals.
|
||||
- Normalizing and denormalizing audio signals.
|
||||
- Griffin-Lim vocoder.
|
||||
|
||||
The `AudioProcessor` needs to be initialized with `TTS.config.shared_configs.BaseAudioConfig`. Any model config
|
||||
also must inherit or initiate `BaseAudioConfig`.
|
||||
|
||||
## AudioProcessor
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.utils.audio.AudioProcessor
|
||||
:members:
|
||||
```
|
||||
|
||||
## BaseAudioConfig
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.config.shared_configs.BaseAudioConfig
|
||||
:members:
|
||||
```
|
|
@ -0,0 +1,25 @@
|
|||
# Datasets
|
||||
|
||||
## TTS Dataset
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.tts.datasets.TTSDataset
|
||||
:members:
|
||||
```
|
||||
|
||||
## Vocoder Dataset
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.vocoder.datasets.gan_dataset.GANDataset
|
||||
:members:
|
||||
```
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.vocoder.datasets.wavegrad_dataset.WaveGradDataset
|
||||
:members:
|
||||
```
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.vocoder.datasets.wavernn_dataset.WaveRNNDataset
|
||||
:members:
|
||||
```
|
|
@ -0,0 +1,12 @@
|
|||
# GAN API
|
||||
|
||||
The {class}`TTS.vocoder.models.gan.GAN` provides an easy way to implementing new GAN based models. You just need
|
||||
to define the model architectures for the generator and the discriminator networks and give them to the `GAN` class
|
||||
to do its ✨️.
|
||||
|
||||
|
||||
## GAN
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.vocoder.models.gan.GAN
|
||||
:members:
|
||||
```
|
|
@ -0,0 +1,24 @@
|
|||
# Model API
|
||||
Model API provides you a set of functions that easily make your model compatible with the `Trainer`,
|
||||
`Synthesizer` and `ModelZoo`.
|
||||
|
||||
## Base TTS Model
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.model.BaseModel
|
||||
:members:
|
||||
```
|
||||
|
||||
## Base `tts` Model
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.tts.models.base_tts.BaseTTS
|
||||
:members:
|
||||
```
|
||||
|
||||
## Base `vocoder` Model
|
||||
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.tts.models.base_vocoder.BaseVocoder`
|
||||
:members:
|
||||
```
|
|
@ -0,0 +1,17 @@
|
|||
# Trainer API
|
||||
|
||||
The {class}`TTS.trainer.Trainer` provides a lightweight, extensible, and feature-complete training run-time. We optimized it for 🐸 but
|
||||
can also be used for any DL training in different domains. It supports distributed multi-gpu, mixed-precision (apex or torch.amp) training.
|
||||
|
||||
|
||||
## Trainer
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.trainer.Trainer
|
||||
:members:
|
||||
```
|
||||
|
||||
## TrainingArgs
|
||||
```{eval-rst}
|
||||
.. autoclass:: TTS.trainer.TrainingArgs
|
||||
:members:
|
||||
```
|
Loading…
Reference in New Issue