You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The default training script uses LibriTTS dataset instead of LibriSpeech.
The model implementation uses time-channel separable 1D convolutional modules without groups and shuffling. For details refer Fig. 2 from the article.
Instead of NovoGrad optimizer I use PyTorch's Adam optimizer with weight decay.
Instead of Cutout for augmentation I use a custom function based on PyTorch's masking functions. You can find more details about augmentation here.
Authors of the article utilise cosine annealing learning rate policy with learning rate warmup. I use PyTorch's implementation of one cycle learning rate policy. This policy is well described in this article.
The text was updated successfully, but these errors were encountered:
There are a number of differences compared to the source:
Differences:
The text was updated successfully, but these errors were encountered: