Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Things that are different compared to the article #3

Open
ivankunyankin opened this issue Jul 5, 2021 · 1 comment
Open

Things that are different compared to the article #3

ivankunyankin opened this issue Jul 5, 2021 · 1 comment
Labels
documentation Improvements or additions to documentation

Comments

@ivankunyankin
Copy link
Owner

ivankunyankin commented Jul 5, 2021

There are a number of differences compared to the source:

Differences:

  1. The default training script uses LibriTTS dataset instead of LibriSpeech.
  2. The model implementation uses time-channel separable 1D convolutional modules without groups and shuffling. For details refer Fig. 2 from the article.
  3. Instead of NovoGrad optimizer I use PyTorch's Adam optimizer with weight decay.
  4. Instead of Cutout for augmentation I use a custom function based on PyTorch's masking functions. You can find more details about augmentation here.
  5. Authors of the article utilise cosine annealing learning rate policy with learning rate warmup. I use PyTorch's implementation of one cycle learning rate policy. This policy is well described in this article.
@ivankunyankin ivankunyankin added the documentation Improvements or additions to documentation label Jul 6, 2021
@raotnameh
Copy link

What is the WER are you able to achieve? Also, can you share the epoch vs WER and CER and LOSS? @ivankunyankin

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

No branches or pull requests

2 participants