You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We are trying to reproduce the training with more dialects:
As a first step, we plan to change the voice of the wave files based on the original corpus set of waves. (I'm running Coqui to convert the voice to another style of voice)
Later on, we plan to use our own set of wave files with the new dialect
The issues I found so far in the first step:
There is no documentation about running the scripts/extract_f0.py part to create the data/pitch_dict.pt file; it would be good to add this to the docs. I learned it by reading some code.
After this part, when I first run the train.py file with my custom configuration, I get an error that the ArabDataset class is not being called correctly; it is passing a cache parameter that doesn't exist in the actual implementation
I removed the parameter to try to brute force the results, but it led to another error:
Hi,
I have wanted to add some updates for a while. If you haven't already, you can take a look at the exp branch and the training.md file there. I have added some support for training without the phonemizer there and started to write some instructions.
The train.py in the root folder doesn't work (I should remove it), the working training scripts are in the /scripts folder. I haven't had much time to update the master branch and it could take some days, but I will mostly add the changes from the exp branch. If anything is unclear about the instructions, just let me know, and I will see if I can add something there.
Great work so far!
We are trying to reproduce the training with more dialects:
The issues I found so far in the first step:
scripts/extract_f0.py
part to create thedata/pitch_dict.pt
file; it would be good to add this to the docs. I learned it by reading some code.train.py
file with my custom configuration, I get an error that theArabDataset
class is not being called correctly; it is passing acache
parameter that doesn't exist in the actual implementationAm I in the right path? Can you please double-check if the tools in the
master
branch are the last version used to train the data?Also, I would like to know if I need to use Python 3.9 for this process; the script's dependencies did not work with newer versions.
The text was updated successfully, but these errors were encountered: