-
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Keyterm data always gets added - and then we always train #476
Comments
@johnml1135 It seems to me like "only train on key terms, then generate a draft for my first book" is a valid but very unusual use-case. It would have to be a new project in an NLLB language. |
We could filter the key terms by book/chapter. Each key term has a list of verses that they occur in. |
@ddaspit, so, filter on any data trained on or pretranslated? That would leave us with the same issue - namely that if you just want to translate from English to Spanish using NLLB200, that is now prevented. If we want to implement that filter, I would consider that a separate enhancement. |
You are correct. It would still train the model. This issue made me realize that we should filter the key terms. We already have the |
If it is, we should test it out (at least manually) and then document it. |
|
Actually, the Serval changes need to be merged in first before this is completed. |
We are waiting on #508. |
Should we add a separate flag for "only pretranslate"? Or should we automagically work if there is no matching corpora, we don't include the keyterms?
The text was updated successfully, but these errors were encountered: