-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hierarchical tokenizers #27
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Welcome to Codecov 🎉Once you merge this PR into your default branch, you're all set! Codecov will compare coverage reports and display results in all future pull requests. Thanks for integrating Codecov - We've got you covered ☂️ |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR introduces a few new things:
1. A new "tokenizer-config" spec
It is now preferred that we instantiate tokenizers with a
tokenizers.toml
file. This was added to support hierarchical tokenizers which might take many BED files to instantiate.Documentation is needed for the
tokenizer.toml
config files2. A new hierarchical tokenizer
This directly addresses discussion over in #25, https://github.com/databio/geniml_dev/issues/85, and https://github.com/databio/geniml_dev/issues/79.
A hierarchical tokenizer can take many universes as input, establishing a priority of tokenization. The primary goal here is to significantly reduce the number of UNK token hits when tokenizing datasets
3. A new MetaTokenizer
Another extension of the tokenizers, the
MetaTokenizer
implements the "meta-tokenization" idea we had. In brief, clusters of highly similar regions (regions that cluster super close in embedding space), are all mapped to a singular "meta token", with the primary goal of drastically reducing vocab sizes for our models to improve training and inference speed and RAM requirements.And of course, python bindings for all of it are implemented. I've also removed a lot of code that was antiquated and unused. While doing so, I spent considerable time revamping the documentation and tests.