You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Issue:
We currently depend on vocabularies, like glove embeddings, that are:
Weirdly biased (although when you backprop to the embeddings, their initial bias is not very relevant anymore),
Depend on being consistent with the tokenizer we use.
Don't necessarily have the same words as our actual text.
Proposed solution project:
Use https://github.com/tensorflow/transform to develop text preprocessing pipelines, e.g. to select tokens that occur sufficiently frequently, and create either random or smarter word embeddings for them.
The text was updated successfully, but these errors were encountered:
iislucas
changed the title
Create a preprocessing pipeline using https://github.com/tensorflow/transform
Create a token/embedding creation preprocessing pipeline using https://github.com/tensorflow/transform
Jul 2, 2018
iislucas
changed the title
Create a token/embedding creation preprocessing pipeline using https://github.com/tensorflow/transform
Create a token/embedding creation preprocessing pipeline using tf-transform
Jul 2, 2018
Issue:
We currently depend on vocabularies, like glove embeddings, that are:
Proposed solution project:
Use https://github.com/tensorflow/transform to develop text preprocessing pipelines, e.g. to select tokens that occur sufficiently frequently, and create either random or smarter word embeddings for them.
The text was updated successfully, but these errors were encountered: