Pretraining Colab Notebook: https://colab.research.google.com/drive/1XYdqwQoAqZPUa524reehwcOe-zcy93Zy?usp=sharing
Finetuning Colab Notebook: https://colab.research.google.com/drive/1dmb7WYaPWnsYVHtO37DsIQVKCNjAM2Q3?usp=sharing
(Use to test the finetuned models)
Evaluation Colab Notebook: https://colab.research.google.com/drive/1IxyxmCI_fCDB1wRaOisNZOicUkUONdgU?usp=sharing
Pretrained Models: https://drive.google.com/drive/folders/17r6JXdVbzpj9qE8ZPSu_XZQEl86N6_g_?usp=sharing
Finetuned Models: https://drive.google.com/drive/folders/1ECsp-OmDDUrXAKPyo5Eqx4F2T6XDffTO?usp=sharing
Results: https://drive.google.com/drive/folders/12fXNKFBKsxZmmrtORfhBlR2AHKMftNoH?usp=sharing
- Many-Mask Numeration
- This will randomly mask one or more of the words in the numeration.
- Three thousand four hundred fifty-five -> <mask> thousand <mask> <mask> fifty-five
- Many-Mask Number
- This will random mask one or more digits in the number
- 56483 -> 5<mask><mask>83 or 5<mask>4<mask>3
- Mixed-Mask
- This will randomly mask one or more digits in the number then mask the complement in the numeration
- Three thousand four hundred fifty-five is the number 3455 -> <mask> <mask> four hundred fifty-<mask> is the number 3<mask>5<mask>
Bulding off of this paper: https://aclanthology.org/2021.emnlp-main.563.pdf we will try finetuning using an e based representations. This takes the number and labels the powers of tens. For example 8721 becomes 8 e 3 7 e 2 2 e 1 1 e 0. The authors of the paper above find that this representations performs well on extrapolation tasks, but have not tested it with this numeration task.
The finetuning domain is from 0-9999. The out of domain values are from 0-999999. We have pretrained 2 sets of models, one set up to the in domain range, and the other up to the out of domain range. We are interested in seeing if the out of domain results--predicting the number from numeration--performs better based on the pretraining techniques and ranges.
In Domain | 0-9999 | 0-999999 |
---|---|---|
T5-Base (Control) | 89.656% | -- |
Many-Mask Numeration | 90.557% | 82.249% |
Many-Mask Number | 99.566% | 99.433% |
Mixed-Mask | 99.666% | 99.833% |
These are the results of testing a pre-trained and fine tuned model on an out of domain range.
Out-Of-Domain | 0-9999 | 0-999999 |
---|---|---|
T5-Base (Control) | 12.575% | -- |
Many-Mask Numeration | 2.311% | 2.888% |
Many-Mask Number | 2.230% | 4.029% |
Mixed-Mask | 2.409% | 4.117% |
In Domain | 0-9999 | 0-999999 |
---|---|---|
T5-Base (Control) | 88.055 | -- |
Many-Mask Numeration | 82.449% | 61.862% |
Many-Mask Number | 99.933% | 98.365% |
Mixed-Mask | 99.766% | 99.933% |
These are the results of testing a pre-trained and fine tuned model on an out of domain range.
Out-Of-Domain | 0-9999 | 0-999999 |
---|---|---|
T5-Base (Control) | 0% | -- |
Many-Mask Numeration | 0% | 0% |
Many-Mask Number | 0% | 0% |
Mixed-Mask | 0% | 0% |