You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is there a way to identify special token positions so that we can strip them after the model calls and then only compare logits of non-special ones here:
What effect does skipping special (label) tokens have on the logits?
It prevents the logits of the special tokens from being included in the computation since they might have different values when next to different tokens.
Seems like a better solution would be to have a separate encode function that never adds special tokens?
Location: https://github.com/bigscience-workshop/lm-evaluation-harness/blob/master/lm_eval/models/huggingface.py#L460
@jon-tow I'm not sure if special tokens should be included as part of the target sequence when doing the LL computation.
The text was updated successfully, but these errors were encountered: