You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"When using the standalone GemmaTokenizerFast make sure to pass padding="max_length" and max_length=64 as that’s how the model was trained." Does Siglip2 support longer text input? If the max_length is set to 256 or 512, will text exceeding 64 be truncated?
The text was updated successfully, but these errors were encountered:
SigLIP 2 was trained with text length 64. The big_vision Gemma tokenizer implementation will pad/truncate to 64 if you set length=64. I'm not sure how other implementations behave (it seems you're referencing the HF transformers implementation). It's unclear how model quality will change if you set the length/max_length to a different value (and resize the positional embedding of the text encoder accordingly), since it was trained with 64.
"When using the standalone GemmaTokenizerFast make sure to pass padding="max_length" and max_length=64 as that’s how the model was trained." Does Siglip2 support longer text input? If the max_length is set to 256 or 512, will text exceeding 64 be truncated?
The text was updated successfully, but these errors were encountered: