"Token indices sequence length is longer than the specified maximum sequence length" when using HHEM-2.1-open
#13
by
ytangccc
- opened
I see the message "Token indices sequence length is longer than the specified maximum sequence length for this model (624 > 512). Running this sequence through the model will result in indexing errors" when using HHEM-2.1-open. It still ran through but I'm just wondering since HHEM-2.1 should have an unlimited context length?
Don't worry about it. This is a notification inherited from the foundation, T5-base.
Is there a restriction on using a different tokenizer?
You have to use the same tokenizer that Google T5 uses. Otherwise, tokens will be mapped to different integer indexes and then mapped to wrong token embeddings. This is a limitation to any Transformer-based models or any model that relies on an embedding layer.
forrest-vectara
changed discussion status to
closed