You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey @MFajcik, sorry for the delayed response. tokenizers have a default padding side set, but models should all be compatible with different padding sides (unless they explicitly error out). Generally speaking, we use right padding by default (for training, single forward passes, etc), and left padding for generation (necessary for the auto regressive generation and kv cache to work out). Mistral should work fine (we've run it). You may need to update to the latest transformers version. If you have an issue there, please send a full repro. Thanks!
Hi, I was wondering regarding your code here.
composer/composer/datasets/in_context_learning_evaluation.py
Line 655 in a7cad7c
Why do you assume right padding (for InContextLearningMultipleChoiceTaskDataset problem, but also some others)?
Thanks for information.
The text was updated successfully, but these errors were encountered: