exceed the model's predefined maximum length (4096)
#2
by
MengboZhou
- opened
When I use the Qwen2.5-Math-7B model for inference, I get the following information:
This is a friendly reminder - the current text generation call will exceed the model's predefined maximum length (4096). Depending on the model, you may observe exceptions, performance degradation, or nothing at all.
What is the context length and maximum number of output tokens for this model?