exceed the model's predefined maximum length (4096)

#2
by MengboZhou - opened

When I use the Qwen2.5-Math-7B model for inference, I get the following information:

This is a friendly reminder - the current text generation call will exceed the model's predefined maximum length (4096). Depending on the model, you may observe exceptions, performance degradation, or nothing at all.

What is the context length and maximum number of output tokens for this model?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment