max_new_tokens Issue

#6
by JungZoona - opened

First of all, thank you for sharing such a great model.

I just wanted to report that in the VLM Example section of the example code in the model card, setting max_new_tokens does not seem to control the number of output tokens. On the other hand, setting max_length does apply the token limit as expected.

While max_new_tokens works as intended in the LLM Example section, it doesn't seem to function properly in the VLM Example.

Could you please check if max_new_tokens is being applied correctly?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment