max_new_tokens Issue
#6
by
JungZoona
- opened
First of all, thank you for sharing such a great model.
I just wanted to report that in the VLM Example section of the example code in the model card, setting max_new_tokens does not seem to control the number of output tokens. On the other hand, setting max_length does apply the token limit as expected.
While max_new_tokens works as intended in the LLM Example section, it doesn't seem to function properly in the VLM Example.
Could you please check if max_new_tokens is being applied correctly?