Max Output Tokens of Llama3.1-8B
#114
by
MengboZhou
- opened
I am wondering what the maximum number of output tokens is for the LLaMA 3.1-8B model during inference. Also, is there a public document listing the output limits?