Max Output Tokens of Llama3.1-8B

#114
by MengboZhou - opened

I am wondering what the maximum number of output tokens is for the LLaMA 3.1-8B model during inference. Also, is there a public document listing the output limits?

Sign up or log in to comment