Text Generation
Transformers
PyTorch
Safetensors
English
hf_olmo
custom_code

Number of tokens used to train 1B

#19
by cz-cb - opened

Hi! The model card says the 1B model is trained with 3T tokens, but the paper says it used 2T tokens. Which one is the correct number of training tokens?

Sign up or log in to comment