No More Adam: Learning Rate Scaling at Initialization is All You Need
Paper
β’
2412.11768
β’
Published
β’
41
datatrove
for all things web-scale data preparation: https://github.com/huggingface/datatrovenanotron
for lightweight 4D parallelism LLM training: https://github.com/huggingface/nanotronlighteval
for in-training fast parallel LLM evaluations: https://github.com/huggingface/lighteval