view post Post 187 the training for SnowflakeCore-G1-1B and 7B would be retaken because now I implemented DeepSpeed and management to use two gpus. See translation
view post Post 250 The development of SnowflakeCore-G1-7B-MoE it getting delay. In the mean time I am working on SnowflakeCore-G1-1B-MoE witch would be a pre-train chatbot. See translation
SnowflakeCore G1 Pre-Train The base models of G1. All the Snowflake models are fully pre-train, not fine-tune of a pre-existing model. FlameF0X/SnowflakeCore-G1-Tiny2 Text Generation • Updated 9 days ago • 117 FlameF0X/SnowflakeCore-G1-Tiny Text Generation • Updated 9 days ago • 151
SnowflakeCore G1 Fine-Tune Fine tunes of G1 FlameF0X/SnowflakeCore-G1-Tiny-Instruct Text Generation • 0.4B • Updated 25 days ago • 43 FlameF0X/SnowflakeCore-G1-Tiny-Chat Text Generation • Updated 7 days ago • 31
SnowflakeCore G1 Pre-Train The base models of G1. All the Snowflake models are fully pre-train, not fine-tune of a pre-existing model. FlameF0X/SnowflakeCore-G1-Tiny2 Text Generation • Updated 9 days ago • 117 FlameF0X/SnowflakeCore-G1-Tiny Text Generation • Updated 9 days ago • 151
SnowflakeCore G1 Fine-Tune Fine tunes of G1 FlameF0X/SnowflakeCore-G1-Tiny-Instruct Text Generation • 0.4B • Updated 25 days ago • 43 FlameF0X/SnowflakeCore-G1-Tiny-Chat Text Generation • Updated 7 days ago • 31
pinned Running SnowflakeCore G1 Benchmark 📊 Explore FlameF0X/SnowflakeCore-G1-Tiny model benchmark results