Tfree-HAT-7b-pretrained Collection Tokenizer free models based on Hierarchical Autoregressive Transformer (https://arxiv.org/abs/2501.10322) trained from scratch. • 2 items • Updated 26 days ago • 7
view article Article From Zero to GPU: A Guide to Building and Scaling Production-Ready CUDA Kernels By drbh and 1 other • 10 days ago • 42
view article Article Say hello to `hf`: a faster, friendlier Hugging Face CLI ✨ By Wauplin and 2 others • Jul 25 • 80
view article Article How Much Power does a SOTA Open Video Model Use? ⚡🎥 By jdelavande and 2 others • Jul 2 • 15
view article Article Bringing Fusion Down to Earth: ML for Stellarator Optimization By cgeorgiaw • Jul 2 • 73
view article Article Groq on Hugging Face Inference Providers 🔥 By sbrandeis and 4 others • Jun 16 • 42
Autoregressive Adversarial Post-Training for Real-Time Interactive Video Generation Paper • 2506.09350 • Published Jun 11 • 48
STARFlow: Scaling Latent Normalizing Flows for High-resolution Image Synthesis Paper • 2506.06276 • Published Jun 6 • 22
view article Article No GPU left behind: Unlocking Efficiency with Co-located vLLM in TRL By toslali-ibm and 5 others • Jun 3 • 84
Model Already Knows the Best Noise: Bayesian Active Noise Selection via Attention in Video Diffusion Model Paper • 2505.17561 • Published May 23 • 31
view article Article Gotchas in Tokenizer Behavior Every Developer Should Know By qgallouedec • Apr 18 • 41