Transformers Can Navigate Mazes With Multi-Step Prediction Paper • 2412.05117 • Published 20 days ago • 5
view article Article Releasing the largest multilingual open pretraining dataset By Pclanglais • Nov 13 • 98
view article Article SauerkrautLM's Multi-Phase Spectrum Training: A Technical Deep Dive By DavidGF • Nov 9 • 9
"Give Me BF16 or Give Me Death"? Accuracy-Performance Trade-Offs in LLM Quantization Paper • 2411.02355 • Published Nov 4 • 46
Fast Matrix Multiplications for Lookup Table-Quantized LLMs Paper • 2407.10960 • Published Jul 15 • 11
Strong German fp8 LLM's Collection Strong Large Language Models for the german language in fp8 format • 6 items • Updated Sep 24 • 3
Parameter-Efficient Fine-Tuning with Discrete Fourier Transform Paper • 2405.03003 • Published May 5 • 7
Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients Paper • 2407.08296 • Published Jul 11 • 31
INT4 LLMs for vLLM Collection Accurate INT4 quantized models by Neural Magic, ready for use with vLLM! • 18 items • Updated Sep 26 • 8
FP8 LLMs for vLLM Collection Accurate FP8 quantized models by Neural Magic, ready for use with vLLM! • 44 items • Updated Oct 17 • 60
Research projects on top of vLLM Collection Papers cited in https://blog.vllm.ai/2024/07/25/lfai-perf.html • 6 items • Updated Jul 29 • 12