view article Article Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA By ybelkada and 4 others • May 24, 2023 • 162
Under 10b iq4_nl gguf Collection Under 10B GGUFs Non-linear Quantized to try on 4GiB VRAM. Leaderboard https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard • 46 items • Updated Jul 10 • 3
FalconMamba 7B Collection This collection features the FalconMamba 7B base model, the instruction-tuned version, their 4-bit and GGUF variants, and the demo. • 15 items • Updated 30 days ago • 34