MXFP4/NVFP4 models
AI & ML interests
Computer Vision, LLMs, Multimodal Models, Model Compression
Recent Activity
HQQ-quantized Qwen models
-
mobiuslabsgmbh/Qwen2.5-7B-Instruct-1M_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 3 • 1 -
mobiuslabsgmbh/Qwen2.5-VL-7B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 591 -
mobiuslabsgmbh/Qwen2.5-VL-3B-Instruct_4bitgs64_hqq_hf
Text Generation • 2B • Updated • 3 • 1 -
mobiuslabsgmbh/Qwen2.5-7B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 3 • 2
-
mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib
Text Generation • Updated • 14 • 55 -
mobiuslabsgmbh/Llama-3.1-8B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 7 • 1 -
mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq
Text Generation • Updated • 12 • 31 -
mobiuslabsgmbh/Llama-3.2-3B-Instruct_4bitgs64_hqq_hf
Text Generation • 2B • Updated • 3 • 1
LLama2 models quantized using https://github.com/mobiusml/hqq
-
mobiuslabsgmbh/Llama-2-7b-hf-4bit_g64-HQQ
Text Generation • Updated • 4 • 2 -
mobiuslabsgmbh/Llama-2-13b-hf-4bit_g64-HQQ
Text Generation • Updated • 18 • 1 -
mobiuslabsgmbh/Llama-2-70b-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 4 • 2 -
mobiuslabsgmbh/Llama-2-70b-chat-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 4 • 3
ViT models quantized using https://github.com/mobiusml/hqq
Quantized models in AO/GemLite format
-
mobiuslabsgmbh/Llama-3.1-8B-Instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 11 • 2 -
mobiuslabsgmbh/Phi-4-mini-instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 1 • 1 -
mobiuslabsgmbh/Qwen2.5-7B-Instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 16 • 2 -
mobiuslabsgmbh/Qwen3-32B_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 3 • 1
Re-distilled DeepSeek R1 models
-
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Llama3-8B-v1.1
Text Generation • 8B • Updated • 13 • • 11 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-7B-v1.1
Text Generation • 8B • Updated • 10 • 15 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.1
Text Generation • 2B • Updated • 66 • 13 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.0
Text Generation • 2B • Updated • 26 • 44
This collection will include language, vision and audio models pre-trained or fine-tuned by Mobius Labs GmbH
4-bit and 2-bit Mixtral models quantized using https://github.com/mobiusml/hqq
-
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-4bit_g64-HQQ
Text Generation • Updated • 4 • 9 -
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 11 • 9 -
mobiuslabsgmbh/Mixtral-8x7B-v0.1-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 6 • 4 -
mobiuslabsgmbh/Mixtral-8x7B-v0.1-hf-4bit_g64-HQQ
Text Generation • Updated • 5 • 1
MXFP4/NVFP4 models
Quantized models in AO/GemLite format
-
mobiuslabsgmbh/Llama-3.1-8B-Instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 11 • 2 -
mobiuslabsgmbh/Phi-4-mini-instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 1 • 1 -
mobiuslabsgmbh/Qwen2.5-7B-Instruct_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 16 • 2 -
mobiuslabsgmbh/Qwen3-32B_gemlite-ao_a16w4_gs_128_pack_32bit
Text Generation • Updated • 3 • 1
HQQ-quantized Qwen models
-
mobiuslabsgmbh/Qwen2.5-7B-Instruct-1M_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 3 • 1 -
mobiuslabsgmbh/Qwen2.5-VL-7B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 591 -
mobiuslabsgmbh/Qwen2.5-VL-3B-Instruct_4bitgs64_hqq_hf
Text Generation • 2B • Updated • 3 • 1 -
mobiuslabsgmbh/Qwen2.5-7B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 3 • 2
Re-distilled DeepSeek R1 models
-
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Llama3-8B-v1.1
Text Generation • 8B • Updated • 13 • • 11 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-7B-v1.1
Text Generation • 8B • Updated • 10 • 15 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.1
Text Generation • 2B • Updated • 66 • 13 -
mobiuslabsgmbh/DeepSeek-R1-ReDistill-Qwen-1.5B-v1.0
Text Generation • 2B • Updated • 26 • 44
-
mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib
Text Generation • Updated • 14 • 55 -
mobiuslabsgmbh/Llama-3.1-8B-Instruct_4bitgs64_hqq_hf
Text Generation • 5B • Updated • 7 • 1 -
mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq
Text Generation • Updated • 12 • 31 -
mobiuslabsgmbh/Llama-3.2-3B-Instruct_4bitgs64_hqq_hf
Text Generation • 2B • Updated • 3 • 1
This collection will include language, vision and audio models pre-trained or fine-tuned by Mobius Labs GmbH
LLama2 models quantized using https://github.com/mobiusml/hqq
-
mobiuslabsgmbh/Llama-2-7b-hf-4bit_g64-HQQ
Text Generation • Updated • 4 • 2 -
mobiuslabsgmbh/Llama-2-13b-hf-4bit_g64-HQQ
Text Generation • Updated • 18 • 1 -
mobiuslabsgmbh/Llama-2-70b-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 4 • 2 -
mobiuslabsgmbh/Llama-2-70b-chat-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 4 • 3
4-bit and 2-bit Mixtral models quantized using https://github.com/mobiusml/hqq
-
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-4bit_g64-HQQ
Text Generation • Updated • 4 • 9 -
mobiuslabsgmbh/Mixtral-8x7B-Instruct-v0.1-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 11 • 9 -
mobiuslabsgmbh/Mixtral-8x7B-v0.1-hf-2bit_g16_s128-HQQ
Text Generation • Updated • 6 • 4 -
mobiuslabsgmbh/Mixtral-8x7B-v0.1-hf-4bit_g64-HQQ
Text Generation • Updated • 5 • 1
ViT models quantized using https://github.com/mobiusml/hqq