|
--- |
|
tags: |
|
- unsloth |
|
- mlx |
|
base_model: unsloth/Qwen3-Coder-30B-A3B-Instruct |
|
library_name: mlx |
|
license: apache-2.0 |
|
license_link: https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# unsloth-Qwen3-Coder-30B-A3B-Instruct-qx4-mlx |
|
|
|
Based on the benchmark results, qx4 would be best suited for: |
|
|
|
Primary Task: BoolQ (Boolean Questions) |
|
|
|
Why BoolQ is the Strength: |
|
|
|
qx4 achieves 0.877 on BoolQ, which is the second-highest score in this dataset |
|
|
|
Only slightly behind q5 (0.883) and qx5 (0.880) |
|
|
|
This represents excellent performance on boolean reasoning tasks |
|
|
|
|
|
Secondary Strengths: |
|
|
|
HellaSwag |
|
|
|
qx4 scores 0.552, which is the highest among all quantized models |
|
|
|
This indicates superior performance on commonsense reasoning and scenario understanding |
|
|
|
Arc_Challenge |
|
|
|
qx4 scores 0.419, which is better than most other quantized models |
|
|
|
Shows strong performance on challenging multiple-choice questions |
|
|
|
|
|
Task Suitability Analysis: |
|
|
|
Best Suited Tasks: |
|
|
|
BoolQ - Strongest performer |
|
|
|
HellaSwag - Highest among quantized models |
|
|
|
Arc_Challenge - Better than most quantizations |
|
|
|
Winogrande - Decent performance (0.567) |
|
|
|
|
|
Other Tasks Where qx4 Performs Well: |
|
|
|
Arc_Easy - 0.531 (solid performance) |
|
|
|
OpenBookQA - 0.426 (adequate for knowledge-based tasks) |
|
|
|
PIQA - 0.723 (good performance) |
|
|
|
|
|
Limitations: |
|
|
|
Weakest in OpenBookQA compared to qm68 (0.426 vs 0.430) |
|
|
|
Below average on Winogrande (0.567) |
|
|
|
Slightly lower than baseline on Arc_Easy |
|
|
|
|
|
Recommendation: |
|
|
|
Use qx4 when Boolean reasoning and commonsense understanding are critical, particularly for applications involving: |
|
|
|
Question answering requiring boolean logic |
|
|
|
Commonsense reasoning scenarios |
|
|
|
Complex multiple-choice question solving |
|
|
|
Tasks where HellaSwag performance is important |
|
|
|
|
|
The model excels at combining logical reasoning (BoolQ) with contextual understanding (HellaSwag), making it ideal for applications that blend precise logical inference with real-world commonsense knowledge. Its performance is particularly strong in scenarios requiring nuanced reasoning about everyday situations and causal relationships. |
|
|
|
Best for: AI assistants, question-answering systems requiring both logical precision and common-sense understanding. |
|
|
|
|
|
This model [unsloth-Qwen3-Coder-30B-A3B-Instruct-qx4-mlx](https://huggingface.co/unsloth-Qwen3-Coder-30B-A3B-Instruct-qx4-mlx) was |
|
converted to MLX format from [unsloth/Qwen3-Coder-30B-A3B-Instruct](https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct) |
|
using mlx-lm version **0.26.3**. |
|
|
|
## Use with mlx |
|
|
|
```bash |
|
pip install mlx-lm |
|
``` |
|
|
|
```python |
|
from mlx_lm import load, generate |
|
|
|
model, tokenizer = load("unsloth-Qwen3-Coder-30B-A3B-Instruct-qx4-mlx") |
|
|
|
prompt = "hello" |
|
|
|
if tokenizer.chat_template is not None: |
|
messages = [{"role": "user", "content": prompt}] |
|
prompt = tokenizer.apply_chat_template( |
|
messages, add_generation_prompt=True |
|
) |
|
|
|
response = generate(model, tokenizer, prompt=prompt, verbose=True) |
|
``` |
|
|