File size: 363 Bytes
57a0c76 b18facf |
1 2 3 4 5 6 7 |
---
license: apache-2.0
base_model: Jinx-org/Jinx-Qwen3-4B
base_model_relation: quantized
---
# Recommended
**Jinx-Qwen3-4B-gguf-q6_k-q-8 (mixed-precision):** selected weights (output, token embeddings, attention/FFN layers in first and last blocks) quantized to **Q8_0**, remaining tensors **Q6_k**, reducing memory footprint while preserving inference fidelity. |