--- license: apache-2.0 pipeline_tag: text-generation library_name: transformers tags: - vllm - mlx - quantized base_model: openai/gpt-oss-120b --- # gpt-oss-120B (6‑bit quantized via MLX‑LM) **A 6‑bit quantized version of `openai/gpt-oss-120b` created using MLX‑LM.** This version significantly reduces inference memory requirements (~90 GB RAM), while retaining most of the model’s original capabilities. ⸻ 🛠️ Quantization Process The model was created using the following steps: * pip uninstall mlx-lm * pip install git+https://github.com/ml-explore/mlx-lm.git@main * mlx_lm.convert \ --hf-path openai/gpt-oss-120b \ --quantize \ --q-bits 6 \ --output-dir gpt-oss-120b-MLX-6bit These commands use the latest MLX‑LM converter to apply a consistent 6‑bit quantization across model weights.