Update README.md
Browse files
README.md
CHANGED
@@ -42,8 +42,6 @@ Built for **Apple Silicon** with Metal acceleration.
|
|
42 |
pip install mlx-lm transformers
|
43 |
```
|
44 |
|
45 |
-
Once set up, you can proceed to run the model by running the snippet below:
|
46 |
-
|
47 |
```python
|
48 |
# Python API (uses tokenizer bundled with this repo)
|
49 |
from mlx_lm import load, generate
|
@@ -81,10 +79,12 @@ Note: Small, domain-specific eval for quick sanity; not a benchmark suite.
|
|
81 |
|
82 |
## Conversion details (provenance)
|
83 |
|
|
|
84 |
python -m mlx_lm convert \
|
85 |
--hf-path openai/gpt-oss-20b \
|
86 |
--mlx-path gpt-oss-20b-mlx-q6-gs32 \
|
87 |
--q-bits 6 --q-group-size 32 -q
|
|
|
88 |
|
89 |
- Some non-expert tensors (embeddings, norms, router) remain FP16.
|
90 |
|
|
|
42 |
pip install mlx-lm transformers
|
43 |
```
|
44 |
|
|
|
|
|
45 |
```python
|
46 |
# Python API (uses tokenizer bundled with this repo)
|
47 |
from mlx_lm import load, generate
|
|
|
79 |
|
80 |
## Conversion details (provenance)
|
81 |
|
82 |
+
```bash
|
83 |
python -m mlx_lm convert \
|
84 |
--hf-path openai/gpt-oss-20b \
|
85 |
--mlx-path gpt-oss-20b-mlx-q6-gs32 \
|
86 |
--q-bits 6 --q-group-size 32 -q
|
87 |
+
```
|
88 |
|
89 |
- Some non-expert tensors (embeddings, norms, router) remain FP16.
|
90 |
|