AI & ML interests

None defined yet.

Recent Activity

depasquale  updated a model about 3 hours ago
mlx-community/Olmo-3.1-32B-Think-bf16
depasquale  published a model about 3 hours ago
mlx-community/Olmo-3.1-32B-Think-bf16
depasquale  updated a model about 3 hours ago
mlx-community/Olmo-3.1-32B-Think-8bit
View all activity

mlx-community 's collections 108

Josiefied and Abliterated Models
Qwen3-Coder-MoE
💻 Significant Performance: among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving ~Claude Sonnet.
Josiefied and Abliterated Qwen3
Abliterated, and further fine-tuned to be the most uncensored models available. Now in MLX
Gemma 3 QAT
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory.
Josiefied and Abliterated Qwen2.5
The best uncensored models
ServiceNow-Apriel
Apriel-1.5-15b-Thinker is a multimodal reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against models 10 time
Qwen2.5
The Qwen 2.5 models are a series of AI models trained on 18 trillion tokens, supporting 29 languages and offering advanced features such as instructio
Llama 3.2
Meta goes small with Llama3.2, both text only 1B and 3B, and the 11B Vision models.
Josiefied and Abliterated Models
ServiceNow-Apriel
Apriel-1.5-15b-Thinker is a multimodal reasoning model in ServiceNow’s Apriel SLM series which achieves competitive performance against models 10 time
Qwen3-Coder-MoE
💻 Significant Performance: among open models on Agentic Coding, Agentic Browser-Use, and other foundational coding tasks, achieving ~Claude Sonnet.
Josiefied and Abliterated Qwen3
Abliterated, and further fine-tuned to be the most uncensored models available. Now in MLX
Gemma 3 QAT
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory.
Josiefied and Abliterated Qwen2.5
The best uncensored models
Qwen2.5
The Qwen 2.5 models are a series of AI models trained on 18 trillion tokens, supporting 29 languages and offering advanced features such as instructio
Llama 3.2
Meta goes small with Llama3.2, both text only 1B and 3B, and the 11B Vision models.