|
--- |
|
license: mit |
|
datasets: |
|
- openai/gsm8k |
|
language: |
|
- en |
|
base_model: |
|
- Qwen/Qwen2.5-3B-Instruct |
|
library_name: transformers |
|
tags: |
|
- fine-tuned |
|
- qwen |
|
- deepseek |
|
- gsm8k |
|
- reasoning |
|
--- |
|
|
|
# Qwen 2.5-3B-Instruct Fine-Tuned on OpenAI GSM8K with DeepSeek Augmentation |
|
|
|
## Model Overview |
|
|
|
This model is a fine-tuned version of **Qwen/Qwen2.5-3B-Instruct**, optimized for mathematical reasoning tasks using the **OpenAI GSM8K** dataset. The fine-tuning process enhances the model's ability to generate step-by-step explanations for grade school math problems, incorporating **reasoning augmentation** through DeepSeek. The model improves upon GSM8K’s standard answers by integrating additional contextual reasoning derived from DeepSeek’s small model. |
|
|
|
### Key Features: |
|
- **Base Model**: Qwen 2.5-3B-Instruct |
|
- **Fine-Tuned On**: OpenAI GSM8K dataset |
|
- **Enhancement**: Answer augmentation with reasoning insights from **DeepSeek-V3-Small** |
|
- **Improved Reasoning**: Model not only provides correct answers but also **augments** explanations with logical steps inspired by DeepSeek’s generative capabilities. |
|
|
|
## Dataset & Training Details |
|
|
|
- **Dataset**: OpenAI’s GSM8K (Grade School Math 8K), a collection of high-quality math problems designed to test problem-solving skills. |
|
- **Enhancement**: After fine-tuning on GSM8K, additional reasoning layers were introduced using **DeepSeek-V3-Small**, leading to richer, more interpretable answers. |
|
- **Training Objective**: Improve step-by-step mathematical reasoning and **enhance logical deductions** in model-generated responses. |
|
|
|
I have adopted some code from Unsloth and here's an updated [notebook](https://colab.research.google.com/drive/1HV0YkyiTD55j1xLRBHwJ_q3ex82W5EXr?usp=sharing) on Colab. Please feel free to copy it and run it yourself. |
|
|
|
You will need: |
|
- Huggingface token |
|
- Together.AI API Key |
|
- Unsloth package |
|
|
|
## How to Use Model via Terminal (Mac) |
|
|
|
**Goal** Run Qwen-2.5-3B Instruct on Your Mac Using `llama.cpp` |
|
|
|
Yes! You can run **Qwen-2.5-3B Instruct** on your Mac using `llama.cpp`. Here’s a step-by-step guide assuming you are starting from a clean macOS installation with only `pyenv` installed. |
|
|
|
### **Step 1: Install Homebrew (if not installed)** |
|
Homebrew is required to install `llama.cpp`. |
|
|
|
1. Open **Terminal** (`Cmd + Space`, type `Terminal`, and press **Enter**). |
|
2. Run: |
|
```sh |
|
/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)" |
|
``` |
|
3. After installation, add Homebrew to your PATH: |
|
```sh |
|
echo 'eval "$(/opt/homebrew/bin/brew shellenv)"' >> ~/.zprofile |
|
eval "$(/opt/homebrew/bin/brew shellenv)" |
|
``` |
|
|
|
--- |
|
|
|
### **Step 2: Install `llama.cpp` via Homebrew** |
|
Run: |
|
```sh |
|
brew install llama.cpp |
|
``` |
|
|
|
Once installed, you should be able to use `llama-cli`. |
|
|
|
--- |
|
|
|
### **Step 3: Run Qwen-2.5-3B Instruct with `llama-cli`** |
|
To run the model, execute: |
|
```sh |
|
llama-cli -hf eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small:Q8_0 |
|
``` |
|
|
|
--- |
|
|
|
### **Step 4: Additional Configurations (If Needed)** |
|
If you encounter issues or need finer control, you may want to: |
|
|
|
#### **A. Verify Installation** |
|
Check if `llama-cli` is installed: |
|
```sh |
|
llama-cli --version |
|
``` |
|
If you see a version output, it’s installed correctly. |
|
|
|
#### **B. Run with Explicit Model Path** |
|
If the default Hugging Face loader doesn't work, you can manually download the model: |
|
1. **Create a models directory:** |
|
```sh |
|
mkdir -p ~/llama_models && cd ~/llama_models |
|
``` |
|
2. **Download the GGUF model file** from [Hugging Face](https://huggingface.co/eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small): |
|
```sh |
|
wget https://huggingface.co/eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small/resolve/main/Q8_0.gguf |
|
``` |
|
3. **Run the model manually**: |
|
```sh |
|
llama-cli -m ~/llama_models/Q8_0.gguf |
|
``` |
|
|
|
--- |
|
|
|
### **Step 5: Test the Model** |
|
Try prompting it: |
|
```sh |
|
llama-cli -m ~/llama_models/Q8_0.gguf -p "Explain quantum computing in simple terms." |
|
``` |
|
or interactively: |
|
```sh |
|
llama-cli -m ~/llama_models/Q8_0.gguf --interactive |
|
``` |
|
|
|
## How to Use Model via Python |
|
|
|
You can load this model with `transformers`: |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model_name = "eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small" |
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
|
# Example prompt |
|
prompt = "A farmer has 24 apples. He gives 6 to each of his 3 children. How many does he have left?" |
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
output = model.generate(**inputs, max_length=200) |
|
print(tokenizer.decode(output[0], skip_special_tokens=True)) |
|
``` |
|
|
|
## Expected Performance |
|
|
|
Compared to the base **Qwen2.5-3B-Instruct**, this fine-tuned model: |
|
- Provides **more detailed explanations** when answering GSM8K math problems. |
|
- Improves **logical reasoning** by incorporating DeepSeek-style augmented reasoning. |
|
- Generates **clearer step-by-step solutions**, making it useful for educational or tutoring applications. |
|
|
|
## Model Directory |
|
|
|
The model is hosted on **Hugging Face Hub**: |
|
👉 **[eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small](https://huggingface.co/eagle0504/qwen-2-5-3b-instruct-using-openai-gsm8k-gguf-data-enhanced-with-deepseek-v3-small)** |
|
|
|
## License |
|
|
|
This model is released under the **MIT License**, allowing open usage and modifications. |
|
|
|
--- |
|
|
|
If you have any questions or suggestions for improvements, feel free to reach out! |
|
|