Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
base_model: unsloth/Qwen3-
|
3 |
tags:
|
4 |
- text-generation-inference
|
5 |
- transformers
|
@@ -17,7 +17,7 @@ pipeline_tag: text-generation
|
|
17 |
library_name: transformers
|
18 |
---
|
19 |
|
20 |
-
# π¦ Riko-Qwen3-
|
21 |
|
22 |
<div align="center">
|
23 |
<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>
|
@@ -25,16 +25,16 @@ library_name: transformers
|
|
25 |
|
26 |
## π Model Overview
|
27 |
|
28 |
-
**Riko-Qwen3-
|
29 |
|
30 |
-
- **Base Model:** unsloth/Qwen3-
|
31 |
- **Source Models:** Kimi K2 + Horizon Beta (alternating turns)
|
32 |
- **Project:** Project Horizon LLM
|
33 |
- **Developer:** subsectmusic
|
34 |
- **Training Framework:** Unsloth + Hugging Face TRL
|
35 |
- **Training Speed:** 2x faster optimization via Unsloth
|
36 |
- **License:** Apache 2.0
|
37 |
-
- **Model Size:**
|
38 |
- **Format Support:** GGUF compatible for Ollama deployment
|
39 |
|
40 |
## π Character Profile: Riko
|
@@ -55,7 +55,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
55 |
import torch
|
56 |
|
57 |
# Load model and tokenizer
|
58 |
-
model_name = "subsectmusic/riko-qwen3-
|
59 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
60 |
model = AutoModelForCausalLM.from_pretrained(
|
61 |
model_name,
|
@@ -68,10 +68,10 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
68 |
|
69 |
```bash
|
70 |
# Pull the GGUF model for Ollama
|
71 |
-
ollama pull subsectmusic/riko-qwen3-
|
72 |
|
73 |
# Start chatting with Riko
|
74 |
-
ollama run subsectmusic/riko-qwen3-
|
75 |
```
|
76 |
|
77 |
### Conversation Template
|
@@ -114,7 +114,7 @@ print(f"Riko: {response}")
|
|
114 |
- **Creative Writing:** Generate authentic tsundere character dialogue and interactions
|
115 |
- **Chatbot Applications:** Personality-driven AI assistant with character consistency
|
116 |
- **Entertainment:** Fun, character-consistent interactions with kitsune AI personality
|
117 |
-
- **Research:** Study knowledge distillation from larger models (Kimi K2 β Qwen3-
|
118 |
- **Educational:** Understanding Project Horizon LLM methodology and alternating training approaches
|
119 |
|
120 |
## π¬ Project Horizon LLM Methodology
|
@@ -125,7 +125,7 @@ print(f"Riko: {response}")
|
|
125 |
- **Source Models:**
|
126 |
- **Kimi K2** (Turn 1, 3, 5... responses)
|
127 |
- **Horizon Beta** (Turn 2, 4, 6... responses) - OpenRouter's cloaked model (#2 Translation, #3 Programming)
|
128 |
-
- **Target Model:** Qwen3-
|
129 |
- **Knowledge Transfer:** Personality traits and response patterns from both high-quality models
|
130 |
- **Character Focus:** Specialized curation for tsundere kitsune personality (Riko)
|
131 |
|
@@ -138,7 +138,7 @@ The training methodology involves:
|
|
138 |
4. **Alternating Pattern:** Continue alternating between Kimi K2 and Horizon Beta for each turn
|
139 |
5. **Response Curation:** Select and refine responses that best match Riko's tsundere personality
|
140 |
6. **Dataset Compilation:** Combine curated human queries with personality-matched responses
|
141 |
-
7. **Fine-tuning:** Train Qwen3-
|
142 |
|
143 |
This approach ensures:
|
144 |
- **Personality Consistency:** Responses align with Riko's tsundere kitsune character
|
@@ -181,7 +181,7 @@ Sequence Length: Dynamic (up to context limit)
|
|
181 |
| Attribute | Details |
|
182 |
|-----------|---------|
|
183 |
| Architecture | Qwen3 Transformer |
|
184 |
-
| Parameters |
|
185 |
| Source Models | Kimi K2 + Horizon Beta (alternating) |
|
186 |
| Project | Project Horizon LLM |
|
187 |
| Context Length | Model dependent |
|
|
|
1 |
---
|
2 |
+
base_model: unsloth/Qwen3-7b-Base-unsloth-bnb-4bit
|
3 |
tags:
|
4 |
- text-generation-inference
|
5 |
- transformers
|
|
|
17 |
library_name: transformers
|
18 |
---
|
19 |
|
20 |
+
# π¦ Riko-Qwen3-7b: Tsundere Kitsune AI
|
21 |
|
22 |
<div align="center">
|
23 |
<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>
|
|
|
25 |
|
26 |
## π Model Overview
|
27 |
|
28 |
+
**Riko-Qwen3-7b** is a specialized conversational AI model fine-tuned to embody the personality of Riko, a tsundere kitsune character. Part of **Project Horizon LLM**, this model was trained using alternating responses from Kimi K2 and Horizon Beta, built on the robust Qwen3-7b foundation, delivering engaging, personality-driven conversations with authentic tsundere characteristics.
|
29 |
|
30 |
+
- **Base Model:** unsloth/Qwen3-7b-Base-unsloth-bnb-4bit
|
31 |
- **Source Models:** Kimi K2 + Horizon Beta (alternating turns)
|
32 |
- **Project:** Project Horizon LLM
|
33 |
- **Developer:** subsectmusic
|
34 |
- **Training Framework:** Unsloth + Hugging Face TRL
|
35 |
- **Training Speed:** 2x faster optimization via Unsloth
|
36 |
- **License:** Apache 2.0
|
37 |
+
- **Model Size:** 7b parameters (4-bit quantized)
|
38 |
- **Format Support:** GGUF compatible for Ollama deployment
|
39 |
|
40 |
## π Character Profile: Riko
|
|
|
55 |
import torch
|
56 |
|
57 |
# Load model and tokenizer
|
58 |
+
model_name = "subsectmusic/riko-qwen3-7b"
|
59 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
60 |
model = AutoModelForCausalLM.from_pretrained(
|
61 |
model_name,
|
|
|
68 |
|
69 |
```bash
|
70 |
# Pull the GGUF model for Ollama
|
71 |
+
ollama pull subsectmusic/riko-qwen3-7b
|
72 |
|
73 |
# Start chatting with Riko
|
74 |
+
ollama run subsectmusic/riko-qwen3-7b
|
75 |
```
|
76 |
|
77 |
### Conversation Template
|
|
|
114 |
- **Creative Writing:** Generate authentic tsundere character dialogue and interactions
|
115 |
- **Chatbot Applications:** Personality-driven AI assistant with character consistency
|
116 |
- **Entertainment:** Fun, character-consistent interactions with kitsune AI personality
|
117 |
+
- **Research:** Study knowledge distillation from larger models (Kimi K2 β Qwen3-7b)
|
118 |
- **Educational:** Understanding Project Horizon LLM methodology and alternating training approaches
|
119 |
|
120 |
## π¬ Project Horizon LLM Methodology
|
|
|
125 |
- **Source Models:**
|
126 |
- **Kimi K2** (Turn 1, 3, 5... responses)
|
127 |
- **Horizon Beta** (Turn 2, 4, 6... responses) - OpenRouter's cloaked model (#2 Translation, #3 Programming)
|
128 |
+
- **Target Model:** Qwen3-7b (student model)
|
129 |
- **Knowledge Transfer:** Personality traits and response patterns from both high-quality models
|
130 |
- **Character Focus:** Specialized curation for tsundere kitsune personality (Riko)
|
131 |
|
|
|
138 |
4. **Alternating Pattern:** Continue alternating between Kimi K2 and Horizon Beta for each turn
|
139 |
5. **Response Curation:** Select and refine responses that best match Riko's tsundere personality
|
140 |
6. **Dataset Compilation:** Combine curated human queries with personality-matched responses
|
141 |
+
7. **Fine-tuning:** Train Qwen3-7b on the curated dataset using Unsloth + TRL
|
142 |
|
143 |
This approach ensures:
|
144 |
- **Personality Consistency:** Responses align with Riko's tsundere kitsune character
|
|
|
181 |
| Attribute | Details |
|
182 |
|-----------|---------|
|
183 |
| Architecture | Qwen3 Transformer |
|
184 |
+
| Parameters | 7b (4-bit quantized) |
|
185 |
| Source Models | Kimi K2 + Horizon Beta (alternating) |
|
186 |
| Project | Project Horizon LLM |
|
187 |
| Context Length | Model dependent |
|