|
---
|
|
library_name: transformers
|
|
license: apache-2.0
|
|
datasets:
|
|
- Manual-Dataset-Creation-Project/Malum-230
|
|
- llm-jp/oasst2-33k-ja
|
|
language:
|
|
- zho
|
|
- eng
|
|
- fra
|
|
- spa
|
|
- por
|
|
- deu
|
|
- ita
|
|
- rus
|
|
- jpn
|
|
- kor
|
|
- vie
|
|
- tha
|
|
- ara
|
|
base_model:
|
|
- Qwen/Qwen2.5-7B
|
|
inference: false
|
|
---
|
|
|
|
# Matsu-7B
|
|
|
|
## Description
|
|
Matsu-7B is a model that was instruction-tuned on the oasst2 and Malum-230, using Qwen2.5-7B as its base model.
|
|
|
|
## Series
|
|
| Variant | Link |
|
|
| --- | --- |
|
|
| Malum-230 | [Manual-Dataset-Creation-Project/Malum-230](https://huggingface.co/datasets/Manual-Dataset-Creation-Project/Malum-230) |
|
|
| Take-7B | [Manual-Dataset-Creation-Project/Take-7B](https://huggingface.co/Manual-Dataset-Creation-Project/Take-7B) |
|
|
|
|
## Contributors
|
|
- [Sudy](https://huggingface.co/sudy-super)
|
|
- [ほーりーふぉっくす](https://huggingface.co/Holy-fox)
|
|
|
|
## Acknowledgments
|
|
We would like to express our gratitude to [VOLTMIND](https://voltmind.jp/) for providing the computational resources used to train this model. |