|
--- |
|
license: apache-2.0 |
|
library_name: peft |
|
base_model: Qwen/Qwen1.5-7B |
|
model-index: |
|
- name: qwen_1.5_odia_7b |
|
results: [] |
|
--- |
|
|
|
# qwen_1.5_odia_7b (Pre-trained) |
|
|
|
Qwen_1.5_Odia_7B is a pre-trained Odia large language model with 7 billion parameters, and it is based on Qwen 1.5-7B. The model is pre-trained on the Culturex-Odia dataset, a filtered version of the original CulturaX dataset for Odia text. The training dataset contains 49 million tokens. The CulturaX-Odia dataset is sourced from mc4 and four distinct OSCAR corpora. |
|
|
|
For more details about the model, data, training procedure, and evaluations, go through the blog [post](https://www.odiagenai.org/blog/odiagenai-released-odia-pre-trained-llm-based-on-qwen). |
|
|
|
## Model Description |
|
* Model type: A 7B pre-trained decoder-only model |
|
* Primary Language(s): Odia and English |
|
* License: Apache-2.0 (Commercial) |
|
|
|
**NOTE** |
|
|
|
This is not an instruction-tuned model, so it may not be able to follow human instructions without using one/few-shot learning or instruction fine-tuning. The model has no moderation mechanisms and may generate harmful or inappropriate responses. |
|
It is recommended to first fine-tune it on the task(s) you are interested in. |
|
|
|
|
|
### Citation Information |
|
|
|
If you find this model useful, please consider giving ๐ and citing: |
|
|
|
``` |
|
@misc{Qwen1.5_odia_7b, |
|
author = {Sambit Sekhar and Shantipriya Parida and Debasish Dhal}, |
|
title = {Introducing OdiaGenAI's Qwen-Based Pre-trained LLM for Odia Language}, |
|
year = {2023}, |
|
publisher = {Hugging Face}, |
|
journal = {Hugging Face repository}, |
|
howpublished = {\url{https://huggingface.co/OdiaGenAI}}, |
|
} |
|
``` |
|
|
|
### Contributions |
|
|
|
- Sambit Sekhar |
|
- Shantipriya Parida |
|
- Debasish Dhal |