Trained polyglot 5.8B with the QLORA method using the Chatbot_data_for_Korean dataset.
The hyper-parameters used for training are as follows.
- batch-size: 16
- max_steps: 3000
- Learning rate: 3e-4
- Lora r: 8
- Lora target modules: query_key_value
Prompt Template:
### ์ง๋ฌธ: {๋ฌธ์ฅ}
### ์๋ต: {๋ฌธ์ฅ}
### ์ ํ: {์ผ๋ฐ ๋๋ ์ฐ์ }
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support