Trained polyglot 1.3B with the QLORA method using the Chatbot_data_for_Korean dataset.

The hyper-parameters used for training are as follows.

  • batch-size: 16
  • max_steps: 3000
  • Learning rate: 3e-4
  • Lora r: 8
  • Lora target modules: query_key_value

Prompt Template:

### ์งˆ๋ฌธ: {๋ฌธ์žฅ}
### ์‘๋‹ต: {๋ฌธ์žฅ}
### ์œ ํ˜•: {์ผ๋ฐ˜ ๋˜๋Š” ์—ฐ์• }
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support