Trained polyglot 1.3B with the QLORA method using the nsmc dataset.

The hyper-parameters used for training are as follows.

  • batch-size: 16
  • max_steps: 10000
  • Learning rate: 3e-4
  • Lora r: 8
  • Lora target modules: query_key_value

Prompt Template:

### ๋ฌธ์žฅ: {๋ฌธ์žฅ}
### ๊ฐ์ •: {๊ธ์ • ๋˜๋Š” ๋ถ€์ •}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support