Trained polyglot 1.3B with the QLORA method using the nsmc dataset.
The hyper-parameters used for training are as follows.
- batch-size: 16
- max_steps: 10000
- Learning rate: 3e-4
- Lora r: 8
- Lora target modules: query_key_value
Prompt Template:
### ๋ฌธ์ฅ: {๋ฌธ์ฅ}
### ๊ฐ์ : {๊ธ์ ๋๋ ๋ถ์ }
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support