Trained polyglot 1.3B with the QLORA method using the nsmc dataset.
The hyper-parameters used for training are as follows.
- batch-size: 16
- max_steps: 10000
- Learning rate: 3e-4
- Lora r: 8
- Lora target modules: query_key_value
Prompt Template:
### ๋ฌธ์ฅ: {๋ฌธ์ฅ}
### ๊ฐ์ : {๊ธ์  ๋๋ ๋ถ์ }
	Inference Providers
	NEW
	
	
	This model isn't deployed by any Inference Provider.
	๐
			
		Ask for provider support