hw-llama-2-7B-nsmc / README.md
Rfy23's picture
Update README.md
5e416e1
|
raw
history blame
1.48 kB
---
library_name: peft
base_model: KT-AI/midm-bitext-S-7B-inst-v1
---
# Model Card for Model ID
* KT-AI/midm-bitext-S-7B-inst-v1
## Model Details
### Model Description
* NSMC μ˜ν™” 리뷰 데이터에 λŒ€ν•˜μ—¬ KT-AI/midm-bitext-S-7B-inst-v1 λ―Έμ„ΈνŠœλ‹.
* μž…λ ₯ ν”„λ‘¬ν”„νŠΈλ₯Ό μ΄μš©ν•˜μ—¬ λ°μ΄ν„°μ…‹μ˜ document(리뷰)κ°€ 긍정적인 λ‚΄μš©μ΄λ©΄ '1'을 뢀정적인 λ‚΄μš©μ΄λ©΄ '0'을 μ˜ˆμΈ‘ν•˜λ„λ‘ 함.
* train data: nsmc train μƒμœ„ 2000개 μƒ˜ν”Œ 이용
* test data: nsmc test μƒμœ„ 2000개 μƒ˜ν”Œ 이용
### Training Data
'nsmc'
* μƒμœ„ 2000개 데이터 이용
### Training Procedure
* prepare_sample_text에 리뷰λ₯Ό 긍정/λΆ€μ •μœΌλ‘œ νŒλ‹¨ν•˜λ„λ‘ μž…λ ₯ ν”„λ‘¬ν”„νŠΈ μˆ˜μ •ν•˜μ˜€μŒ.
#### Training Hyperparameters
* per_device_train_batch_size: 1
* per_device_eval_batch_size: 1
* learning_rate: 1e-4
* gradient_accumulation_steps: 2
* optimizer: paged_adamw_32bit
* lr_scheduler_type: cosine
* lr_scheduler_warmup_ratio: 0.03
* training_args.logging_steps: 50
* training_args.max_steps : 1000
* trainable params: trainable params: 16,744,448 || all params: 7,034,347,520 || trainable%: 0.23803839591934178
### Results
TrainOutput(global_step=1000, training_loss=1.0208648338317872, metrics={'train_runtime': 1128.0266, 'train_samples_per_second': 1.773, 'train_steps_per_second': 0.887, 'total_flos': 3.1051694997504e+16, 'train_loss': 1.0208648338317872, 'epoch': 1.0})
#### Accruacy
λ―Έμ„ΈνŠœλ‹ ν›„ λͺ¨λΈμ˜ 정확도:0.61