λ°μ΄ν° μ
LIMO
- GAIR/LIMO (μμ΄, μλ³Έ)
LIMO νκ΅μ΄ λ²μ
- exp-models/GAIR-LIMO-KOREAN (νκ΅μ΄ λ²μ)
- junnei/ko-limo (νκ΅μ΄ λ²μ)
νΉμ΄μ¬ν
- μλ LIMOμμλ 15 epoch νμ΅μ μνν¨
- μμ΄1+νκ΅μ΄2 λ°μ΄ν° μ μ μμ ν 5 epoch νμ΅μμΌ μλ νμ΅ λ°©λ²κ³Ό μ μ¬ν νμλ§νΌ, κ·Έλ¬λ μ½κ°μ λ³νμ΄ μλλ‘ νμ΅μν€λ €κ³ ν¨
- κ·Έλ¬λ μ μ± νκ°μμ 4 epoch μμ μ checkpointκ° κ°μ₯ μ±λ₯μ΄ μ’μ 보μμ
Training Details
- 4xH200 SXM, 13.5 Hours
Axolotl config
base_model: beomi/EXAONE-3.5-32B-Instruct-Llamafied
model_type: AutoModelForCausalLM
tokenizer_config: beomi/EXAONE-3.5-32B-Instruct-Llamafied
tokenizer_type: AutoTokenizer
load_in_8bit: false
load_in_4bit: false
strict: false
datasets:
- path: werty1248/kk_oo_llliiimmmooo
field_messages: conversations
type: chat_template
chat_template: tokenizer_default
dataset_prepared_path: ./data_preparation
output_dir: /workspace/data
hf_use_auth_token: true
sequence_len: 32768
sample_packing: false
pad_to_sequence_len: true
plugins:
- axolotl.integrations.liger.LigerPlugin
liger_rope: true
liger_rms_norm: true
liger_layer_norm: true
liger_glu_activation: true
liger_fused_linear_cross_entropy: true
wandb_project:
#wandb_entity:
#wandb_watch:
wandb_name:
#wandb_log_model:
gradient_accumulation_steps: 2
micro_batch_size: 1
num_epochs: 5
optimizer: paged_adamw_8bit
lr_scheduler: cosine
learning_rate: 5.0e-6
train_on_inputs: false
group_by_length: false
bf16: auto
fp16:
tf32: false
gradient_checkpointing: true
early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention:
flash_attention: true
warmup_ratio: 0.05
eval_table_size:
save_total_limit: 2
deepspeed: ./deepspeed_configs/zero3_bf16.json
special_tokens:
pad_token: "[|endofturn|]"
- Downloads last month
- 58
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.