metadata
library_name: transformers
tags:
- llama-factory
license: apache-2.0
datasets:
- hon9kon9ize/yue-alpaca
- Nin8520/words
language:
- zh
- en
base_model:
- Qwen/Qwen2.5-7B-Instruct
Model Details
- Base Model: Qwen2.5-7B-Instruct
Training Details
Training Data
Training Setting
args = dict(
stage="sft",
do_train=True,
model_name_or_path="Qwen/Qwen2.5-7B-Instruct",
dataset="yue_1,yue_2",
template="alpaca",
finetuning_type="lora",
lora_target="all",
output_dir="Qwen2.5_lora",
per_device_train_batch_size=2,
gradient_accumulation_steps=4,
lr_scheduler_type="cosine",
logging_steps=5,
warmup_ratio=0.1,
save_steps=1000,
learning_rate=5e-5,
num_train_epochs=1.0,
max_samples=300,
max_grad_norm=1.0,
loraplus_lr_ratio=16.0,
fp16=True,
report_to="none",
)
Evaluation
Not done