Configuration Parsing Warning: In adapter_config.json: "peft.base_model_name_or_path" must be a string

meta-llama/Llama-3.2-3B Fine-tuned with QLora

This model is a fine-tuned version of meta-llama/Llama-3.2-3B using the LoRA on the glue/qnli dataset.

πŸš€ Training Details

Fine-tuning Configuration

  • Base Model: meta-llama/Llama-3.2-3B
  • Quantization: 4-bit (NF4) compute.
  • LoRA Rank: 16
  • LoRA Alpha: 32
  • Batch Size: 8 (per device)
  • Gradient Accumulation: 4
  • Learning Rate: 2e-5
  • Sequence Length: 1024 tokens
  • Gradient Checkpointing: Enabled

πŸ“Š Training Metrics

  • Total Steps: 3274
  • Final Loss: 1.6037503279194865
  • Trainable Params: 24,313,856

## βš–οΈ License
This model inherits the Apache 2.0 license.
Downloads last month
3
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for D1zzYzz/QNLI-QLORA-llama-3.2-3B

Adapter
(208)
this model