meta-llama/Llama-3.2-3B Fine-tuned with Lora

This model is a fine-tuned version of meta-llama/Llama-3.2-3B using the LoRA on the google/boolq dataset.

πŸš€ Training Details

Fine-tuning Configuration

  • Base Model: meta-llama/Llama-3.2-3B
  • Quantization: 4-bit
  • LoRA Rank: 16
  • LoRA Alpha: 32
  • Batch Size: 8 (per device)
  • Gradient Accumulation: 4
  • Learning Rate: 2e-5
  • Sequence Length: 1024 tokens
  • Gradient Checkpointing: Enabled

πŸ“Š Training Metrics

  • Total Steps: 295
  • Final Loss: 1.5720999766204318
  • Trainable Params: 24,313,856

## βš–οΈ License
This model inherits the Apache 2.0 license.
Downloads last month
3
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for D1zzYzz/BOOLQ-LORA-llama-3.2-3B

Adapter
(212)
this model

Dataset used to train D1zzYzz/BOOLQ-LORA-llama-3.2-3B