Up and running in Hugging Face Space using 2 virtual cpu's and 16 GB RAM!! CFR-FineTuned_III

Llama-3.2-1B Fine-tuned on the Code of Federal Regulations (CFR)

This is a fine-tuned version of meta-llama/Meta-Llama-3.2-1B trained on all sections from the United States Code of Federal Regulations (CFR). The goal: provide a specialized assistant for navigating and answering questions about U.S. federal regulations.

Model Description

  • Base Model: Llama-3.2-1B
  • Method: QLoRA, 4-bit quantization
  • Dataset: Custom, parsed from CFR XML (Titles 1-50)
  • Epochs: 3
  • Tokens Seen: ~243M
  • Final Training Loss: 1.267
  • Mean Token Accuracy: 0.739
  • Training Time: ~5h 17m

Hardware/Environment:
Training was conducted on Modal using a single NVIDIA H200 GPU.
Training speed: ~1.10 steps/sec, 35 samples/sec.

Note: This loss is typical for a Llama-3 1B model on legal/complex text. For comparison: random output would yield >2.0; perfect memorization of a small dataset would yield <1.0. This is in the “actually learned something useful” range for this setup.

Intended Uses & Limitations

Intended Uses

  • Regulatory Q&A
  • Summarization of CFR text
  • Text generation related to U.S. federal regulations

Limitations

  • NOT a substitute for legal advice. Output may be incorrect or outdated (data as of 2024-06-25).
  • Can hallucinate—don’t trust answers without checking against the source.
  • Validation/test loss is not reported here (evaluate on your own task/data before using in production).

How to Use

You can use this model directly with the transformers library.

Downloads last month
65
Safetensors
Model size
1.24B params
Tensor type
F32
·
Inference Providers NEW