metadata
license: apache-2.0
language:
- en
pipeline_tag: text-generation
tags:
- pytorch
- gpt2
Nepal Legal GPT-2 (From Scratch Implementation)
This is a custom GPT-2 style transformer model trained from scratch on Nepal's legal documents. The model was implemented entirely in PyTorch without relying on pre-trained weights, specifically designed to understand and generate text related to Nepal's legal domain.
Model Details
- Model Architecture: GPT-2 style Transformer
- Parameters: ~1 million
- Context Length: 128 tokens
- Layers: 6
- Attention Heads: 6
- Embedding Dimension: 384
- Vocabulary Size: 50,257 (GPT-2 tokenizer)
- Training Data: Nepal Legal QA English Dataset
- License: MIT
Training Data Experimental
The model was trained on a specialized dataset of Nepal's legal documents in English, containing:
- Legal questions and answers
- Procedural instructions
- Legal definitions and explanations
- Court procedures and regulations