--- language: multilingual license: mit tags: - transformer - summarization - translation - question-answering - english - arabic datasets: - miscovery/arabic_egypt_english_world_facts pipeline_tag: summarization library_name: transformers --- # Miscovery Transformer Model This model is a transformer-based encoder-decoder model for multiple NLP tasks: - Text summarization - Translation (English-Arabic) - Question-answering ## Model Architecture - Model type: miscovery - Number of parameters: 485674144 - Encoder layers: 12 - Decoder layers: 12 - Attention heads: 12 - Hidden size: 768 - Feed-forward size: 3072 ## Training The model was trained in two stages: 1. Pre-training on sentence rearrangement tasks 2. Fine-tuning on downstream tasks ## Usage 1. Install the package: ```bash pip install miscovery-model ``` 2. Run the model using a script: ```python from miscovery_model import standard_pipeline # Create a pipeline model = standard_pipeline("miscovery/model") # Use it result = model("Translate this to Arabic: What year did World War I begin?") print(result) ``` ## Limitations This model was trained on specific datasets and may not generalize well to all domains.