Miscovery Transformer Model
This model is a transformer-based encoder-decoder model for multiple NLP tasks:
- Text summarization
- Translation (English-Arabic)
- Question-answering
Model Architecture
- Model type: miscovery
- Number of parameters: 485674144
- Encoder layers: 12
- Decoder layers: 12
- Attention heads: 12
- Hidden size: 768
- Feed-forward size: 3072
Training
The model was trained in two stages:
- Pre-training on sentence rearrangement tasks
- Fine-tuning on downstream tasks
Usage
- Install the package:
pip install miscovery-model
- Run the model using a script:
from miscovery_model import standard_pipeline
# Create a pipeline
model = standard_pipeline("miscovery/model")
# Use it
result = model("Translate this to Arabic: What year did World War I begin?")
print(result)
Limitations
This model was trained on specific datasets and may not generalize well to all domains.
- Downloads last month
- 258
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support