Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ This is a distilled cross-encoder model based on ModernBERT-large, trained to pr
|
|
11 |
- **Base Model**: [answerdotai/ModernBERT-large](https://huggingface.co/answerdotai/ModernBERT-large)
|
12 |
- **Architecture**: Cross-encoder with MLP head (1024 → 512 → 256 → 1)
|
13 |
- **Max Sequence Length**: 4096
|
14 |
-
- **Training Data**: MMLU Pro Subset (500 queries) scored by 35 different LM Judges and reward models, aggregated to form sample-level scores with Weaver
|
15 |
- **Training Objective**: Binary classification (correct/incorrect answer prediction)
|
16 |
|
17 |
## Usage
|
|
|
11 |
- **Base Model**: [answerdotai/ModernBERT-large](https://huggingface.co/answerdotai/ModernBERT-large)
|
12 |
- **Architecture**: Cross-encoder with MLP head (1024 → 512 → 256 → 1)
|
13 |
- **Max Sequence Length**: 4096
|
14 |
+
- **Training Data**: [MMLU Pro](https://huggingface.co/datasets/TIGER-Lab/MMLU-Pro) Subset (500 queries) scored by 35 different LM Judges and reward models, aggregated to form sample-level scores with Weaver
|
15 |
- **Training Objective**: Binary classification (correct/incorrect answer prediction)
|
16 |
|
17 |
## Usage
|