Hierarchical BERT
Collection
Set of BERT models with Hierarchical attention pre-trained on conversational data to process multiple utterances at once
•
8 items
•
Updated
This model is a fine-tuned version of /gpfs/home/ikuzmin/hier-bert-pytorch/data/hibial-model on an unknown dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss | Accuracy |
---|---|---|---|---|
2.7518 | 1.55 | 25000 | 2.5873 | 0.5213 |
2.2587 | 3.1 | 50000 | 2.1487 | 0.5824 |