--- library_name: transformers license: apache-2.0 base_model: t5-small tags: - summarization - generated_from_trainer metrics: - rouge model-index: - name: t5-small-finetuned-stock-news-2 results: [] --- # t5-small-finetuned-stock-news-2 This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.8004 - Rouge1: 44.167 - Rouge2: 36.0987 - Rougel: 42.3071 - Rougelsum: 42.7469 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2.5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | 0.9333 | 1.0 | 567 | 0.8685 | 43.2839 | 35.0416 | 41.431 | 41.724 | | 0.8108 | 2.0 | 1134 | 0.8320 | 43.8366 | 35.6114 | 42.0042 | 42.3429 | | 0.7762 | 3.0 | 1701 | 0.8183 | 43.6494 | 35.4373 | 41.755 | 42.1571 | | 0.7487 | 4.0 | 2268 | 0.8084 | 43.8891 | 35.6917 | 41.9799 | 42.3663 | | 0.7289 | 5.0 | 2835 | 0.8036 | 43.7749 | 35.5598 | 41.851 | 42.2422 | | 0.7166 | 6.0 | 3402 | 0.8018 | 44.1163 | 36.0261 | 42.335 | 42.6929 | | 0.7014 | 7.0 | 3969 | 0.8015 | 44.0648 | 35.8877 | 42.1801 | 42.6115 | | 0.7119 | 8.0 | 4536 | 0.8004 | 44.167 | 36.0987 | 42.3071 | 42.7469 | ### Framework versions - Transformers 4.51.3 - Pytorch 2.6.0+cu124 - Datasets 3.5.1 - Tokenizers 0.21.1