add paper link
Browse files
README.md
CHANGED
|
@@ -84,6 +84,8 @@ A fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/goo
|
|
| 84 |
- 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
|
| 85 |
- all training used 16384 token input / 1024 max output
|
| 86 |
|
|
|
|
|
|
|
| 87 |
## How-To in Python
|
| 88 |
|
| 89 |
Install/update transformers `pip install -U transformers`
|
|
|
|
| 84 |
- 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
|
| 85 |
- all training used 16384 token input / 1024 max output
|
| 86 |
|
| 87 |
+
Read the paper by Guo et al. here: [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/pdf/2112.07916.pdf)
|
| 88 |
+
|
| 89 |
## How-To in Python
|
| 90 |
|
| 91 |
Install/update transformers `pip install -U transformers`
|