Update README.md
Browse files
README.md
CHANGED
@@ -61,6 +61,7 @@ To this end, we introduce two new corpora: one comprises PNAS abstract-significa
|
|
61 |
We finetune the scientifc abstract simplification task using an encoder-decoder Transformer model (a variant of Flan-T5).
|
62 |
Our model is first tuned with multiple discrete instructions by mixing four relevant tasks in a challenge-proportional manner.
|
63 |
Then we continue tuning the model solely with the abstract-significance corpus.
|
|
|
64 |
We hope our work can pave the last mile of scientific understanding and let people better enjoy the fruits of open science.
|
65 |
|
66 |
|
@@ -75,7 +76,6 @@ We hope our work can pave the last mile of scientific understanding and let peop
|
|
75 |
- **Parent Model:** [FLAN-T5-large](https://huggingface.co/google/flan-t5-large)
|
76 |
|
77 |
|
78 |
-
|
79 |
# Usage
|
80 |
|
81 |
Use the code below to get started with the model. Remember to prepend the `INSTRUCTION` for best performance.
|
@@ -109,7 +109,7 @@ print(tokenizer.decode(decoded_ids[0], skip_special_tokens=True))
|
|
109 |
|----------------------------------|-----------------------------|------------------------------------|--------------------------------------|--------------------------------|----------------------------------------------------------------------------------------------------------------------------------------|
|
110 |
| Scientific Abstract-Significance | 3,030/200/200 | 707,071, 375,433 | 45,697, 24,901 | 46,985, 24,426 | - |
|
111 |
| Editor Abstract | 732/91/92 | 154,808, 194,721 | 19,675, 24,421 | 19,539, 24,332 | - |
|
112 |
-
| Wiki Auto | 28,364/1,000/1,000 | 18,239,990, 12,547,272 | 643,157, 444,034 | 642549,
|
113 |
| CNN/DailyMail | 287,113/13,368/11,490 | - | - | - | We used the 2.0 version, adopted from Huggingface datasets. |
|
114 |
|
115 |
|
@@ -130,7 +130,7 @@ We finetuned the base model (flan-t5-large) on multiple relevant tasks with stan
|
|
130 |
|
131 |
- Retuning: In this stage, we continued finetuning the checkpoint solely with the Scientific Abstract-Significance corpus till optimal validation loss was observed.
|
132 |
|
133 |
-
The multi-instruction tuning and the retuning took roughly 63 hours and 8 hours, respectively, on two NVIDIA RTX A5000 (24GB memory each) GPUs. We saved the checkpoint with the lowest validation loss for inference. We used the AdamW optimizer and a learning rate of 3e-5 with fully sharded data parallel strategy across training stages.
|
134 |
|
135 |
|
136 |
# Evaluation
|
|
|
61 |
We finetune the scientifc abstract simplification task using an encoder-decoder Transformer model (a variant of Flan-T5).
|
62 |
Our model is first tuned with multiple discrete instructions by mixing four relevant tasks in a challenge-proportional manner.
|
63 |
Then we continue tuning the model solely with the abstract-significance corpus.
|
64 |
+
The model can generate better lay summaries compared with models finetuned only with the abstract-significance corpus and models finetuned with task mixtures in traditonal ways.
|
65 |
We hope our work can pave the last mile of scientific understanding and let people better enjoy the fruits of open science.
|
66 |
|
67 |
|
|
|
76 |
- **Parent Model:** [FLAN-T5-large](https://huggingface.co/google/flan-t5-large)
|
77 |
|
78 |
|
|
|
79 |
# Usage
|
80 |
|
81 |
Use the code below to get started with the model. Remember to prepend the `INSTRUCTION` for best performance.
|
|
|
109 |
|----------------------------------|-----------------------------|------------------------------------|--------------------------------------|--------------------------------|----------------------------------------------------------------------------------------------------------------------------------------|
|
110 |
| Scientific Abstract-Significance | 3,030/200/200 | 707,071, 375,433 | 45,697, 24,901 | 46,985, 24,426 | - |
|
111 |
| Editor Abstract | 732/91/92 | 154,808, 194,721 | 19,675, 24,421 | 19,539, 24,332 | - |
|
112 |
+
| Wiki Auto | 28,364/1,000/1,000 | 18,239,990, 12,547,272 | 643,157, 444,034 | 642549, 444,883 | We used the ACL version, adopted from Huggingface datasets. The validation and test samples are split from the corpus and kept frozen. |
|
113 |
| CNN/DailyMail | 287,113/13,368/11,490 | - | - | - | We used the 2.0 version, adopted from Huggingface datasets. |
|
114 |
|
115 |
|
|
|
130 |
|
131 |
- Retuning: In this stage, we continued finetuning the checkpoint solely with the Scientific Abstract-Significance corpus till optimal validation loss was observed.
|
132 |
|
133 |
+
The multi-instruction tuning and the retuning took roughly 63 hours and 8 hours, respectively, on two NVIDIA RTX A5000 (24GB memory each) GPUs. We saved the checkpoint with the lowest validation loss for inference. We used the AdamW optimizer and a learning rate of 3e-5 with fully sharded data parallel strategy across training stages. The batch size equals to 1.
|
134 |
|
135 |
|
136 |
# Evaluation
|