You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+8-9Lines changed: 8 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,9 @@
1
1
# Tk-Instruct
2
2
3
-
- This repo releases our implementation for the Tk-Instruct model in the [Natural Instructions V2 paper](https://arxiv.org/abs/2204.07705).
3
+
- This repo releases our implementation for the Tk-Instruct model in the [Super-NaturalInstructions paper](https://arxiv.org/abs/2204.07705).
4
4
- Tk-Instruct is a preliminary attempt towards general-purpose AI that can solve many NLP tasks by following in-context instructions (plain language task definitions or k-shot examples).
5
5
- It is built based on the pretrained [T5 model](https://arxiv.org/abs/1910.10683), and finetuned on our [data](https://github.com/allenai/natural-instructions).
6
-
- We are going to publish our demo soon. Stay tuned!
7
-
<!-- - You can play with this model via our online [demo](https://instructions.apps.allenai.org/demo)! -->
6
+
- You can play with the 11B model via our online [demo](https://instructions.apps.allenai.org/demo)!
8
7
9
8
## Requirements
10
9
@@ -26,13 +25,13 @@ Note: after the main exploration with 3B model, we train our 11B model on TPUs u
26
25
27
26
## Data
28
27
29
-
Our models are trained and evaluated on [Natural Instructions V2](https://github.com/allenai/natural-instructions), which can be cloned by running:
28
+
Our models are trained and evaluated on [Super-NaturalInstructions](https://github.com/allenai/natural-instructions), which can be cloned by running:
30
29
31
30
```bash
32
31
git clone git@github.com:allenai/natural-instructions.git data
33
32
```
34
33
35
-
Since Natural Instructions V2 didn't provide an official split for the development set, in order to do evaluation during training time, you can mannualy create a `dev_tasks.txt` in the `data/splits/default` folder. We found it unclear what should be a meaningful validation set, under such cross-task generalization setting. You can use a part of the training tasks for validation, or you can set apart tasks in some categories for validation.
34
+
Since Super-NaturalInstructions didn't provide an official split for the development set, in order to do evaluation during training time, you can mannualy create a `dev_tasks.txt` in the `data/splits/default` folder. We found it unclear what should be a meaningful validation set, under such cross-task generalization setting. You can use a part of the training tasks for validation, or you can set apart tasks in some categories for validation.
36
35
37
36
If you want to use the T5 code [here](https://github.com/google-research/text-to-text-transfer-transformer), you can convert the data into text2text format with [`scripts/convert_data_to_s2s.sh`](scripts/convert_data_to_s2s.sh).
38
37
@@ -119,10 +118,10 @@ We will keep adding the predictions and performance of new models into this repo
119
118
## Citation
120
119
121
120
```bib
122
-
@article{wang2022benchmarking,
123
-
title={Benchmarking Generalization via In-Context Instructions on 1,600+ Language Tasks},
124
-
author={Wang, Yizhong and Mishra, Swaroop and Alipoormolabashi, Pegah and Kordi, Yeganeh and others},
125
-
journal={arXiv preprint arXiv:2204.07705},
121
+
@inproceedings{supernaturalinstructions,
122
+
title={Super-NaturalInstructions:Generalization via Declarative Instructions on 1600+ Tasks},
123
+
author={Wang, Yizhong and Mishra, Swaroop and Alipoormolabashi, Pegah and Kordi, Yeganeh and Mirzaei, Amirreza and Arunkumar, Anjana and Ashok, Arjun and Dhanasekaran, Arut Selvan and Naik, Atharva and Stap, David and others},
0 commit comments