Skip to content

Commit 7c0e400

Browse files
committed
update README
1 parent 05e19ae commit 7c0e400

File tree

1 file changed

+8
-9
lines changed

1 file changed

+8
-9
lines changed

README.md

Lines changed: 8 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,9 @@
11
# Tk-Instruct
22

3-
- This repo releases our implementation for the Tk-Instruct model in the [Natural Instructions V2 paper](https://arxiv.org/abs/2204.07705).
3+
- This repo releases our implementation for the Tk-Instruct model in the [Super-NaturalInstructions paper](https://arxiv.org/abs/2204.07705).
44
- Tk-Instruct is a preliminary attempt towards general-purpose AI that can solve many NLP tasks by following in-context instructions (plain language task definitions or k-shot examples).
55
- It is built based on the pretrained [T5 model](https://arxiv.org/abs/1910.10683), and finetuned on our [data](https://github.com/allenai/natural-instructions).
6-
- We are going to publish our demo soon. Stay tuned!
7-
<!-- - You can play with this model via our online [demo](https://instructions.apps.allenai.org/demo)! -->
6+
- You can play with the 11B model via our online [demo](https://instructions.apps.allenai.org/demo)!
87

98
## Requirements
109

@@ -26,13 +25,13 @@ Note: after the main exploration with 3B model, we train our 11B model on TPUs u
2625

2726
## Data
2827

29-
Our models are trained and evaluated on [Natural Instructions V2](https://github.com/allenai/natural-instructions), which can be cloned by running:
28+
Our models are trained and evaluated on [Super-NaturalInstructions](https://github.com/allenai/natural-instructions), which can be cloned by running:
3029

3130
```bash
3231
git clone git@github.com:allenai/natural-instructions.git data
3332
```
3433

35-
Since Natural Instructions V2 didn't provide an official split for the development set, in order to do evaluation during training time, you can mannualy create a `dev_tasks.txt` in the `data/splits/default` folder. We found it unclear what should be a meaningful validation set, under such cross-task generalization setting. You can use a part of the training tasks for validation, or you can set apart tasks in some categories for validation.
34+
Since Super-NaturalInstructions didn't provide an official split for the development set, in order to do evaluation during training time, you can mannualy create a `dev_tasks.txt` in the `data/splits/default` folder. We found it unclear what should be a meaningful validation set, under such cross-task generalization setting. You can use a part of the training tasks for validation, or you can set apart tasks in some categories for validation.
3635

3736
If you want to use the T5 code [here](https://github.com/google-research/text-to-text-transfer-transformer), you can convert the data into text2text format with [`scripts/convert_data_to_s2s.sh`](scripts/convert_data_to_s2s.sh).
3837

@@ -119,10 +118,10 @@ We will keep adding the predictions and performance of new models into this repo
119118
## Citation
120119

121120
```bib
122-
@article{wang2022benchmarking,
123-
title={Benchmarking Generalization via In-Context Instructions on 1,600+ Language Tasks},
124-
author={Wang, Yizhong and Mishra, Swaroop and Alipoormolabashi, Pegah and Kordi, Yeganeh and others},
125-
journal={arXiv preprint arXiv:2204.07705},
121+
@inproceedings{supernaturalinstructions,
122+
title={Super-NaturalInstructions:Generalization via Declarative Instructions on 1600+ Tasks},
123+
author={Wang, Yizhong and Mishra, Swaroop and Alipoormolabashi, Pegah and Kordi, Yeganeh and Mirzaei, Amirreza and Arunkumar, Anjana and Ashok, Arjun and Dhanasekaran, Arut Selvan and Naik, Atharva and Stap, David and others},
124+
booktitle={EMNLP},
126125
year={2022}
127126
}
128127
```

0 commit comments

Comments
 (0)