| 
1 | 1 | dataset: squad_adversarial  | 
2 | 2 | subset: AddSent  | 
3 | 3 | templates:  | 
4 |  | -  048c2159-2c8c-40e2-90f7-18c9623381ba: !Template  | 
 | 4 | +  22a2f318-5302-479e-93be-215453060624: !Template  | 
5 | 5 |     answer_choices: null  | 
6 |  | -    id: 048c2159-2c8c-40e2-90f7-18c9623381ba  | 
7 |  | -    jinja: 'Generate a possible question for the following short passage:  | 
 | 6 | +    id: 22a2f318-5302-479e-93be-215453060624  | 
 | 7 | +    jinja: '{{context}}  | 
8 | 8 | 
  | 
9 | 9 | 
  | 
10 |  | -      {{context}} |||  | 
11 |  | -
  | 
12 |  | -      {{question}}'  | 
13 |  | -    metadata: !TemplateMetadata  | 
14 |  | -      choices_in_prompt: null  | 
15 |  | -      metrics: []  | 
16 |  | -      original_task: false  | 
17 |  | -    name: possible_qn  | 
18 |  | -    reference: ''  | 
19 |  | -  08fb6eac-6321-4b25-8578-14a799a103ed: !Template  | 
20 |  | -    answer_choices: null  | 
21 |  | -    id: 08fb6eac-6321-4b25-8578-14a799a103ed  | 
22 |  | -    jinja: 'After reading the following paragraph, please answer this question: {{question}}  | 
23 |  | -
  | 
24 |  | -
  | 
25 |  | -      {{context}}  | 
 | 10 | +      Q: {{question}}  | 
26 | 11 | 
  | 
27 | 12 | 
  | 
28 |  | -      |||  | 
29 |  | -
  | 
30 |  | -      {{answers[''text''] | most_frequent | choice}}'  | 
 | 13 | +      Referring to the passage above, the correct answer to the given question is  | 
 | 14 | +      ||| {{answers["text"][0]}}'  | 
31 | 15 |     metadata: !TemplateMetadata  | 
32 |  | -      choices_in_prompt: null  | 
33 |  | -      metrics: []  | 
 | 16 | +      choices_in_prompt: false  | 
 | 17 | +      metrics:  | 
 | 18 | +      - Squad  | 
34 | 19 |       original_task: true  | 
35 |  | -    name: after  | 
36 |  | -    reference: ''  | 
37 |  | -  1f2c2108-441a-4b3c-a5c8-8ece28edb6e1: !Template  | 
38 |  | -    answer_choices: null  | 
39 |  | -    id: 1f2c2108-441a-4b3c-a5c8-8ece28edb6e1  | 
40 |  | -    jinja: 'At what character does the text "{{answers["text"][0]}}" start in the  | 
41 |  | -      following paragraph?  | 
42 |  | -
  | 
43 |  | -
  | 
44 |  | -      {{context}}  | 
45 |  | -
  | 
46 |  | -
  | 
47 |  | -      |||  | 
48 |  | -
  | 
49 |  | -      {{answers["answer_start"][0]}}'  | 
50 |  | -    metadata: !TemplateMetadata  | 
51 |  | -      choices_in_prompt: null  | 
52 |  | -      metrics: []  | 
53 |  | -      original_task: false  | 
54 |  | -    name: find text  | 
55 |  | -    reference: ''  | 
56 |  | -  279e4019-8d67-498d-8832-a7905bc0c68d: !Template  | 
57 |  | -    answer_choices: null  | 
58 |  | -    id: 279e4019-8d67-498d-8832-a7905bc0c68d  | 
59 |  | -    jinja: 'Use the following non-answers to generate a possible short passage-question  | 
60 |  | -      pair:  | 
61 |  | -
  | 
62 |  | -      {{answers["text"]|join('', '')}} |||  | 
63 |  | -
  | 
64 |  | -      {{context}}  | 
65 |  | -
  | 
66 |  | -      {{question}}  | 
67 |  | -
  | 
68 |  | -      '  | 
69 |  | -    metadata: !TemplateMetadata  | 
70 |  | -      choices_in_prompt: null  | 
71 |  | -      metrics: []  | 
72 |  | -      original_task: false  | 
73 |  | -    name: answers_question  | 
 | 20 | +    name: answer_given_context_and_question  | 
74 | 21 |     reference: ''  | 
75 |  | -  44df6bac-bffa-4e46-b2d4-f3eb5b43cefa: !Template  | 
 | 22 | +  402adce7-4857-4524-8ad3-6270b66a5e0f: !Template  | 
76 | 23 |     answer_choices: null  | 
77 |  | -    id: 44df6bac-bffa-4e46-b2d4-f3eb5b43cefa  | 
78 |  | -    jinja: 'Generate a title for the following short passage:  | 
 | 24 | +    id: 402adce7-4857-4524-8ad3-6270b66a5e0f  | 
 | 25 | +    jinja: 'Refer to the passage below and answer the following question:  | 
79 | 26 | 
  | 
80 | 27 | 
  | 
81 |  | -      {{context}} |||  | 
 | 28 | +      Passage: {{context}}  | 
82 | 29 | 
  | 
83 |  | -      {{title|replace("_"," ")}}  | 
84 |  | -
  | 
85 |  | -      '  | 
86 |  | -    metadata: !TemplateMetadata  | 
87 |  | -      choices_in_prompt: null  | 
88 |  | -      metrics: []  | 
89 |  | -      original_task: false  | 
90 |  | -    name: title  | 
91 |  | -    reference: ''  | 
92 |  | -  60ae905d-d5fa-4f60-bbcb-acb8d0ec2cf1: !Template  | 
93 |  | -    answer_choices: null  | 
94 |  | -    id: 60ae905d-d5fa-4f60-bbcb-acb8d0ec2cf1  | 
95 |  | -    jinja: "Q: {{question}}\n\nA: \n|||\n{{answers['text'] | most_frequent | choice}}"  | 
96 |  | -    metadata: !TemplateMetadata  | 
97 |  | -      choices_in_prompt: null  | 
98 |  | -      metrics: []  | 
99 |  | -      original_task: false  | 
100 |  | -    name: cbqa qa  | 
101 |  | -    reference: ''  | 
102 |  | -  6118ec43-d051-4599-b24f-8779f66b9ad6: !Template  | 
103 |  | -    answer_choices: null  | 
104 |  | -    id: 6118ec43-d051-4599-b24f-8779f66b9ad6  | 
105 |  | -    jinja: '{{question}}  | 
106 | 30 | 
  | 
 | 31 | +      Question: {{question}}  | 
107 | 32 | 
  | 
108 | 33 |       |||  | 
109 | 34 | 
  | 
110 |  | -
  | 
111 |  | -      {{answers[''text''] | most_frequent | choice}}'  | 
 | 35 | +      {{answers["text"][0]}}'  | 
112 | 36 |     metadata: !TemplateMetadata  | 
113 |  | -      choices_in_prompt: null  | 
114 |  | -      metrics: []  | 
115 |  | -      original_task: false  | 
116 |  | -    name: cbqa  | 
117 |  | -    reference: ''  | 
118 |  | -  754e8bad-454f-4ae3-9747-299506955569: !Template  | 
119 |  | -    answer_choices: null  | 
120 |  | -    id: 754e8bad-454f-4ae3-9747-299506955569  | 
121 |  | -    jinja: 'Please come up with a good question to test reading comprehension about  | 
122 |  | -      the following paragraph:  | 
123 |  | -
  | 
124 |  | -
  | 
125 |  | -      {{context}}  | 
126 |  | -
  | 
127 |  | -
  | 
128 |  | -      |||  | 
129 |  | -
  | 
130 |  | -
  | 
131 |  | -      {{question}}'  | 
132 |  | -    metadata: !TemplateMetadata  | 
133 |  | -      choices_in_prompt: null  | 
134 |  | -      metrics: []  | 
135 |  | -      original_task: null  | 
136 |  | -    name: generate question  | 
137 |  | -    reference: ''  | 
138 |  | -  7ff4bc14-08d4-47c1-9cd3-b7473d6505e7: !Template  | 
139 |  | -    answer_choices: null  | 
140 |  | -    id: 7ff4bc14-08d4-47c1-9cd3-b7473d6505e7  | 
141 |  | -    jinja: 'For the following passage-question pair, list all possible wrong answers  | 
142 |  | -      (pitfalls) test-takers may choose:  | 
143 |  | -
  | 
144 |  | -
  | 
145 |  | -      {{context}}  | 
146 |  | -
  | 
147 |  | -      {{question}} |||  | 
148 |  | -
  | 
149 |  | -      {{answers["text"]|join(", ")}}'  | 
150 |  | -    metadata: !TemplateMetadata  | 
151 |  | -      choices_in_prompt: null  | 
152 |  | -      metrics: []  | 
 | 37 | +      choices_in_prompt: false  | 
 | 38 | +      metrics:  | 
 | 39 | +      - Squad  | 
153 | 40 |       original_task: true  | 
154 |  | -    name: possible_pitfalls  | 
 | 41 | +    name: answer_question_given_context  | 
155 | 42 |     reference: ''  | 
156 |  | -  88b952a3-3784-43bb-a463-4a34478785d5: !Template  | 
 | 43 | +  b4994c82-bfb2-4e0c-a5d7-081053830097: !Template  | 
157 | 44 |     answer_choices: null  | 
158 |  | -    id: 88b952a3-3784-43bb-a463-4a34478785d5  | 
159 |  | -    jinja: '{{["Question", "Problem"]  | choice}} {{range(1, 12) | choice}}: {{question}}  | 
160 |  | -
  | 
161 |  | -
  | 
162 |  | -      Hint: {{context}}  | 
 | 45 | +    id: b4994c82-bfb2-4e0c-a5d7-081053830097  | 
 | 46 | +    jinja: '{{context}}  | 
163 | 47 | 
  | 
164 | 48 | 
  | 
165 |  | -      |||  | 
166 |  | -
  | 
167 |  | -      {{answers["text"] | most_frequent | choice}}'  | 
 | 49 | +      From the above passage, a reasonable question with "{{answers["text"][0]}}"  | 
 | 50 | +      as the answer would be: ||| {{question}}'  | 
168 | 51 |     metadata: !TemplateMetadata  | 
169 |  | -      choices_in_prompt: null  | 
170 |  | -      metrics: []  | 
 | 52 | +      choices_in_prompt: false  | 
 | 53 | +      metrics:  | 
 | 54 | +      - BLEU  | 
 | 55 | +      - ROUGE  | 
171 | 56 |       original_task: false  | 
172 |  | -    name: question/hint  | 
173 |  | -    reference: ''  | 
174 |  | -  8bcc0d77-6925-4fa1-b8cc-e6da3b272197: !Template  | 
 | 57 | +    name: jeopardy  | 
 | 58 | +    reference: jeopardy style- wiki_qa  | 
 | 59 | +  b60cd43d-7026-434b-abf8-f67cc965316a: !Template  | 
175 | 60 |     answer_choices: null  | 
176 |  | -    id: 8bcc0d77-6925-4fa1-b8cc-e6da3b272197  | 
177 |  | -    jinja: "Question: {{question}}\n\nAnswer: \n|||\n{{answers['text'] | most_frequent\  | 
178 |  | -      \ | choice}}"  | 
179 |  | -    metadata: !TemplateMetadata  | 
180 |  | -      choices_in_prompt: null  | 
181 |  | -      metrics: []  | 
182 |  | -      original_task: false  | 
183 |  | -    name: cbqa question answer  | 
184 |  | -    reference: ''  | 
185 |  | -  a99d7cf5-d723-4c7a-b843-e2b8a476754d: !Template  | 
186 |  | -    answer_choices: null  | 
187 |  | -    id: a99d7cf5-d723-4c7a-b843-e2b8a476754d  | 
188 |  | -    jinja: 'I''ve always wondered: {{question}}  | 
189 |  | -
  | 
190 |  | -
  | 
191 |  | -      I searched Wikipedia and this is what I found. What''s the answer?  | 
192 |  | -
  | 
193 |  | -
  | 
194 |  | -      {{context}}  | 
195 |  | -
  | 
 | 61 | +    id: b60cd43d-7026-434b-abf8-f67cc965316a  | 
 | 62 | +    jinja: '{{context}}  | 
196 | 63 | 
  | 
197 |  | -      |||  | 
198 | 64 | 
  | 
199 |  | -      {{answers[''text''] | most_frequent | choice}}'  | 
 | 65 | +      Generate a question from the above passage : ||| {{question}}'  | 
200 | 66 |     metadata: !TemplateMetadata  | 
201 |  | -      choices_in_prompt: null  | 
202 |  | -      metrics: []  | 
 | 67 | +      choices_in_prompt: false  | 
 | 68 | +      metrics:  | 
 | 69 | +      - BLEU  | 
 | 70 | +      - ROUGE  | 
203 | 71 |       original_task: false  | 
204 |  | -    name: wondered  | 
 | 72 | +    name: given_context_generate_question  | 
205 | 73 |     reference: ''  | 
206 |  | -  a9d70ff7-8080-4eaa-9be2-1b67fe9b44f4: !Template  | 
 | 74 | +  dada0334-1dc2-4e39-a7e1-258ac622ab4f: !Template  | 
207 | 75 |     answer_choices: null  | 
208 |  | -    id: a9d70ff7-8080-4eaa-9be2-1b67fe9b44f4  | 
209 |  | -    jinja: 'I''m working on the final exam for my class and am trying to figure out  | 
210 |  | -      the answer to the question "{{question}}" I found the following info on Wikipedia  | 
211 |  | -      and I think it has the answer. Can you tell me the answer?  | 
212 |  | -
  | 
213 |  | -
  | 
214 |  | -      {{context}}  | 
215 |  | -
  | 
216 |  | -
  | 
217 |  | -      |||  | 
218 |  | -
  | 
219 |  | -      {{answers[''text''] | most_frequent | choice}}'  | 
220 |  | -    metadata: !TemplateMetadata  | 
221 |  | -      choices_in_prompt: null  | 
222 |  | -      metrics: []  | 
223 |  | -      original_task: false  | 
224 |  | -    name: exam  | 
225 |  | -    reference: ''  | 
226 |  | -  f086fa63-6ca2-48d2-857d-179ab88fce48: !Template  | 
227 |  | -    answer_choices: null  | 
228 |  | -    id: f086fa63-6ca2-48d2-857d-179ab88fce48  | 
229 |  | -    jinja: 'I''m creating a final exam for my reading class. Can you please come up  | 
230 |  | -      with a good question to quiz how well students have read the following text  | 
231 |  | -      snippet?  | 
232 |  | -
  | 
233 |  | -
  | 
234 |  | -      {{context}}  | 
235 |  | -
  | 
236 |  | -
  | 
237 |  | -      |||  | 
238 |  | -
  | 
239 |  | -
  | 
240 |  | -      {{question}}'  | 
 | 76 | +    id: dada0334-1dc2-4e39-a7e1-258ac622ab4f  | 
 | 77 | +    jinja: "{{context}}\n\nWith reference to the above context, {{question}} ||| \n\  | 
 | 78 | +      \n{{answers.text[0]}}"  | 
241 | 79 |     metadata: !TemplateMetadata  | 
242 |  | -      choices_in_prompt: null  | 
243 |  | -      metrics: []  | 
244 |  | -      original_task: null  | 
245 |  | -    name: exam creation help  | 
246 |  | -    reference: ''  | 
247 |  | -  f9b51e3b-a41a-47a5-b929-76a1e0efd430: !Template  | 
248 |  | -    answer_choices: null  | 
249 |  | -    id: f9b51e3b-a41a-47a5-b929-76a1e0efd430  | 
250 |  | -    jinja: 'Count the characters up until "{{answers["text"][0]}}" appears in the  | 
251 |  | -      following chunk of text.  | 
252 |  | -
  | 
253 |  | -
  | 
254 |  | -      {{context}}  | 
255 |  | -
  | 
256 |  | -
  | 
257 |  | -      |||  | 
258 |  | -
  | 
259 |  | -
  | 
260 |  | -      {{answers["answer_start"][0]}}'  | 
261 |  | -    metadata: !TemplateMetadata  | 
262 |  | -      choices_in_prompt: null  | 
263 |  | -      metrics: []  | 
264 |  | -      original_task: null  | 
265 |  | -    name: count letters  | 
 | 80 | +      choices_in_prompt: false  | 
 | 81 | +      metrics:  | 
 | 82 | +      - Squad  | 
 | 83 | +      original_task: true  | 
 | 84 | +    name: answer_the_question  | 
266 | 85 |     reference: ''  | 
267 |  | -  fb81ba4d-341a-43f0-a94f-fa7e350d10c0: !Template  | 
 | 86 | +  e638bc9e-5059-4ace-a6f9-4871f548342f: !Template  | 
268 | 87 |     answer_choices: null  | 
269 |  | -    id: fb81ba4d-341a-43f0-a94f-fa7e350d10c0  | 
270 |  | -    jinja: 'List all possible non-answers that have a lot of words in common with  | 
271 |  | -      the following context-question pair:  | 
272 |  | -
  | 
 | 88 | +    id: e638bc9e-5059-4ace-a6f9-4871f548342f  | 
 | 89 | +    jinja: '{{context}}  | 
273 | 90 | 
  | 
274 |  | -      {{context}}  | 
275 | 91 | 
  | 
276 |  | -      {{question}} |||  | 
 | 92 | +      Q: {{question}}  | 
277 | 93 | 
  | 
278 |  | -      {{answers["text"]|join('', '')}}  | 
279 | 94 | 
  | 
280 |  | -      '  | 
 | 95 | +      A: ||| {{answers["text"][0]}}'  | 
281 | 96 |     metadata: !TemplateMetadata  | 
282 |  | -      choices_in_prompt: null  | 
283 |  | -      metrics: []  | 
 | 97 | +      choices_in_prompt: false  | 
 | 98 | +      metrics:  | 
 | 99 | +      - Squad  | 
284 | 100 |       original_task: true  | 
285 |  | -    name: incorrect_answers  | 
 | 101 | +    name: given_context_answer_question_variation  | 
286 | 102 |     reference: ''  | 
0 commit comments