|
21 | 21 | Open Data for Education and Outreach. |
22 | 22 | ''' |
23 | 23 |
|
| 24 | +# For working with the metadata records |
24 | 25 | import json |
25 | 26 |
|
| 27 | +# For making deep copies of the evergreen data, since we want to change some things |
| 28 | +import copy |
| 29 | + |
26 | 30 | # Need new recids and DOIs |
27 | 31 |
|
28 | 32 | # Map of skim names into pretty-print descriptions |
|
78 | 82 | skim = 'no' # Fun little hack to fix the English... |
79 | 83 | name_short = '-'.join(aline.split('_')[2:4]).lower() |
80 | 84 | if 'Data' in aline: |
81 | | - name = f'Run 2 2015+2016 proton-proton collision data, {skim} skim' |
| 85 | + name = f'Run 2 2015+2016 proton-proton collision data beta release, {skim} skim' |
82 | 86 | else: |
83 | | - name = f'MC simulation, {skim} skim' |
| 87 | + name = f'MC simulation, 2015+2016 proton-proton collisions beta release, {skim} skim' |
84 | 88 | rec_doi = recid_doi_pairs.pop() |
85 | 89 | dataset_files[ aline.strip() ] = {'name_short':name_short, 'name':name, |
86 | 90 | 'categories':{'source':'ATLAS Collaboration'},'doi':rec_doi[1],'recid':rec_doi[0]} |
|
135 | 139 | }, |
136 | 140 | { |
137 | 141 | "description": "More about this ntuple format", |
138 | | - "url": "https://opendata.atlas.cern/docs/documentation/data_format/FEB2025_ntuple/" |
| 142 | + "url": "https://opendata.atlas.cern/docs/data/for_education/13TeV25_details" |
139 | 143 | }, |
140 | 144 | { |
141 | 145 | "description": "Ntuple making framework", |
|
165 | 169 | big_total_events = 0 |
166 | 170 | big_total_size = 0 |
167 | 171 |
|
| 172 | +# Now loop through all the datasets that we are going to publish |
168 | 173 | for adataset in dataset_files: |
169 | | - my_json = {} |
170 | | - # Update with the stuff that's always good |
171 | | - my_json.update(evergreen_data) |
| 174 | + # Start from the stuff that's always good |
| 175 | + my_json = copy.deepcopy(evergreen_data) |
172 | 176 | # Simple abstract for the collection |
173 | 177 | my_json['abstract'] = {'description':dataset_files[adataset]['name']+' from the ATLAS experiment'} |
174 | 178 | # Name of the collections, systematically set |
|
186 | 190 | my_json['doi'] = dataset_files[adataset]['doi'] |
187 | 191 | # Update the methodology section with the skim description |
188 | 192 | skim = adataset.split('_')[2] |
189 | | - my_json['methodology']['description'] += skim_name_map[skim] |
| 193 | + my_json['methodology']['description'] = evergreen_data['methodology']['description']+skim_name_map[skim] |
190 | 194 | # Add a record of the files for this dataset |
191 | 195 | my_json['files'] = [] |
192 | 196 | # Make list of files for this dataset |
|
235 | 239 | # Update with the stuff that's always good |
236 | 240 | my_json.update(evergreen_data) |
237 | 241 | # Simple abstract for the collection |
238 | | -my_json['abstract'] = {'description':'2015 Pb-Pb Open Data for Research from the ATLAS experiment'} |
| 242 | +my_json['abstract'] = {'description':'Run 2 2015+2016 proton-proton collision data and corresponding MC simulation Open Data for Education and Outreach from the ATLAS experiment'} |
239 | 243 | # Name of the collections, systematically set |
240 | 244 | my_json['collections'] = ['ATLAS-Simulated-Datasets','ATLAS-Primary-Datasets'] |
241 | 245 | my_json['type']['secondary'] = ['Simulated','Collision'] |
| 246 | +# Description needs a simple update |
| 247 | +my_json['methodology']['description'] = evergreen_data['methodology']['description'].replace('the following skimming selection was applied: ','several event pre-selections are available to accelerate analysis, as well as an inclusive set of all events.') |
242 | 248 | # Add categories, mostly for MC datasets |
243 | 249 | my_json['categories'] = {'source':'ATLAS Collaboration'} |
244 | | -my_json['title'] = 'ROOT ntuple format 2015-2016 proton-proton Open Data for Education and Outreach from the ATLAS experiment' |
| 250 | +my_json['title'] = 'ROOT ntuple format 2015-2016 proton-proton Open Data for Education and Outreach beta release from the ATLAS experiment' |
245 | 251 | # Add a record ID for CERN Open Data. Reserved range for this release |
246 | 252 | my_json['recid'] = '93910' |
247 | 253 | # Add the DOI - these are pre-reserved by the Open Data Portal team |
|
0 commit comments