|
50 | 50 | from templateflow.api import templates as _get_template_list
|
51 | 51 | from ..utils.bids import _init_layout, relative_to_root
|
52 | 52 | from ..utils.images import set_consumables, unsafe_write_nifti_header_and_data
|
53 |
| -from ..utils.misc import splitext as _splitext, _copy_any, unlink |
| 53 | +from ..utils.misc import _copy_any, unlink |
54 | 54 |
|
55 | 55 | regz = re.compile(r"\.gz$")
|
56 | 56 | _pybids_spec = loads(Path(_pkgres("niworkflows", "data/nipreps.json")).read_text())
|
@@ -710,21 +710,6 @@ def _run_interface(self, runtime):
|
710 | 710 | }
|
711 | 711 | )
|
712 | 712 | if self._metadata:
|
713 |
| - out_file = Path(self._results["out_file"][0]) |
714 |
| - # 1.3.x hack |
715 |
| - # For dtseries, we have been generating weird non-BIDS JSON files. |
716 |
| - # We can safely keep producing them to avoid breaking derivatives, but |
717 |
| - # only the existing keys should keep going into them. |
718 |
| - if out_file.name.endswith(".dtseries.nii"): |
719 |
| - legacy_metadata = {} |
720 |
| - for key in ("grayordinates", "space", "surface", "surface_density", "volume"): |
721 |
| - if key in self._metadata: |
722 |
| - legacy_metadata[key] = self._metadata.pop(key) |
723 |
| - if legacy_metadata: |
724 |
| - sidecar = out_file.parent / f"{_splitext(str(out_file))[0]}.json" |
725 |
| - unlink(sidecar, missing_ok=True) |
726 |
| - sidecar.write_text(dumps(legacy_metadata, sort_keys=True, indent=2)) |
727 |
| - # The future: the extension is the first . and everything after |
728 | 713 | sidecar = out_file.parent / f"{out_file.name.split('.', 1)[0]}.json"
|
729 | 714 | unlink(sidecar, missing_ok=True)
|
730 | 715 | sidecar.write_text(dumps(self._metadata, sort_keys=True, indent=2))
|
|
0 commit comments