Skip to content

Reconsider how chunk sizes should be computed #374

@ieivanov

Description

@ieivanov
INFO:iohub.convert:Found Dataset D:\2026_02_08_mantis_v2_rebuild\argolight_2d_dots_oblique_1 with dimensions (P, T, C, Z, Y, X): (1, 1, 1, 1187, 256, 1600)
WARNING:iohub.convert:Chunk size 1187 on axis 2 adjusted to 1 (dimension 1187).

Chunking of size 1 along z is inefficient. What is the core requirement of aligned chunks for the Dask library, as implemented in _adjust_chunks_for_divisibility? Couldn't this dataset be split into three chunks of size, say, [(512, 256, 1600), (512, 256, 1600), (163, 256, 1600)]

This dataset is available for testing at: /hpc/instruments/cm.mantis/2026_02_08_mantis_v2_rebuild\argolight_bullseye_cal_1

Metadata

Metadata

Assignees

Labels

area: zarrZarr integration, storage backends (zarr-python/tensorstore/zarrs)design discussionNeeds design input before implementationperformanceSpeed and memory usage of the code

Type

Projects

No projects

Relationships

None yet

Development

No branches or pull requests

Issue actions