6
6
from urllib .parse import urlparse
7
7
8
8
import cfdm
9
- from cfdm .read_write .exceptions import FileTypeError
9
+ from cfdm .read_write .exceptions import DatasetTypeError
10
10
from cfdm .read_write .netcdf import NetCDFRead
11
11
12
12
from ..aggregate import aggregate as cf_aggregate
@@ -384,7 +384,7 @@ class read(cfdm.read):
384
384
385
385
.. versionadded:: NEXTVERSION
386
386
387
- {{read store_hdf5_chunks : `bool`, optional}}
387
+ {{read store_dataset_chunks : `bool`, optional}}
388
388
389
389
.. versionadded:: NEXTVERSION
390
390
@@ -502,7 +502,7 @@ def __new__(
502
502
unpack = True ,
503
503
warn_valid = False ,
504
504
dask_chunks = "storage-aligned" ,
505
- store_hdf5_chunks = True ,
505
+ store_dataset_chunks = True ,
506
506
domain = False ,
507
507
cfa = None ,
508
508
cfa_write = None ,
@@ -728,7 +728,7 @@ def __new__(
728
728
storage_options = storage_options ,
729
729
netcdf_backend = netcdf_backend ,
730
730
dask_chunks = dask_chunks ,
731
- store_hdf5_chunks = store_hdf5_chunks ,
731
+ store_dataset_chunks = store_dataset_chunks ,
732
732
cache = cache ,
733
733
cfa = cfa ,
734
734
cfa_write = cfa_write ,
@@ -737,7 +737,7 @@ def __new__(
737
737
unsqueeze = unsqueeze ,
738
738
file_type = file_type ,
739
739
)
740
- except FileTypeError as error :
740
+ except DatasetTypeError as error :
741
741
if file_type is None :
742
742
file_format_errors .append (error )
743
743
else :
@@ -764,7 +764,7 @@ def __new__(
764
764
domain = domain ,
765
765
file_type = file_type ,
766
766
)
767
- except FileTypeError as error :
767
+ except DatasetTypeError as error :
768
768
if file_type is None :
769
769
file_format_errors .append (error )
770
770
else :
@@ -773,7 +773,7 @@ def __new__(
773
773
774
774
if file_format_errors :
775
775
error = "\n " .join (map (str , file_format_errors ))
776
- raise FileTypeError (f"\n { error } " )
776
+ raise DatasetTypeError (f"\n { error } " )
777
777
778
778
if domain :
779
779
file_contents = DomainList (file_contents )
0 commit comments