1
- Importre logging
1
+ import logging
2
2
import math
3
3
from functools import partial , reduce
4
4
from itertools import product
@@ -456,12 +456,12 @@ def __contains__(self, value):
456
456
return False
457
457
458
458
# 'cf_contains' has its own calls to 'cfdm_to_memory', so
459
- # we can set '_force_in_memory =False'.
460
- value = value .to_dask_array (_force_in_memory = False )
459
+ # we can set '_force_to_memory =False'.
460
+ value = value .to_dask_array (_force_to_memory = False )
461
461
462
462
# 'cf_contains' has its own calls to 'cfdm_to_memory', so we
463
- # can set '_force_in_memory =False'.
464
- dx = self .to_dask_array (_force_in_memory = False )
463
+ # can set '_force_to_memory =False'.
464
+ dx = self .to_dask_array (_force_to_memory = False )
465
465
466
466
out_ind = tuple (range (dx .ndim ))
467
467
dx_ind = out_ind
@@ -1567,8 +1567,8 @@ def percentile(
1567
1567
axes = tuple (sorted (d ._parse_axes (axes )))
1568
1568
1569
1569
# 'cf_percentile' has its own call to 'cfdm_to_memory', so we
1570
- # can set '_force_in_memory =False'.
1571
- dx = d .to_dask_array (_force_in_memory = False )
1570
+ # can set '_force_to_memory =False'.
1571
+ dx = d .to_dask_array (_force_to_memory = False )
1572
1572
dtype = dx .dtype
1573
1573
shape = dx .shape
1574
1574
@@ -2129,8 +2129,8 @@ def _asdatetime(self, inplace=False):
2129
2129
2130
2130
if not d ._isdatetime ():
2131
2131
# 'cf_rt2dt' has its own call to 'cfdm_to_memory', so we
2132
- # can set '_force_in_memory =False'.
2133
- dx = d .to_dask_array (_force_in_memory = False )
2132
+ # can set '_force_to_memory =False'.
2133
+ dx = d .to_dask_array (_force_to_memory = False )
2134
2134
dx = dx .map_blocks (cf_rt2dt , units = units , dtype = object )
2135
2135
d ._set_dask (dx )
2136
2136
@@ -2186,8 +2186,8 @@ def _asreftime(self, inplace=False):
2186
2186
2187
2187
if d ._isdatetime ():
2188
2188
# 'cf_dt2rt' has its own call to 'cfdm_to_memory', so we
2189
- # can set '_force_in_memory =False'.
2190
- dx = d .to_dask_array (_force_in_memory = False )
2189
+ # can set '_force_to_memory =False'.
2190
+ dx = d .to_dask_array (_force_to_memory = False )
2191
2191
dx = dx .map_blocks (cf_dt2rt , units = units , dtype = float )
2192
2192
d ._set_dask (dx )
2193
2193
@@ -2784,8 +2784,8 @@ def _regrid(
2784
2784
)
2785
2785
2786
2786
# 'regrid' has its own calls to 'cfdm_to_memory', so we can set
2787
- # '_force_in_memory =False'.
2788
- dx = self .to_dask_array (_force_in_memory = False )
2787
+ # '_force_to_memory =False'.
2788
+ dx = self .to_dask_array (_force_to_memory = False )
2789
2789
2790
2790
# Rechunk so that each chunk contains data in the form
2791
2791
# expected by the regrid operator, i.e. the regrid axes all
@@ -3028,10 +3028,10 @@ def concatenate(
3028
3028
copied = not copy # to avoid making two copies in a given case
3029
3029
3030
3030
# Get data as dask arrays and apply concatenation
3031
- # operation. We can set '_force_in_memory =False' because at compute
3031
+ # operation. We can set '_force_to_memory =False' because at compute
3032
3032
# time the concatenation operation does not need to access the
3033
3033
# actual data.
3034
- dxs = [d .to_dask_array (_force_in_memory = False ) for d in processed_data ]
3034
+ dxs = [d .to_dask_array (_force_to_memory = False ) for d in processed_data ]
3035
3035
dx = da .concatenate (dxs , axis = axis )
3036
3036
3037
3037
# Set the CFA write status
@@ -3066,7 +3066,7 @@ def concatenate(
3066
3066
# If and only if any two input Data objects have
3067
3067
# different __in_memory__ values, then set
3068
3068
# in_memory=False on the concatenation.
3069
- in_memory = False
3069
+ in_memory = False
3070
3070
break
3071
3071
3072
3072
# Set the new dask array
@@ -3464,8 +3464,8 @@ def Units(self, value):
3464
3464
cf_func = partial (cf_units , from_units = old_units , to_units = value )
3465
3465
3466
3466
# 'cf_units' has its own call to 'cfdm_to_memory', so we
3467
- # can set '_force_in_memory =False'.
3468
- dx = self .to_dask_array (_force_in_memory = False )
3467
+ # can set '_force_to_memory =False'.
3468
+ dx = self .to_dask_array (_force_to_memory = False )
3469
3469
dx = dx .map_blocks (cf_func , dtype = dtype )
3470
3470
3471
3471
# Setting equivalent units doesn't affect the CFA write
@@ -3508,8 +3508,8 @@ def is_masked(self):
3508
3508
3509
3509
"""
3510
3510
# 'cf_is_masked' has its own call to 'cfdm_to_memory', so we
3511
- # can set '_force_in_memory =False'.
3512
- dx = self .to_dask_array (_force_in_memory = False )
3511
+ # can set '_force_to_memory =False'.
3512
+ dx = self .to_dask_array (_force_to_memory = False )
3513
3513
3514
3514
out_ind = tuple (range (dx .ndim ))
3515
3515
dx_ind = out_ind
@@ -4215,8 +4215,8 @@ def convert_reference_time(
4215
4215
d .Units = units0
4216
4216
4217
4217
# 'cf_rt2dt' its own call to 'cfdm_to_memory', so we can set
4218
- # '_force_in_memory =False'.
4219
- dx = d .to_dask_array (_force_in_memory = False )
4218
+ # '_force_to_memory =False'.
4219
+ dx = d .to_dask_array (_force_to_memory = False )
4220
4220
4221
4221
# Convert to the correct date-time objects
4222
4222
dx = dx .map_blocks (cf_rt2dt , units = units0 , dtype = object )
@@ -4283,10 +4283,10 @@ def get_deterministic_name(self):
4283
4283
units = self ._Units
4284
4284
4285
4285
# The dask graph is never going to be computed, so we can set
4286
- # '_force_in_memory =False'.
4286
+ # '_force_to_memory =False'.
4287
4287
return tokenize (
4288
4288
self .to_dask_array (
4289
- _apply_mask_hardness = False , _force_in_memory = False
4289
+ _force_mask_hardness = False , _force_to_memory = False
4290
4290
).name ,
4291
4291
units .formatted (definition = True , names = True ),
4292
4292
units ._canonical_calendar ,
@@ -4324,8 +4324,8 @@ def add_file_location(self, location):
4324
4324
updated = False
4325
4325
4326
4326
# The dask graph is never going to be computed, so we can set
4327
- # '_force_in_memory =False'.
4328
- dsk = self .todict (_force_in_memory = False )
4327
+ # '_force_to_memory =False'.
4328
+ dsk = self .todict (_force_to_memory = False )
4329
4329
for key , a in dsk .items ():
4330
4330
try :
4331
4331
dsk [key ] = a .add_file_location (location )
@@ -4338,7 +4338,7 @@ def add_file_location(self, location):
4338
4338
updated = True
4339
4339
4340
4340
if updated :
4341
- dx = self .to_dask_array (_force_in_memory = False )
4341
+ dx = self .to_dask_array (_force_to_memory = False )
4342
4342
dx = da .Array (dsk , dx .name , dx .chunks , dx .dtype , dx ._meta )
4343
4343
self ._set_dask (dx , clear = self ._NONE , in_memory = None )
4344
4344
@@ -5725,8 +5725,8 @@ def unique(self, split_every=None):
5725
5725
d .soften_mask ()
5726
5726
5727
5727
# The applicable chunk function will have its own call to
5728
- # 'cfdm_to_memory', so we can set '_force_in_memory =False'.
5729
- dx = d .to_dask_array (_force_in_memory = False )
5728
+ # 'cfdm_to_memory', so we can set '_force_to_memory =False'.
5729
+ dx = d .to_dask_array (_force_to_memory = False )
5730
5730
dx = Collapse ().unique (dx , split_every = split_every )
5731
5731
5732
5732
d ._set_dask (dx )
@@ -6240,8 +6240,8 @@ def file_locations(self):
6240
6240
out = set ()
6241
6241
6242
6242
# The dask graph is never going to be computed, so we can set
6243
- # '_force_in_memory =False'.
6244
- for key , a in self .todict (_force_in_memory = False ).items ():
6243
+ # '_force_to_memory =False'.
6244
+ for key , a in self .todict (_force_to_memory = False ).items ():
6245
6245
try :
6246
6246
out .update (a .file_locations ())
6247
6247
except AttributeError :
@@ -6801,8 +6801,8 @@ def del_file_location(self, location):
6801
6801
updated = False
6802
6802
6803
6803
# The dask graph is never going to be computed, so we can set
6804
- # '_force_in_memory =False'.
6805
- dsk = self .todict (_force_in_memory = False )
6804
+ # '_force_to_memory =False'.
6805
+ dsk = self .todict (_force_to_memory = False )
6806
6806
for key , a in dsk .items ():
6807
6807
try :
6808
6808
dsk [key ] = a .del_file_location (location )
@@ -6815,7 +6815,7 @@ def del_file_location(self, location):
6815
6815
updated = True
6816
6816
6817
6817
if updated :
6818
- dx = self .to_dask_array (_force_in_memory = False )
6818
+ dx = self .to_dask_array (_force_to_memory = False )
6819
6819
dx = da .Array (dsk , dx .name , dx .chunks , dx .dtype , dx ._meta )
6820
6820
self ._set_dask (dx , clear = self ._NONE , in_memory = None )
6821
6821
@@ -7852,8 +7852,8 @@ def where(
7852
7852
# hardness has been applied.
7853
7853
#
7854
7854
# 'cf_where' has its own calls to 'cfdm_to_memory', so we can
7855
- # set '_force_in_memory =False'.
7856
- dx = d .to_dask_array (_force_in_memory = False )
7855
+ # set '_force_to_memory =False'.
7856
+ dx = d .to_dask_array (_force_to_memory = False )
7857
7857
7858
7858
units = d .Units
7859
7859
@@ -7869,8 +7869,8 @@ def where(
7869
7869
condition = type (self ).asdata (condition )
7870
7870
condition = where_broadcastable (d , condition , "condition" )
7871
7871
# 'cf_where' has its own calls to 'cfdm_to_memory', so we can
7872
- # set '_force_in_memory =False'.
7873
- condition = condition .to_dask_array (_force_in_memory = False )
7872
+ # set '_force_to_memory =False'.
7873
+ condition = condition .to_dask_array (_force_to_memory = False )
7874
7874
7875
7875
# If x or y is self then change it to None. This prevents an
7876
7876
# unnecessary copy; and, at compute time, an unncessary numpy
0 commit comments