Skip to content

Commit 0b8343c

Browse files
PProfizigithub-actions[bot]
authored andcommitted
update generated code
1 parent cddb7e7 commit 0b8343c

File tree

6 files changed

+26
-18
lines changed

6 files changed

+26
-18
lines changed

doc/source/_static/dpf_operators.html

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

src/ansys/dpf/core/operators/result/migrate_to_h5dpf.py

Lines changed: 25 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -18,31 +18,33 @@
1818
class migrate_to_h5dpf(Operator):
1919
r"""Read mesh properties from the results files contained in the streams or
2020
data sources and make those properties available through a mesh
21-
selection manager in output.
21+
selection manager in output.User can input a GenericDataContainer that
22+
will map an item to a result name. Example of Map: {{ default: wf1},
23+
{EUL: wf2}, {ENG_SE: wf3}}.
2224
2325
2426
Parameters
2527
----------
26-
dataset_size_compression_threshold: int, optional
28+
dataset_size_compression_threshold: int or GenericDataContainer, optional
2729
Integer value that defines the minimum dataset size (in bytes) to use h5 native compression Applicable for arrays of floats, doubles and integers.
28-
h5_native_compression: int or DataTree, optional
30+
h5_native_compression: int or DataTree or GenericDataContainer, optional
2931
Integer value / DataTree that defines the h5 native compression used For Integer Input {0: No Compression (default); 1-9: GZIP Compression : 9 provides maximum compression but at the slowest speed.}For DataTree Input {type: None / GZIP / ZSTD; level: GZIP (1-9) / ZSTD (1-20); num_threads: ZSTD (>0)}
30-
export_floats: bool, optional
32+
export_floats: bool or GenericDataContainer, optional
3133
Converts double to float to reduce file size (default is true).If False, nodal results are exported as double precision and elemental results as single precision.
3234
filename: str
3335
filename of the migrated file
3436
comma_separated_list_of_results: str, optional
3537
list of results (source operator names) separated by semicolons that will be stored. (Example: U;S;EPEL). If empty, all available results will be converted.
3638
all_time_sets: bool, optional
37-
default is false
39+
Deprecated. Please use filtering workflows instead to select time scoping. Default is false.
3840
streams_container: StreamsContainer, optional
3941
streams (result file container) (optional)
4042
data_sources: DataSources, optional
4143
if the stream is null then we need to get the file path from the data sources
4244
compression_workflow: Workflow or GenericDataContainer, optional
43-
BETA Option: Applies input compression workflow. User can input a GenericDataContainer that will map a compression workflow to a result name. Example of Map: {{ default: wf1}, {EUL: wf2}, {ENG_SE: wf3}}
45+
BETA Option: Applies input compression workflow.
4446
filtering_workflow: Workflow or GenericDataContainer, optional
45-
Applies input filtering workflow. User can input a GenericDataContainer of the format described for Pin(6) that will map a filtering workflow to a result name.
47+
Applies input filtering workflow.
4648
4749
Returns
4850
-------
@@ -142,26 +144,32 @@ def __init__(
142144
def _spec() -> Specification:
143145
description = r"""Read mesh properties from the results files contained in the streams or
144146
data sources and make those properties available through a mesh
145-
selection manager in output.
147+
selection manager in output.User can input a GenericDataContainer that
148+
will map an item to a result name. Example of Map: {{ default: wf1},
149+
{EUL: wf2}, {ENG_SE: wf3}}.
146150
"""
147151
spec = Specification(
148152
description=description,
149153
map_input_pin_spec={
150154
-5: PinSpecification(
151155
name="dataset_size_compression_threshold",
152-
type_names=["int32"],
156+
type_names=["int32", "generic_data_container"],
153157
optional=True,
154158
document=r"""Integer value that defines the minimum dataset size (in bytes) to use h5 native compression Applicable for arrays of floats, doubles and integers.""",
155159
),
156160
-2: PinSpecification(
157161
name="h5_native_compression",
158-
type_names=["int32", "abstract_data_tree"],
162+
type_names=[
163+
"int32",
164+
"abstract_data_tree",
165+
"generic_data_container",
166+
],
159167
optional=True,
160168
document=r"""Integer value / DataTree that defines the h5 native compression used For Integer Input {0: No Compression (default); 1-9: GZIP Compression : 9 provides maximum compression but at the slowest speed.}For DataTree Input {type: None / GZIP / ZSTD; level: GZIP (1-9) / ZSTD (1-20); num_threads: ZSTD (>0)}""",
161169
),
162170
-1: PinSpecification(
163171
name="export_floats",
164-
type_names=["bool"],
172+
type_names=["bool", "generic_data_container"],
165173
optional=True,
166174
document=r"""Converts double to float to reduce file size (default is true).If False, nodal results are exported as double precision and elemental results as single precision.""",
167175
),
@@ -181,7 +189,7 @@ def _spec() -> Specification:
181189
name="all_time_sets",
182190
type_names=["bool"],
183191
optional=True,
184-
document=r"""default is false""",
192+
document=r"""Deprecated. Please use filtering workflows instead to select time scoping. Default is false.""",
185193
),
186194
3: PinSpecification(
187195
name="streams_container",
@@ -199,13 +207,13 @@ def _spec() -> Specification:
199207
name="compression_workflow",
200208
type_names=["workflow", "generic_data_container"],
201209
optional=True,
202-
document=r"""BETA Option: Applies input compression workflow. User can input a GenericDataContainer that will map a compression workflow to a result name. Example of Map: {{ default: wf1}, {EUL: wf2}, {ENG_SE: wf3}}""",
210+
document=r"""BETA Option: Applies input compression workflow.""",
203211
),
204212
7: PinSpecification(
205213
name="filtering_workflow",
206214
type_names=["workflow", "generic_data_container"],
207215
optional=True,
208-
document=r"""Applies input filtering workflow. User can input a GenericDataContainer of the format described for Pin(6) that will map a filtering workflow to a result name.""",
216+
document=r"""Applies input filtering workflow.""",
209217
),
210218
},
211219
map_output_pin_spec={
@@ -437,7 +445,7 @@ def comma_separated_list_of_results(self) -> Input:
437445
def all_time_sets(self) -> Input:
438446
r"""Allows to connect all_time_sets input to the operator.
439447
440-
default is false
448+
Deprecated. Please use filtering workflows instead to select time scoping. Default is false.
441449
442450
Returns
443451
-------
@@ -500,7 +508,7 @@ def data_sources(self) -> Input:
500508
def compression_workflow(self) -> Input:
501509
r"""Allows to connect compression_workflow input to the operator.
502510
503-
BETA Option: Applies input compression workflow. User can input a GenericDataContainer that will map a compression workflow to a result name. Example of Map: {{ default: wf1}, {EUL: wf2}, {ENG_SE: wf3}}
511+
BETA Option: Applies input compression workflow.
504512
505513
Returns
506514
-------
@@ -521,7 +529,7 @@ def compression_workflow(self) -> Input:
521529
def filtering_workflow(self) -> Input:
522530
r"""Allows to connect filtering_workflow input to the operator.
523531
524-
Applies input filtering workflow. User can input a GenericDataContainer of the format described for Pin(6) that will map a filtering workflow to a result name.
532+
Applies input filtering workflow.
525533
526534
Returns
527535
-------
5.5 KB
Binary file not shown.
0 Bytes
Binary file not shown.
5.05 KB
Binary file not shown.
0 Bytes
Binary file not shown.

0 commit comments

Comments
 (0)