Skip to content

Commit 2bdaba2

Browse files
Merge pull request #270234 from jonburchel/patch-39
Remove references to memory optimized compute
2 parents 20b7d79 + 92b3858 commit 2bdaba2

File tree

5 files changed

+7
-38
lines changed

5 files changed

+7
-38
lines changed

articles/data-factory/.openpublishing.redirection.data-factory.json

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1090,7 +1090,12 @@
10901090
"source_path_from_root": "/articles/data-factory/connector-troubleshoot-google-adwords.md",
10911091
"redirect_url": "/azure/data-factory/connector-troubleshoot-google-ads",
10921092
"redirect_document_id": false
1093-
}
1093+
},
1094+
{
1095+
"source_path_from_root": "/articles/data-factory/memory-optimized-compute.md",
1096+
"redirect_url": "azure/data-factory/control-flow-execute-data-flow-activity#type-properties",
1097+
"redirect_document_id": false
1098+
}
10941099
]
10951100
}
10961101

articles/data-factory/TOC.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -240,8 +240,6 @@ items:
240240
href: concepts-data-flow-performance-transformations.md
241241
- name: Using data flows in pipelines
242242
href: concepts-data-flow-performance-pipelines.md
243-
- name: Memory optimized compute
244-
href: memory-optimized-compute.md
245243
- name: Integration Runtime performance
246244
href: concepts-integration-runtime-performance.md
247245
- name: Manage data flow canvas

articles/data-factory/concepts-integration-runtime-performance.md

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -19,14 +19,6 @@ For more information how to create an Integration Runtime, see [Integration Runt
1919

2020
The easiest way to get started with data flow integration runtimes is to choose small, medium, or large from the compute size picker. See the mappings to cluster configurations for those sizes below.
2121

22-
## Cluster type
23-
24-
There are two available options for the type of Spark cluster to utilize: general purpose & memory optimized.
25-
26-
**General purpose** clusters are the default selection and will be ideal for most data flow workloads. These tend to be the best balance of performance and cost.
27-
28-
If your data flow has many joins and lookups, you may want to use a **memory optimized** cluster. Memory optimized clusters can store more data in memory and will minimize any out-of-memory errors you may get. Memory optimized have the highest price-point per core, but also tend to result in more successful pipelines. If you experience any out of memory errors when executing data flows, switch to a memory optimized Azure IR configuration.
29-
3022
## Cluster size
3123

3224
Data flows distribute the data processing over different cores in a Spark cluster to perform operations in parallel. A Spark cluster with more cores increases the number of cores in the compute environment. More cores increase the processing power of the data flow. Increasing the size of the cluster is often an easy way to reduce the processing time.

articles/data-factory/control-flow-execute-data-flow-activity.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ Property | Description | Allowed values | Required
7070
dataflow | The reference to the Data Flow being executed | DataFlowReference | Yes
7171
integrationRuntime | The compute environment the data flow runs on. If not specified, the autoresolve Azure integration runtime is used. | IntegrationRuntimeReference | No
7272
compute.coreCount | The number of cores used in the spark cluster. Can only be specified if the autoresolve Azure Integration runtime is used | 8, 16, 32, 48, 80, 144, 272 | No
73-
compute.computeType | The type of compute used in the spark cluster. Can only be specified if the autoresolve Azure Integration runtime is used | "General", "MemoryOptimized" | No
73+
compute.computeType | The type of compute used in the spark cluster. Can only be specified if the autoresolve Azure Integration runtime is used | "General" | No
7474
staging.linkedService | If you're using an Azure Synapse Analytics source or sink, specify the storage account used for PolyBase staging.<br/><br/>If your Azure Storage is configured with VNet service endpoint, you must use managed identity authentication with "allow trusted Microsoft service" enabled on storage account, refer to [Impact of using VNet Service Endpoints with Azure storage](/azure/azure-sql/database/vnet-service-endpoint-rule-overview#impact-of-using-virtual-network-service-endpoints-with-azure-storage). Also learn the needed configurations for [Azure Blob](connector-azure-blob-storage.md#managed-identity) and [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md#managed-identity) respectively.<br/> | LinkedServiceReference | Only if the data flow reads or writes to an Azure Synapse Analytics
7575
staging.folderPath | If you're using an Azure Synapse Analytics source or sink, the folder path in blob storage account used for PolyBase staging | String | Only if the data flow reads or writes to Azure Synapse Analytics
7676
traceLevel | Set logging level of your data flow activity execution | Fine, Coarse, None | No

articles/data-factory/memory-optimized-compute.md

Lines changed: 0 additions & 26 deletions
This file was deleted.

0 commit comments

Comments
 (0)