Skip to content

Commit f0a8e28

Browse files
authored
Merge pull request #252939 from ekote/unsupported24
Spark 2.4 EOL - runtime is not supported as of Sep 29
2 parents cb4edeb + 0f8cede commit f0a8e28

14 files changed

+69
-61
lines changed

articles/synapse-analytics/machine-learning/setup-environment-cognitive-services.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ To get started on Azure Kubernetes Service, follow these steps:
9292

9393
1. [Deploy an Azure Kubernetes Service (AKS) cluster using the Azure portal](../../aks/learn/quick-kubernetes-deploy-portal.md)
9494

95-
1. [Install the Apache Spark 2.4.0 helm chart](https://hub.helm.sh/charts/microsoft/spark)
95+
1. [Install the Apache Spark 2.4.0 helm chart](https://hub.helm.sh/charts/microsoft/spark) - warning: [Spark 2.4](../spark/apache-spark-24-runtime.md) is retired and out of the support.
9696

9797
1. [Install an Azure AI container using Helm](../../ai-services/computer-vision/deploy-computer-vision-on-premises.md)
9898

articles/synapse-analytics/overview-faq.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -138,7 +138,7 @@ sections:
138138
- question: |
139139
What versions of Spark are available?
140140
answer: |
141-
As of May 2021, Azure Synapse Apache Spark fully supports Spark 2.4 and Spark 3.1. As of April 2022, Spark 3.2 is in preview. For a full list of core components and currently supported versions see [Apache Spark version support](./spark/apache-spark-version-support.md).
141+
As of September 2023, Azure Synapse Apache Spark fully supports Spark 3.3. For a full list of core components and currently supported versions see [Apache Spark version support](./spark/apache-spark-version-support.md).
142142
143143
- question: |
144144
Is there an equivalent of DButils in Azure Synapse Spark?

articles/synapse-analytics/spark/apache-spark-24-runtime.md

Lines changed: 11 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: Azure Synapse Runtime for Apache Spark 2.4 (EOLA)
3-
description: Supported versions of Spark, Scala, Python, and .NET for Apache Spark 2.4.
4-
author: eskot
2+
title: Azure Synapse Runtime for Apache Spark 2.4 (unsupported)
3+
description: Versions of Spark, Scala, Python, and .NET for Apache Spark 2.4.
4+
author: ekote
55
ms.service: synapse-analytics
66
ms.topic: reference
77
ms.subservice: spark
@@ -10,14 +10,17 @@ ms.author: eskot
1010
ms.custom: has-adal-ref, devx-track-dotnet, devx-track-extended-java, devx-track-python
1111
---
1212

13-
# Azure Synapse Runtime for Apache Spark 2.4 (EOLA)
13+
# Azure Synapse Runtime for Apache Spark 2.4 (unsupported)
1414

1515
Azure Synapse Analytics supports multiple runtimes for Apache Spark. This document will cover the runtime components and versions for the Azure Synapse Runtime for Apache Spark 2.4.
1616

17-
> [!IMPORTANT]
18-
> * End of life announced (EOLA) for Azure Synapse Runtime for Apache Spark 2.4 has been announced July 29, 2022.
19-
> * In accordance with the Synapse runtime for Apache Spark lifecycle policy, Azure Synapse runtime for Apache Spark 2.4 will be retired and disabled as of September 29, 2023. After the EOL date, the retired runtimes are unavailable for new Spark pools and existing workflows can't execute. Metadata will temporarily remain in the Synapse workspace.
20-
> * We recommend that you upgrade your Apache Spark 2.4 workloads to version 3.3 at your earliest convenience.
17+
> [!WARNING]
18+
> End of Support Notification for Azure Synapse Runtime for Apache Spark 2.4
19+
> * Effective September 29, 2023, the Azure Synapse will discontinue official support for Spark 2.4 Runtimes.
20+
> * Post September 29, we will not be addressing any support tickets related to Spark 2.4. There will be no release pipeline in place for bug or security fixes for Spark 2.4. Utilizing Spark 2.4 post the support cutoff date is undertaken at one's own risk. We strongly discourage its continued use due to potential security and functionality concerns.
21+
> * Recognizing that certain customers may need additional time to transition to a higher runtime version, we are temporarily extending the usage option for Spark 2.4, but we will not provide any official support for it.
22+
> * We strongly advise to proactively upgrade their workloads to a more recent version of the runtime (e.g., [Azure Synapse Runtime for Apache Spark 3.3 (GA)](./apache-spark-33-runtime.md)).
23+
2124

2225
## Component versions
2326
| Component | Version |

articles/synapse-analytics/spark/apache-spark-external-metastore.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Azure Synapse Analytics allows Apache Spark pools in the same workspace to share
1616

1717
## Supported Hive Metastore versions
1818

19-
The feature works with both Spark 2.4 and Spark 3.1. The following table shows the supported Hive Metastore versions for each Spark version.
19+
The feature works with Spark 3.1. The following table shows the supported Hive Metastore versions for each Spark version.
2020

2121
|Spark Version|HMS 0.13.X|HMS 1.2.X|HMS 2.1.X|HMS 2.3.x|HMS 3.1.X|
2222
|--|--|--|--|--|--|
@@ -161,7 +161,7 @@ If the underlying data of your Hive tables are stored in Azure Blob storage acco
161161
3. Provide **Name** of the linked service. Record the name of the linked service, this info will be used in Spark configuration shortly.
162162
4. Select the Azure Blob Storage account. Make sure Authentication method is **Account key**. Currently Spark pool can only access Blob Storage account via account key.
163163
5. **Test connection** and click **Create**.
164-
6. After creating the linked service to Blob Storage account, when you run Spark queries, make sure you run below Spark code in the notebook to get access to the the Blob Storage account for the Spark session. Learn more about why you need to do this [here](./apache-spark-secure-credentials-with-tokenlibrary.md).
164+
6. After creating the linked service to Blob Storage account, when you run Spark queries, make sure you run below Spark code in the notebook to get access to the Blob Storage account for the Spark session. Learn more about why you need to do this [here](./apache-spark-secure-credentials-with-tokenlibrary.md).
165165

166166
```python
167167
%%pyspark
@@ -190,7 +190,7 @@ After setting up storage connections, you can query the existing tables in the H
190190
No credentials found for account xxxxx.blob.core.windows.net in the configuration, and its container xxxxx is not accessible using anonymous credentials. Please check if the container exists first. If it is not publicly available, you have to provide account credentials.
191191
```
192192

193-
When use key authentication to your storage account via linked service, you need to take an extra step to get the token for Spark session. Run below code to configure your Spark session before running the query. Learn more about why you need to do this here.
193+
When using key authentication to your storage account via linked service, you need to take an extra step to get the token for Spark session. Run below code to configure your Spark session before running the query. Learn more about why you need to do this here.
194194

195195
```python
196196
%%pyspark
@@ -254,4 +254,4 @@ You can easily fix this issue by appending `/usr/hdp/current/hadoop-client/*` to
254254
```text
255255
Eg:
256256
spark.sql.hive.metastore.jars":"/opt/hive-metastore/lib-2.3/*:/usr/hdp/current/hadoop-client/lib/*:/usr/hdp/current/hadoop-client/*
257-
```
257+
```

articles/synapse-analytics/spark/apache-spark-intelligent-cache-concept.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@ You won't see the benefit of this feature if:
9393

9494
* Your workload requires large amounts of shuffle, then disabling the Intelligent Cache will free up available space to prevent your job from failing due to insufficient storage space.
9595

96-
* You're using a Spark 2.4 pool, you'll need to upgrade your pool to the latest version of Spark.
96+
* You're using a Spark 3.1 pool, you'll need to upgrade your pool to the latest version of Spark.
9797

9898

9999
## Learn more

articles/synapse-analytics/spark/apache-spark-performance-hyperspace.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ This document is also available in notebook form, for [Python](https://github.co
3131
## Setup
3232

3333
>[!Note]
34-
> Hyperspace is supported in Azure Synapse Runtime for Apache Spark 2.4 (EOLA), Azure Synapse Runtime for Apache Spark 3.1 (EOLA), and Azure Synapse Runtime for Apache Spark 3.2 (EOLA). However, it should be noted that Hyperspace is not supported in Azure Synapse Runtime for Apache Spark 3.3 (GA).
34+
> Hyperspace is supported in Azure Synapse Runtime for Apache Spark 3.1 (EOLA), and Azure Synapse Runtime for Apache Spark 3.2 (EOLA). However, it should be noted that Hyperspace is not supported in Azure Synapse Runtime for Apache Spark 3.3 (GA).
3535
3636
To begin with, start a new Spark session. Since this document is a tutorial merely to illustrate what Hyperspace can offer, you will make a configuration change that allows us to highlight what Hyperspace is doing on small datasets.
3737

articles/synapse-analytics/spark/apache-spark-version-support.md

Lines changed: 30 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: Apache Spark version support
33
description: Supported versions of Spark, Scala, Python, .NET
4-
author: eskot
4+
author: ekote
55
ms.service: synapse-analytics
66
ms.topic: reference
77
ms.subservice: spark
@@ -13,14 +13,39 @@ ms.reviewer: eskot
1313

1414
# Azure Synapse runtimes
1515

16-
Apache Spark pools in Azure Synapse use runtimes to tie together essential component versions such as Azure Synapse optimizations, packages, and connectors with a specific Apache Spark version. Each runtime will be upgraded periodically to include new improvements, features, and patches.
17-
18-
When you create a serverless Apache Spark pool, you will have the option to select the corresponding Apache Spark version. Based on this, the pool will come pre-installed with the associated runtime components and packages. The runtimes have the following advantages:
19-
16+
Apache Spark pools in Azure Synapse use runtimes to tie together essential component versions such as Azure Synapse optimizations, packages, and connectors with a specific Apache Spark version. Each runtime will be upgraded periodically to include new improvements, features, and patches. When you create a serverless Apache Spark pool, you will have the option to select the corresponding Apache Spark version. Based on this, the pool will come pre-installed with the associated runtime components and packages. The runtimes have the following advantages:
2017
- Faster session startup times
2118
- Tested compatibility with specific Apache Spark versions
2219
- Access to popular, compatible connectors and open-source packages
2320

21+
22+
## Supported Azure Synapse runtime releases
23+
24+
> [!WARNING]
25+
> End of Support Notification for Azure Synapse Runtime for Apache Spark 2.4
26+
> * Effective September 29, 2023, the Azure Synapse will discontinue official support for Spark 2.4 Runtimes.
27+
> * Post September 29, we will not be addressing any support tickets related to Spark 2.4. There will be no release pipeline in place for bug or security fixes for Spark 2.4. Utilizing Spark 2.4 post the support cutoff date is undertaken at one's own risk. We strongly discourage its continued use due to potential security and functionality concerns.
28+
> * Recognizing that certain customers may need additional time to transition to a higher runtime version, we are temporarily extending the usage option for Spark 2.4, but we will not provide any official support for it.
29+
> * We strongly advise to proactively upgrade their workloads to a more recent version of the runtime (e.g., [Azure Synapse Runtime for Apache Spark 3.3 (GA)](./apache-spark-33-runtime.md)).
30+
31+
The following table lists the runtime name, Apache Spark version, and release date for supported Azure Synapse Runtime releases.
32+
33+
| Runtime name | Release date | Release stage | End of life announcement date | End of life effective date |
34+
|----------------------------------------------------------------------------|-------------------|---------------------------------|-------------------------------|----------------------------|
35+
| [Azure Synapse Runtime for Apache Spark 3.3](./apache-spark-33-runtime.md) | Nov 17, 2022 | GA (as of Feb 23, 2023) | Nov 17, 2023 | Nov 17, 2024 |
36+
| [Azure Synapse Runtime for Apache Spark 3.2](./apache-spark-32-runtime.md) | July 8, 2022 | __End of Life Announced (EOLA)__ | July 8, 2023 | July 8, 2024 |
37+
| [Azure Synapse Runtime for Apache Spark 3.1](./apache-spark-3-runtime.md) | May 26, 2021 | __End of Life Announced (EOLA)__ | January 26, 2023 | January 26, 2024 |
38+
| [Azure Synapse Runtime for Apache Spark 2.4](./apache-spark-24-runtime.md) | December 15, 2020 | __End of Life (EOL)__ | __July 29, 2022__ | __September 29, 2023__ |
39+
40+
## Runtime release stages
41+
42+
For the complete runtime for Apache Spark lifecycle and support policies, refer to [Synapse runtime for Apache Spark lifecycle and supportability](./runtime-for-apache-spark-lifecycle-and-supportability.md).
43+
44+
## Runtime patching
45+
46+
Azure Synapse runtime for Apache Spark patches are rolled out monthly containing bug, feature and security fixes to the Apache Spark core engine, language environments, connectors and libraries.
47+
48+
2449
> [!NOTE]
2550
> - Maintenance updates will be automatically applied to new sessions for a given serverless Apache Spark pool.
2651
> - You should test and validate that your applications run properly when using new runtime versions.
@@ -41,25 +66,6 @@ When you create a serverless Apache Spark pool, you will have the option to sele
4166
> * ```org/apache/log4j/chainsaw/*```
4267
>
4368
> While the above classes were not used in the default Log4j configurations in Synapse, it is possible that some user application could still depend on it. If your application needs to use these classes, use Library Management to add a secure version of Log4j to the Spark Pool. __Do not use Log4j version 1.2.17__, as it would be reintroducing the vulnerabilities.
44-
>
45-
46-
## Supported Azure Synapse runtime releases
47-
The following table lists the runtime name, Apache Spark version, and release date for supported Azure Synapse Runtime releases.
48-
49-
| Runtime name | Release date | Release stage | End of life announcement date | End of life effective date |
50-
|----------------------------------------------------------------------------|-------------------|----------------------------------|-------------------------------|----------------------------|
51-
| [Azure Synapse Runtime for Apache Spark 3.3](./apache-spark-33-runtime.md) | Nov 17, 2022 | GA (as of Feb 23, 2023) | Nov 17, 2023 | Nov 17, 2024 |
52-
| [Azure Synapse Runtime for Apache Spark 3.2](./apache-spark-32-runtime.md) | July 8, 2022 | __End of Life Announced (EOLA)__ | July 8, 2023 | July 8, 2024 |
53-
| [Azure Synapse Runtime for Apache Spark 3.1](./apache-spark-3-runtime.md) | May 26, 2021 | __End of Life Announced (EOLA)__ | January 26, 2023 | January 26, 2024 |
54-
| [Azure Synapse Runtime for Apache Spark 2.4](./apache-spark-24-runtime.md) | December 15, 2020 | __End of Life Announced (EOLA)__ | __July 29, 2022__ | __September 29, 2023__ |
55-
56-
## Runtime release stages
57-
58-
For the complete runtime for Apache Spark lifecycle and support policies, refer to [Synapse runtime for Apache Spark lifecycle and supportability](./runtime-for-apache-spark-lifecycle-and-supportability.md).
59-
60-
## Runtime patching
61-
62-
Azure Synapse runtime for Apache Spark patches are rolled out monthly containing bug, feature and security fixes to the Apache Spark core engine, language environments, connectors and libraries.
6369
6470
The patch policy differs based on the [runtime lifecycle stage](./runtime-for-apache-spark-lifecycle-and-supportability.md):
6571
1. Generally Available (GA) runtime: Receive no upgrades on major versions (i.e. 3.x -> 4.x). And will upgrade a minor version (i.e. 3.x -> 3.y) as long as there are no deprecation or regression impacts.

articles/synapse-analytics/spark/data-sources/apache-spark-cdm-connector.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ For information on defining Common Data Model documents by using Common Data Mod
2020

2121
At a high level, the connector supports:
2222

23-
* Spark 2.4, 3.1, and 3.2.
23+
* 3.1, and 3.2., and 3.3.
2424
* Reading data from an entity in a Common Data Model folder into a Spark DataFrame.
2525
* Writing from a Spark DataFrame to an entity in a Common Data Model folder based on a Common Data Model entity definition.
2626
* Writing from a Spark DataFrame to an entity in a Common Data Model folder based on the DataFrame schema.

articles/synapse-analytics/spark/data-sources/apache-spark-kusto-connector.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -12,20 +12,19 @@ author: midesa
1212
---
1313

1414
# Azure Data Explorer (Kusto) connector for Apache Spark
15-
The Azure Data Explorer (Kusto) connector for Apache Spark is designed to efficiently transfer data between Kusto clusters and Spark. This connector is available in Python, Java, and .NET. It is built in to the Azure Synapse Apache Spark 2.4 runtime (EOLA).
15+
The Azure Data Explorer (Kusto) connector for Apache Spark is designed to efficiently transfer data between Kusto clusters and Spark. This connector is available in Python, Java, and .NET.
1616

1717
## Authentication
1818
When using Azure Synapse Notebooks or Apache Spark job definitions, the authentication between systems is made seamless with the linked service. The Token Service connects with Azure Active Directory to obtain security tokens for use when accessing the Kusto cluster.
1919

20-
For Azure Synapse Pipelines, the authentication will use the service principal name. Currently, managed identities are not supported with the Azure Data Explorer connector.
20+
For Azure Synapse Pipelines, the authentication uses the service principal name. Currently, managed identities aren't supported with the Azure Data Explorer connector.
2121

2222
## Prerequisites
23-
- [Connect to Azure Data Explorer](../../quickstart-connect-azure-data-explorer.md): You will need to set up a Linked Service to connect to an existing Kusto cluster.
23+
- [Connect to Azure Data Explorer](../../quickstart-connect-azure-data-explorer.md): You need to set up a Linked Service to connect to an existing Kusto cluster.
2424

2525
## Limitations
26-
- The Azure Data Explorer (Kusto) connector is currently only supported on the Azure Synapse Apache Spark 2.4 runtime (EOLA).
2726
- The Azure Data Explorer linked service can only be configured with the Service Principal Name.
28-
- Within Azure Synapse Notebooks or Apache Spark Job Definitions, the Azure Data Explorer connector will use Azure AD pass-through to connect to the Kusto Cluster.
27+
- Within Azure Synapse Notebooks or Apache Spark Job Definitions, the Azure Data Explorer connector uses Azure AD pass-through to connect to the Kusto Cluster.
2928

3029

3130
## Use the Azure Data Explorer (Kusto) connector

articles/synapse-analytics/spark/low-shuffle-merge-for-apache-spark.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -29,9 +29,9 @@ It's available on Synapse Pools for Apache Spark versions 3.2 and 3.3.
2929

3030
|Version| Availability | Default |
3131
|--|--|--|
32-
| Delta 0.6 / Spark 2.4 | No | - |
33-
| Delta 1.2 / Spark 3.2 | Yes | false |
34-
| Delta 2.2 / Spark 3.3 | Yes | true |
32+
| Delta 0.6 / [Spark 2.4](./apache-spark-24-runtime.md) | No | - |
33+
| Delta 1.2 / [Spark 3.2](./apache-spark-32-runtime.md) | Yes | false |
34+
| Delta 2.2 / [Spark 3.3](./apache-spark-33-runtime.md) | Yes | true |
3535

3636

3737
## Benefits of Low Shuffle Merge

0 commit comments

Comments
 (0)