Skip to content

Commit 9049048

Browse files
Merge pull request #50594 from theresa-i/use-apache-spark
Updates
2 parents a6bc856 + 58156e7 commit 9049048

21 files changed

+28
-88
lines changed

learn-pr/wwl/use-apache-spark-work-files-lakehouse/1-introduction.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Introduction
44
metadata:
55
title: Introduction
66
description: "Introduction"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 1
1612
content: |
1713
[!include[](includes/1-introduction.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/2-spark.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Prepare to use Apache Spark
44
metadata:
55
title: Prepare to use Apache Spark
66
description: "Prepare to use Apache Spark"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 3
1612
content: |
1713
[!include[](includes/2-spark.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/3-spark-code.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Run Spark code
44
metadata:
55
title: Run Spark code
66
description: "Run Spark code"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 3
1612
content: |
1713
[!include[](includes/3-spark-code.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/4-dataframe.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Work with data in a Spark dataframe
44
metadata:
55
title: Work with data in a Spark dataframe
66
description: "Work with data in a Spark dataframe"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 5
1612
content: |
1713
[!include[](includes/4-dataframe.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/5-spark-sql.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Work with data using Spark SQL
44
metadata:
55
title: Work with data using Spark SQL
66
description: "Work with data using Spark SQL"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 7
1612
content: |
1713
[!include[](includes/5-spark-sql.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/6-visualize-data.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Visualize data in a Spark notebook
44
metadata:
55
title: Visualize data in a Spark notebook
66
description: "Visualize data in a Spark notebook"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 5
1612
content: |
1713
[!include[](includes/6-visualize-data.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/7-exercise-spark.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Exercise - Analyze data with Apache Spark
44
metadata:
55
title: Exercise - Analyze data with Apache Spark
66
description: "Exercise - Analyze data with Apache Spark"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 45
1612
content: |
1713
[!include[](includes/7-exercise-spark.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/8-knowledge-check.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -6,14 +6,10 @@ metadata:
66
ai_generated_module_assessment: true
77
title: Module assessment
88
description: "Knowledge check"
9-
ms.date: 04/16/2025
10-
author: wwlpublish
9+
ms.date: 05/22/2025
10+
author: theresa-i
1111
ms.author: theresai
1212
ms.topic: unit
13-
ms.custom:
14-
- build-2023
15-
- build-2023-dataai
16-
- build-2023-fabric
1713
durationInMinutes: 3
1814
quiz:
1915
title: "Check your knowledge"

learn-pr/wwl/use-apache-spark-work-files-lakehouse/9-summary.yml

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,14 +4,10 @@ title: Summary
44
metadata:
55
title: Summary
66
description: "Summary"
7-
ms.date: 04/16/2025
8-
author: wwlpublish
7+
ms.date: 05/22/2025
8+
author: theresa-i
99
ms.author: theresai
1010
ms.topic: unit
11-
ms.custom:
12-
- build-2023
13-
- build-2023-dataai
14-
- build-2023-fabric
1511
durationInMinutes: 1
1612
content: |
1713
[!include[](includes/9-summary.md)]

learn-pr/wwl/use-apache-spark-work-files-lakehouse/includes/2-spark.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -22,9 +22,9 @@ Microsoft Fabric provides a *starter pool* in each workspace, enabling Spark job
2222
Additionally, you can create custom Spark pools with specific node configurations that support your particular data processing needs.
2323

2424
> [!NOTE]
25-
> The ability to customize Spark pool settings can be disabled by Fabric administrators at the Fabric Capacity level. For more information, see **[Capacity administration settings for Data Engineering and Data Science](/fabric/data-engineering/capacity-settings-overview)** in the Fabric documentation.
25+
> The ability to customize Spark pool settings can be disabled by Fabric administrators at the Fabric Capacity level. For more information, see **[Capacity administration settings for Data Engineering and Data Science](/fabric/data-engineering/capacity-settings-overview?azure-portal=true)** in the Fabric documentation.
2626
27-
You can manage settings for the starter pool and create new Spark pools in the **Data Engineering/Science** section of the workspace settings.
27+
You can manage settings for the starter pool and create new Spark pools in the **Admin portal** section of the workspace settings, under **Capacity settings**, then **Data Engineering/Science Settings.**
2828

2929
![Screenshot of the Spark settings page in Microsoft Fabric.](../media/spark-settings.png)
3030

@@ -37,7 +37,7 @@ Specific configuration settings for Spark pools include:
3737
If you create one or more custom Spark pools in a workspace, you can set one of them (or the starter pool) as the default pool to be used if a specific pool is not specified for a given Spark job.
3838

3939
> [!TIP]
40-
> For more information about managing Spark pools in Microsoft Fabric, see **[Configuring starter pools in Microsoft Fabric](/fabric/data-engineering/configure-starter-pools)** and **[How to create custom Spark pools in Microsoft Fabric](/fabric/data-engineering/create-custom-spark-pools)** in the Microsoft Fabric documentation.
40+
> For more information about managing Spark pools in Microsoft Fabric, see **[Configuring starter pools in Microsoft Fabric](/fabric/data-engineering/configure-starter-pools?azure-portal=true)** and **[How to create custom Spark pools in Microsoft Fabric](/fabric/data-engineering/create-custom-spark-pools?azure-portal=true)** in the Microsoft Fabric documentation.
4141
4242
## Runtimes and environments
4343

@@ -50,7 +50,7 @@ In some cases, organizations may need to define multiple *environments* to suppo
5050
Microsoft Fabric supports multiple Spark runtimes, and will continue to add support for new runtimes as they are released. You can use the workspace settings interface to specify the Spark runtime that is used by default environment when a Spark pool is started.
5151

5252
> [!TIP]
53-
> For more information about Spark runtimes in Microsoft Fabric, see **[Apache Spark Runtimes in Fabric](/fabric/data-engineering/runtime)** in the Microsoft Fabric documentation.
53+
> For more information about Spark runtimes in Microsoft Fabric, see **[Apache Spark Runtimes in Fabric](/fabric/data-engineering/runtime?azure-portal=true)** in the Microsoft Fabric documentation.
5454
5555
### Environments in Microsoft Fabric
5656

@@ -71,7 +71,7 @@ When creating an environment, you can:
7171
After creating at least one custom environment, you can specify it as the default environment in the workspace settings.
7272

7373
> [!TIP]
74-
> For more information about using custom environments in Microsoft Fabric, see **[Create, configure, and use an environment in Microsoft Fabric](/fabric/data-engineering/create-and-use-environment)** in the Microsoft Fabric documentation.
74+
> For more information about using custom environments in Microsoft Fabric, see **[Create, configure, and use an environment in Microsoft Fabric](/fabric/data-engineering/create-and-use-environment=azure-portal=true)** in the Microsoft Fabric documentation.
7575
7676
## Additional Spark configuration options
7777

@@ -99,7 +99,7 @@ To enable the native execution engine for a specific script or notebook, you can
9999
```
100100

101101
> [!TIP]
102-
> For more information about the native execution engine, see **[Native execution engine for Fabric Spark](/fabric/data-engineering/native-execution-engine-overview)** in the Microsoft Fabric documentation.
102+
> For more information about the native execution engine, see **[Native execution engine for Fabric Spark](/fabric/data-engineering/native-execution-engine-overview?azure-portal=true)** in the Microsoft Fabric documentation.
103103
104104
### High concurrency mode
105105

@@ -108,7 +108,7 @@ When you run Spark code in Microsoft Fabric, a Spark session is initiated. You c
108108
To enable high concurrency mode, use the **Data Engineering/Science** section of the workspace settings interface.
109109

110110
> [!TIP]
111-
> For more information about high concurrency mode, see **[High concurrency mode in Apache Spark for Fabric](/fabric/data-engineering/high-concurrency-overview)** in the Microsoft Fabric documentation.
111+
> For more information about high concurrency mode, see **[High concurrency mode in Apache Spark for Fabric](/fabric/data-engineering/high-concurrency-overview?azure-portal=true)** in the Microsoft Fabric documentation.
112112
113113
### Automatic MLFlow logging
114114

@@ -119,5 +119,5 @@ MLFlow is an open source library that is used in data science workloads to manag
119119
Administrators can manage Spark settings at a Fabric capacity level, enabling them to restrict and override Spark settings in workspaces within an organization.
120120

121121
> [!TIP]
122-
> For more information about managing Spark configuration at the Fabric capacity level, see **[Configure and manage data engineering and data science settings for Fabric capacities](/fabric/data-engineering/capacity-settings-management)** in the Microsoft Fabric documentation.
122+
> For more information about managing Spark configuration at the Fabric capacity level, see **[Configure and manage data engineering and data science settings for Fabric capacities](/fabric/data-engineering/capacity-settings-management?azure-portal-true)** in the Microsoft Fabric documentation.
123123

0 commit comments

Comments
 (0)