Skip to content

Commit 0f79e71

Browse files
authored
Merge pull request #91925 from guyhay/Investigate-Spark-from-U-SQL
Understanding spark for U-SQL developers.
2 parents df18418 + 06f4b38 commit 0f79e71

6 files changed

+359
-21
lines changed

articles/data-lake-analytics/TOC.yml

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,8 +40,8 @@
4040
href: data-lake-analytics-manage-use-nodejs.md
4141
- name: Add users
4242
href: data-lake-analytics-add-users.md
43-
- name: Policies
44-
href: data-lake-analytics-policies.md
43+
- name: Account policies
44+
href: data-lake-analytics-account-policies.md
4545
- name: Secure job folders
4646
href: data-lake-analytics-secure.md
4747
- name: Access diagnostic logs
@@ -118,6 +118,14 @@
118118
href: data-lake-analytics-cicd-test.md
119119
- name: U-SQL SDK
120120
href: data-lake-analytics-u-sql-sdk.md
121+
- name: Understand Apache Spark for U-SQL developers
122+
items:
123+
- name: Understand Apache Spark for U-SQL developers
124+
href: data-lake-analytics-understand-spark-for-usql-developers.md
125+
- name: Understand Apache Spark data formats
126+
href: data-lake-analytics-understand-spark-data-formats.md
127+
- name: Understand Apache Spark code concepts
128+
href: data-lake-analytics-understand-spark-code-concepts.md
121129
- name: Reference
122130
items:
123131
- name: Azure PowerShell

articles/data-lake-analytics/data-lake-analytics-policies.md renamed to articles/data-lake-analytics/data-lake-analytics-account-policies.md

Lines changed: 16 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -1,19 +1,19 @@
11
---
2-
title: Manage Azure Data Lake Analytics policies
3-
description: Learn how to use policies to control usage of a Data Lake Analytics account, such as maximum AUs and maximum jobs.
2+
title: Manage Azure Data Lake Analytics Account Policies
3+
description: Learn how to use account policies to control usage of a Data Lake Analytics account, such as maximum AUs and maximum jobs.
44
services: data-lake-analytics
55
ms.service: data-lake-analytics
6-
author: saveenr
7-
ms.author: saveenr
6+
author: guyhay
7+
ms.author: guyhay
88

99
ms.reviewer: jasonwhowell
1010
ms.assetid: 0a6102d1-7554-4df2-b487-4dae9a7287b6
1111
ms.topic: conceptual
1212
ms.date: 04/30/2018
1313
---
14-
# Manage Azure Data Lake Analytics using policies
14+
# Manage Azure Data Lake Analytics using Account Policies
1515

16-
Using account policies, you can control how resources an Azure Data Lake Analytics account are used. These policies allow you to control the cost of using Azure Data Lake Analytics. For example, with these policies you can prevent unexpected cost spikes by limiting how many AUs the account can simultaneously use.
16+
Account policies help you control how resources an Azure Data Lake Analytics account are used. These policies allow you to control the cost of using Azure Data Lake Analytics. For example, with these policies you can prevent unexpected cost spikes by limiting how many AUs the account can simultaneously use.
1717

1818
## Account-level policies
1919

@@ -25,7 +25,7 @@ A policy controls the total number of Analytics Units (AUs) your Data Lake Analy
2525
To change the number of AUs for your Data Lake Analytics account:
2626

2727
1. In the Azure portal, go to your Data Lake Analytics account.
28-
2. Click **Properties**.
28+
2. Click **Limits and policies**.
2929
3. Under **Maximum AUs**, move the slider to select a value, or enter the value in the text box.
3030
4. Click **Save**.
3131

@@ -34,12 +34,12 @@ To change the number of AUs for your Data Lake Analytics account:
3434
>
3535
3636
### Maximum number of jobs that can run simultaneously
37-
A policy controls how many jobs can run at the same time. By default, this value is set to 20. If your Data Lake Analytics has AUs available, new jobs are scheduled to run immediately until the total number of running jobs reaches the value of this policy. When you reach the maximum number of jobs that can run simultaneously, subsequent jobs are queued in priority order until one or more running jobs complete (depending on AU availability).
37+
This policy limits how many jobs can run simultaneously. By default, this value is set to 20. If your Data Lake Analytics has AUs available, new jobs are scheduled to run immediately until the total number of running jobs reaches the value of this policy. When you reach the maximum number of jobs that can run simultaneously, subsequent jobs are queued in priority order until one or more running jobs complete (depending on available AUs).
3838

3939
To change the number of jobs that can run simultaneously:
4040

4141
1. In the Azure portal, go to your Data Lake Analytics account.
42-
2. Click **Properties**.
42+
2. Click **Limits and policies**.
4343
3. Under **Maximum Number of Running Jobs**, move the slider to select a value, or enter the value in the text box.
4444
4. Click **Save**.
4545

@@ -48,25 +48,25 @@ To change the number of jobs that can run simultaneously:
4848
>
4949
5050
### How long to keep job metadata and resources
51-
When your users run U-SQL jobs, the Data Lake Analytics service retains all related files. Related files include the U-SQL script, the DLL files referenced in the U-SQL script, compiled resources, and statistics. The files are in the /system/ folder of the default Azure Data Lake Storage account. This policy controls how long these resources are stored before they are automatically deleted (the default is 30 days). You can use these files for debugging, and for performance-tuning of jobs that you'll rerun in the future.
51+
When your users run U-SQL jobs, the Data Lake Analytics service keeps all related files. These files include the U-SQL script, the DLL files referenced in the U-SQL script, compiled resources, and statistics. The files are in the /system/ folder of the default Azure Data Lake Storage account. This policy controls how long these resources are stored before they are automatically deleted (the default is 30 days). You can use these files for debugging, and for performance-tuning of jobs that you'll rerun in the future.
5252

5353
To change how long to keep job metadata and resources:
5454

5555
1. In the Azure portal, go to your Data Lake Analytics account.
56-
2. Click **Properties**.
56+
2. Click **Limits and policies**.
5757
3. Under **Days to Retain Job Queries**, move the slider to select a value, or enter the value in the text box.
5858
4. Click **Save**.
5959

6060
## Job-level policies
6161

62-
With job-level policies, you can control the maximum AUs and the maximum priority that individual users (or members of specific security groups) can set on jobs that they submit. This policy lets you control the costs incurred by users. It also lets you control the effect that scheduled jobs might have on high-priority production jobs that are running in the same Data Lake Analytics account.
62+
Job-level policies allow you to control the maximum AUs and the maximum priority that individual users (or members of specific security groups) can set on jobs that they submit. This policy lets you control the costs incurred by users. It also lets you control the effect that scheduled jobs might have on high-priority production jobs that are running in the same Data Lake Analytics account.
6363

6464
Data Lake Analytics has two policies that you can set at the job level:
6565

6666
* **AU limit per job**: Users can only submit jobs that have up to this number of AUs. By default, this limit is the same as the maximum AU limit for the account.
6767
* **Priority**: Users can only submit jobs that have a priority lower than or equal to this value. A higher number indicates a lower priority. By default, this limit is set to 1, which is the highest possible priority.
6868

69-
There is a default policy set on every account. The default policy applies to all users of the account. You can set additional policies for specific users and groups.
69+
There is a default policy set on every account. The default policy applies to all users of the account. You can create additional policies for specific users and groups.
7070

7171
> [!NOTE]
7272
> Account-level policies and job-level policies apply simultaneously.
@@ -75,7 +75,7 @@ There is a default policy set on every account. The default policy applies to al
7575
### Add a policy for a specific user or group
7676

7777
1. In the Azure portal, go to your Data Lake Analytics account.
78-
2. Click **Properties**.
78+
2. Click **Limits and policies**.
7979
3. Under **Job Submission Limits**, click the **Add Policy** button. Then, select or enter the following settings:
8080
1. **Compute Policy Name**: Enter a policy name, to remind you of the purpose of the policy.
8181
2. **Select User or Group**: Select the user or group this policy applies to.
@@ -89,7 +89,7 @@ There is a default policy set on every account. The default policy applies to al
8989
### Delete or edit an existing policy
9090

9191
1. In the Azure portal, go to your Data Lake Analytics account.
92-
2. Click **Properties**.
92+
2. Click **Limits and policies**.
9393
3. Under **Job Submission Limits**, find the policy you want to edit.
9494
4. To see the **Delete** and **Edit** options, in the rightmost column of the table, click `...`.
9595

@@ -102,5 +102,4 @@ There is a default policy set on every account. The default policy applies to al
102102

103103
* [Overview of Azure Data Lake Analytics](data-lake-analytics-overview.md)
104104
* [Get started with Data Lake Analytics by using the Azure portal](data-lake-analytics-get-started-portal.md)
105-
* [Manage Azure Data Lake Analytics by using Azure PowerShell](data-lake-analytics-manage-use-powershell.md)
106-
105+
* [Manage Azure Data Lake Analytics by using Azure PowerShell](data-lake-analytics-manage-use-powershell.md)

0 commit comments

Comments
 (0)