Skip to content

Commit 74a047d

Browse files
committed
Merge remote-tracking branch 'refs/remotes/MicrosoftDocs/master' into nitinme-freshness-updates
2 parents 55dca93 + 97dc104 commit 74a047d

File tree

6 files changed

+140
-5
lines changed

6 files changed

+140
-5
lines changed

articles/azure-monitor/insights/container-insights-prometheus-integration.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Configure Azure Monitor for containers Prometheus Integration | Microsoft Docs
33
description: This article describes how you can configure the Azure Monitor for containers agent to scrape metrics from Prometheus with your Kubernetes cluster.
44
ms.topic: conceptual
5-
ms.date: 01/13/2020
5+
ms.date: 04/16/2020
66
---
77

88
# Configure scraping of Prometheus metrics with Azure Monitor for containers
@@ -18,7 +18,6 @@ ms.date: 01/13/2020
1818
Scraping of Prometheus metrics is supported with Kubernetes clusters hosted on:
1919

2020
- Azure Kubernetes Service (AKS)
21-
- Azure Container Instances
2221
- Azure Stack or on-premises
2322
- Azure Red Hat OpenShift
2423

articles/lab-services/classroom-labs/TOC.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -50,6 +50,8 @@
5050
href: class-type-jupyter-notebook.md
5151
- name: Mobile app development with Android Studio
5252
href: class-type-mobile-dev-android-studio.md
53+
- name: Big data analytics
54+
href: class-type-big-data-analytics.md
5355
- name: Create and configure lab accounts (lab account owner)
5456
items:
5557
- name: Create and manage lab accounts
Lines changed: 132 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,132 @@
1+
---
2+
title: Set up a lab to teach big data analytics using Azure Lab Services | Microsoft Docs
3+
description: Learn how to set up a lab to teach the big data analytics using Docker deployment of Hortonworks Data Platform (HDP).
4+
services: lab-services
5+
documentationcenter: na
6+
author: nicolela
7+
manager:
8+
editor: ''
9+
10+
ms.service: lab-services
11+
ms.workload: na
12+
ms.tgt_pltfrm: na
13+
ms.devlang: na
14+
ms.topic: article
15+
ms.date: 4/15/2020
16+
ms.author: nicolela
17+
18+
---
19+
# Set up a lab to big data analytics using Docker deployment of HortonWorks Data Platform
20+
21+
This article shows you how to set up a lab to teach a big data analytics class. With this type of class, students learn how to handle large volumes of data and apply machine and statistical learning algorithms to derive data insights. A key objective for students is to learn to use data analytics tools, such as [Apache Hadoop's open-source software package](https://hadoop.apache.org/) which provides tools for storing, managing, and processing big data.
22+
23+
In this lab, students will use a popular commercial version of Hadoop provided by [Cloudera](https://www.cloudera.com/), called [Hortonworks Data Platform (HDP)](https://www.cloudera.com/products/hdp.html). Specifically, students will use [HDP Sandbox 3.0.1](https://www.cloudera.com/tutorials/getting-started-with-hdp-sandbox/1.html) which is a simplified, easy-to-use version of the platform that is free of cost and intended for learning and experimentation. Although this class may use either Windows or Linux virtual machines (VM) with HDP Sandbox deployed, this article will show how to use Windows.
24+
25+
Another interesting aspect of this lab, is that we will deploy HDP Sandbox on the lab VMs using [Docker](https://www.docker.com/) containers. Each Docker container provides its own isolated environment for software applications to run inside. Conceptually, Docker containers are like nested VMs and can be used to easily deploy and run a wide variety of software applications based on container images provided on [Docker Hub](https://www.docker.com/products/docker-hub). Cloudera's deployment script for HDP Sandbox automatically pulls the [HDP Sandbox 3.0.1 Docker image](https://hub.docker.com/r/hortonworks/sandbox-hdp) from Docker Hub and runs two Docker containers:
26+
- sandbox-hdp
27+
- sandbox-proxy
28+
29+
## Lab configuration
30+
31+
To set up this lab, you need an Azure subscription and lab account to get started. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/) before you begin. Once you get an Azure subscription, you can create a new lab account in Azure Lab Services. For more information about creating a new lab account, see [Tutorial to Setup a Lab Account](tutorial-setup-lab-account.md). You can also use an existing lab account.
32+
33+
### Lab account settings
34+
35+
Enable the settings described in the table below for the lab account. For more information about how to enable marketplace images, see [Specify Marketplace images available to lab creators](https://docs.microsoft.com/azure/lab-services/classroom-labs/specify-marketplace-images).
36+
37+
| Lab account setting | Instructions |
38+
| ------------------- | ------------ |
39+
|Marketplace image| Enable the Windows 10 Pro image for use within your lab account.|
40+
41+
### Lab settings
42+
43+
Use the settings in the table below when setting up a classroom lab. For more information how to create a classroom lab, see [set up a classroom lab tutorial](tutorial-setup-classroom-lab.md).
44+
45+
| Lab settings | Value/instructions |
46+
| ------------ | ------------------ |
47+
|Virtual Machine Size| Medium (Nested Virtualization). This VM size is best suited for relational databases, in-memory caching, and analytics. This size also supports nested virtualization.|
48+
|Virtual Machine Image| Windows 10 Pro|
49+
50+
> [!NOTE]
51+
> We need to use Medium (Nested Virtualization) since deploying HDP Sandbox using Docker requires:
52+
> - Windows Hyper-V with nested virtualization
53+
> - At least 10 GB of RAM
54+
55+
## Template machine configuration
56+
57+
To set up the template machine, we will:
58+
- Install Docker
59+
- Deploy HDP Sandbox
60+
- Use PowerShell and Windows Task Scheduler to automatically start the Docker containers
61+
62+
### Install Docker
63+
64+
The steps in this section are based on [Cloudera's instructions for deploying with Docker containers](https://www.cloudera.com/tutorials/sandbox-deployment-and-install-guide/3.html).
65+
66+
To use Docker containers, you must first install Docker Desktop on the template VM:
67+
68+
1. Follow the steps in the [Prerequisites section](https://www.cloudera.com/tutorials/sandbox-deployment-and-install-guide/3.html#prerequisites) to install [Docker for Windows](https://docs.docker.com/docker-for-windows/install/).
69+
70+
> [!IMPORTANT]
71+
> Ensure that the **Use Windows containers instead of Linux containers** configuration option is left unchecked.
72+
73+
1. Ensure that **Windows Containers and Hyper-V features** are turned on.
74+
![Turn Windows features on or off](../media/class-type-big-data-analytics/windows-hyperv-features.png)
75+
76+
1. Follow the steps in the [Memory for Windows](https://www.cloudera.com/tutorials/sandbox-deployment-and-install-guide/3.html#memory-for-windows) section to configure Docker's memory configuration.
77+
78+
> [!WARNING]
79+
> If you inadvertently check the **Use Windows containers instead of Linux containers** option when installing Docker, you won't see the memory configuration settings. To fix this, you can switch to using Linux containers by [clicking on the Docker icon in Windows System tray](https://docs.docker.com/docker-for-windows/#docker-settings-dialog); when the Docker Desktop menu opens, select **Switch to Linux containers**.
80+
81+
### Deploy HDP Sandbox
82+
83+
In this section, you will deploy HDP Sandbox and then also access HDP Sandbox using the browser.
84+
85+
1. Ensure that you have installed [Git Bash](https://gitforwindows.org/) as listed in the [Prerequisites section](https://www.cloudera.com/tutorials/sandbox-deployment-and-install-guide/3.html#prerequisites) of the guide since this is recommended for completing the next steps.
86+
87+
1. Using [Cloudera's Deployment and Install Guide for Docker](https://www.cloudera.com/tutorials/sandbox-deployment-and-install-guide/3.html), complete the steps in the following sections:
88+
89+
- Deploy HDP Sandbox
90+
- Verify HDP Sandbox
91+
92+
> [!WARNING]
93+
> When you download the latest .zip file for HDP, ensure that you do *not* save the .zip file in a directory path that includes whitespace.
94+
95+
> [!NOTE]
96+
> If you receive an exception during deployment stating **Drive has not been shared**, you need to share your C drive with Docker so that HDP's Linux containers can access local Windows files. To fix this, [click on the Docker icon in Windows System tray](https://docs.docker.com/docker-for-windows/#docker-settings-dialog) to open the Docker Desktop menu and select **Settings**. When **Docker's Settings** dialog opens, select **Resources > File Sharing** and check the **C** drive. You can then repeat the steps to deploy HDP Sandbox.
97+
98+
1. Once the Docker containers for HDP Sandbox are deployed and running, you can access the environment by launching your browser and following Cloudera's instructions for opening the [Sandbox Welcome Page](https://www.cloudera.com/tutorials/learning-the-ropes-of-the-hdp-sandbox.html#welcome-page) and launching the HDP Dashboard.
99+
100+
> [!NOTE]
101+
> These instructions assume that you have first mapped the local IP address of the sandbox environment to the sandbox-hdp.hortonworks.com in the host file on your template VM. If you do **not** do this mapping, you can access the Sandbox Welcome page by navigating to [http://localhost:8080](http://localhost:8080).
102+
103+
### Automatically start Docker containers when students log in
104+
105+
To provide an easy to use experience for students, we'll use a PowerShell script that automatically:
106+
- Starts the HDP Sandbox Docker containers when a student starts and connects to their lab VM.
107+
- Launches the browser and navigates to the Sandbox Welcome Page.
108+
We'll also use Windows Task Scheduler to automatically run this script when a student logs into their VM.
109+
To set this up, follow these steps: [Big Data Analytics scripting](https://github.com/Azure/azure-devtestlab/tree/master/samples/ClassroomLabs/Scripts/BigDataAnalytics/).
110+
111+
## Cost estimate
112+
113+
If you would like to estimate the cost of this lab, you can use the following example.
114+
115+
For a class of 25 students with 20 hours of scheduled class time and 10 hours of quota for homework or assignments, the price for the lab would be:
116+
- 25 students * (20 + 10) hours * 55 Lab Units * 0.01 USD per hour = 412.50 USD
117+
118+
Further more details on pricing, see [Azure Lab Services Pricing](https://azure.microsoft.com/pricing/details/lab-services/).
119+
120+
## Conclusion
121+
122+
This article walked you through the steps necessary to create a lab for a big data analytics class that using Hortonworks Data Platform deployed with Docker. The setup for this class type may be used for similar data analytics classes. This setup may also be applicable to other types of classes that use Docker for deployment.
123+
124+
## Next steps
125+
126+
Next steps are common to setting up any lab.
127+
128+
- [Create and manage a template](how-to-create-manage-template.md)
129+
- [Add users](tutorial-setup-classroom-lab.md#add-users-to-the-lab)
130+
- [Set quota](how-to-configure-student-usage.md#set-quotas-for-users)
131+
- [Set a schedule](tutorial-setup-classroom-lab.md#set-a-schedule-for-the-lab)
132+
- [Email registration links to students](how-to-configure-student-usage.md#send-invitations-to-users)
26.3 KB
Loading

articles/postgresql/concepts-read-replicas.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -93,14 +93,14 @@ For additional insight, query the master server directly to get the replication
9393
In PostgreSQL version 10:
9494

9595
```SQL
96-
select pg_wal_lsn_diff(pg_current_wal_lsn(), stat.replay_lsn)
96+
select pg_wal_lsn_diff(pg_current_wal_lsn(), replay_lsn)
9797
AS total_log_delay_in_bytes from pg_stat_replication;
9898
```
9999

100100
In PostgreSQL version 9.6 and earlier:
101101

102102
```SQL
103-
select pg_xlog_location_diff(pg_current_xlog_location(), stat.replay_location)
103+
select pg_xlog_location_diff(pg_current_xlog_location(), replay_location)
104104
AS total_log_delay_in_bytes from pg_stat_replication;
105105
```
106106

includes/virtual-machines-disks-incremental-snapshots-description.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,4 +16,6 @@ There are a few differences between an incremental snapshot and a regular snapsh
1616

1717
Incremental snapshots also offer a differential capability, only available to managed disks. They enable you to get the changes between two incremental snapshots of the same managed disks, down to the block level. You can use this capability to reduce your data footprint when copying snapshots across regions. For example, you can download the first incremental snapshot as a base blob in another region. For the subsequent incremental snapshots, you can copy only the changes since the last snapshot to the base blob. After copying the changes, you can take snapshots on the base blob that represent your point in time backup of the disk in another region. You can restore your disk either from the base blob or from a snapshot on the base blob in another region.
1818

19-
:::image type="content" source="media/virtual-machines-disks-incremental-snapshots-description/incremental-snapshot-diagram.png" alt-text="Diagram depicting incremental snapshots copied across regions. Snapshots make various API calls until eventually forming page blobs per each snapshot.":::
19+
:::image type="content" source="media/virtual-machines-disks-incremental-snapshots-description/incremental-snapshot-diagram.png" alt-text="Diagram depicting incremental snapshots copied across regions. Snapshots make various API calls until eventually forming page blobs per each snapshot.":::
20+
21+
You can see the used size of your snapshots by looking at the [Azure usage report](https://docs.microsoft.com/azure/billing/billing-understand-your-bill). For example, if the used data size of a snapshot is 10 GiB, the **daily** usage report will show 10 GiB/(31 days) = 0.3226 as the consumed quantity.

0 commit comments

Comments
 (0)