Skip to content

Commit abc5a5c

Browse files
authored
Merge pull request #89338 from hrasheed-msft/hdi_abfs_correction
removing abfss URI scheme
2 parents 2784795 + df50e14 commit abc5a5c

13 files changed

+20
-22
lines changed

articles/hdinsight/hadoop/apache-hadoop-hive-java-udf.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ Learn how to create a Java-based user-defined function (UDF) that works with Apa
2020
* A Hadoop cluster on HDInsight. See [Get Started with HDInsight on Linux](./apache-hadoop-linux-tutorial-get-started.md).
2121
* [Java Developer Kit (JDK) version 8](https://aka.ms/azure-jdks)
2222
* [Apache Maven](https://maven.apache.org/download.cgi) properly [installed](https://maven.apache.org/install.html) according to Apache. Maven is a project build system for Java projects.
23-
* The [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be wasb:// for Azure Storage, abfs:// for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage or Data Lake Storage Gen2, the URI would be wasbs:// or abfss://, respectively See also, [secure transfer](../../storage/common/storage-require-secure-transfer.md).
23+
* The [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be wasb:// for Azure Storage, abfs:// for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage, the URI would be `wasbs://`. See also, [secure transfer](../../storage/common/storage-require-secure-transfer.md).
2424

2525
* A text editor or Java IDE
2626

articles/hdinsight/hadoop/apache-hadoop-on-premises-migration-best-practices-storage.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -104,7 +104,7 @@ In the past, cloud-based analytics had to compromise in areas of performance, ma
104104

105105
One of the following formats can be used to access data that is stored in ADLS Gen2:
106106
- `abfs:///`: Access the default Data Lake Storage for the cluster.
107-
- `abfs[s]://file_system@account_name.dfs.core.windows.net`: Used when communicating with a non-default Data Lake Storage.
107+
- `abfs://file_system@account_name.dfs.core.windows.net`: Used when communicating with a non-default Data Lake Storage.
108108

109109
For more information, see the following articles:
110110

articles/hdinsight/hadoop/apache-hadoop-use-hive-beeline.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -106,7 +106,7 @@ When connecting directly from the cluster head node, or from a resource inside t
106106

107107
* A Hadoop cluster on HDInsight. See [Get Started with HDInsight on Linux](./apache-hadoop-linux-tutorial-get-started.md).
108108

109-
* Notice the [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your cluster's primary storage. For example, `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2, or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage or Data Lake Storage Gen2, the URI is `wasbs://` or `abfss://`, respectively. For more information, see [secure transfer](../../storage/common/storage-require-secure-transfer.md).
109+
* Notice the [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your cluster's primary storage. For example, `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2, or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage, the URI is `wasbs://`. For more information, see [secure transfer](../../storage/common/storage-require-secure-transfer.md).
110110

111111

112112
* Option 1: An SSH client. For more information, see [Connect to HDInsight (Apache Hadoop) using SSH](../hdinsight-hadoop-linux-use-ssh-unix.md). Most of the steps in this document assume that you are using Beeline from an SSH session to the cluster.

articles/hdinsight/hadoop/python-udf-hdinsight.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ HDInsight also includes Jython, which is a Python implementation written in Java
2323

2424
* **A Hadoop cluster on HDInsight**. See [Get Started with HDInsight on Linux](apache-hadoop-linux-tutorial-get-started.md).
2525
* **An SSH client**. For more information, see [Connect to HDInsight (Apache Hadoop) using SSH](../hdinsight-hadoop-linux-use-ssh-unix.md).
26-
* The [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be wasb:// for Azure Storage, abfs:// for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage or Data Lake Storage Gen2, the URI would be wasbs:// or abfss://, respectively See also, [secure transfer](../../storage/common/storage-require-secure-transfer.md).
26+
* The [URI scheme](../hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage, the URI would be wasbs://. See also, [secure transfer](../../storage/common/storage-require-secure-transfer.md).
2727
* **Possible change to storage configuration.** See [Storage configuration](#storage-configuration) if using storage account kind `BlobStorage`.
2828
* Optional. If Planning to use PowerShell, you will need the [AZ module](https://docs.microsoft.com/powershell/azure/new-azureps-module-az) installed.
2929

articles/hdinsight/hdinsight-apache-storm-with-kafka.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -368,7 +368,7 @@ The project contains a file named `dev.properties` that is used to pass paramete
368368
| `kafka.broker.hosts` | The Kafka broker hosts (worker nodes). |
369369
| `kafka.topic` | The Kafka topic that the topologies use. |
370370
| `hdfs.write.dir` | The directory that the Kafka-reader topology writes to. |
371-
| `hdfs.url` | The file system used by the Storm cluster. For Azure Storage accounts, use a value of `wasb:///`. For Azure Data Lake Storage Gen2, use a value of `abfs:///`. For Azure Data Lake Storage Gen1, use a value of `adl:///`. |
371+
| `hdfs.url` | The file system used by the Storm cluster. For Azure Storage accounts, use a value of `wasb://`. For Azure Data Lake Storage Gen2, use a value of `abfs://`. For Azure Data Lake Storage Gen1, use a value of `adl://`. |
372372

373373
## Create the clusters
374374

articles/hdinsight/hdinsight-hadoop-compare-storage-options.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -90,9 +90,9 @@ For more information, see [The Azure Blob Filesystem driver (ABFS): A dedicated
9090

9191
Azure Data Lake Storage Gen2 uses a new URI scheme to access files in Azure Storage from HDInsight:
9292

93-
`abfs[s]://<FILE_SYSTEM_NAME>@<ACCOUNT_NAME>.dfs.core.windows.net/<PATH>`
93+
`abfs://<FILE_SYSTEM_NAME>@<ACCOUNT_NAME>.dfs.core.windows.net/<PATH>`
9494

95-
The URI scheme provides SSL-encrypted access (`abfss://` prefix) and unencrypted access (`abfs://` prefix). Use `abfss` wherever possible, even when accessing data that lives inside the same region in Azure.
95+
The URI scheme provides SSL-encrypted access.
9696

9797
`<FILE_SYSTEM_NAME>` identifies the path of the file system Data Lake Storage Gen2.
9898

@@ -103,8 +103,8 @@ The URI scheme provides SSL-encrypted access (`abfss://` prefix) and unencrypted
103103
If values for `<FILE_SYSTEM_NAME>` and `<ACCOUNT_NAME>` aren't specified, the default file system is used. For the files on the default file system, use a relative path or an absolute path. For example, the `hadoop-mapreduce-examples.jar` file that comes with HDInsight clusters can be referred to by using one of the following paths:
104104

105105
```
106-
abfss://[email protected]/example/jars/hadoop-mapreduce-examples.jar
107-
abfss:///example/jars/hadoop-mapreduce-examples.jar /example/jars/hadoop-mapreduce-examples.jar
106+
abfs://[email protected]/example/jars/hadoop-mapreduce-examples.jar
107+
abfs:///example/jars/hadoop-mapreduce-examples.jar /example/jars/hadoop-mapreduce-examples.jar
108108
```
109109

110110
> [!Note]

articles/hdinsight/hdinsight-hadoop-linux-information.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -119,11 +119,9 @@ When using __Azure Storage__, use one of the following URI schemes:
119119

120120
* `wasb://<container-name>@<account-name>.blob.core.windows.net/`: Used when communicating with a non-default storage account. For example, when you have an additional storage account or when accessing data stored in a publicly accessible storage account.
121121

122-
When using __Azure Data Lake Storage Gen2__, use one of the following URI schemes:
122+
When using __Azure Data Lake Storage Gen2__, use the following URI scheme:
123123

124-
* `abfs:///`: Access default storage using unencrypted communication.
125-
126-
* `abfss:///`: Access default storage using encrypted communication. The abfss scheme is supported only from HDInsight version 3.6 onwards.
124+
* `abfs://`: Access default storage using encrypted communication.
127125

128126
* `abfs://<container-name>@<account-name>.dfs.core.windows.net/`: Used when communicating with a non-default storage account. For example, when you have an additional storage account or when accessing data stored in a publicly accessible storage account.
129127

articles/hdinsight/hdinsight-storage-sharedaccesssignature-permissions.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ HDInsight has full access to data in the Azure Storage accounts associated with
3737

3838
* If using C#, Visual Studio must be version 2013 or higher.
3939

40-
* The [URI scheme](./hdinsight-hadoop-linux-information.md#URI-and-scheme) for your storage account. This would be `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2 or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage or Data Lake Storage Gen2, the URI would be `wasbs://` or `abfss://`, respectively See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
40+
* The [URI scheme](./hdinsight-hadoop-linux-information.md#URI-and-scheme) for your storage account. This would be `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2 or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage, the URI would be `wasbs://`. See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
4141

4242
* An existing HDInsight cluster to add a Shared Access Signature to. If not, you can use Azure PowerShell to create a cluster and add a Shared Access Signature during cluster creation.
4343

articles/hdinsight/hdinsight-use-oozie-linux-mac.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ You can also use Oozie to schedule jobs that are specific to a system, like Java
3232

3333
* **An Azure SQL Database**. See [Create an Azure SQL database in the Azure portal](../sql-database/sql-database-get-started.md). This article uses a database named `oozietest`.
3434

35-
* The [URI scheme](./hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2 or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage or Data Lake Storage Gen2, the URI would be `wasbs://` or `abfss://`, respectively See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
35+
* The [URI scheme](./hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be `wasb://` for Azure Storage, `abfs://` for Azure Data Lake Storage Gen2 or `adl://` for Azure Data Lake Storage Gen1. If secure transfer is enabled for Azure Storage, the URI would be `wasbs://`. See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
3636

3737
## Example workflow
3838

articles/hdinsight/hdinsight-using-spark-query-hbase.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ Apache HBase is typically queried either with its low-level API (scans, gets, an
2020

2121
* An SSH client. For more information, see [Connect to HDInsight (Apache Hadoop) using SSH](hdinsight-hadoop-linux-use-ssh-unix.md).
2222

23-
* The [URI scheme](hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be wasb:// for Azure Blob Storage, abfs:// for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Blob Storage or Data Lake Storage Gen2, the URI would be wasbs:// or abfss://, respectively See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
23+
* The [URI scheme](hdinsight-hadoop-linux-information.md#URI-and-scheme) for your clusters primary storage. This would be wasb:// for Azure Blob Storage, abfs:// for Azure Data Lake Storage Gen2 or adl:// for Azure Data Lake Storage Gen1. If secure transfer is enabled for Blob Storage, the URI would be `wasbs://`. See also, [secure transfer](../storage/common/storage-require-secure-transfer.md).
2424

2525
## Overall process
2626

0 commit comments

Comments
 (0)