Skip to content

Commit a9c2978

Browse files
author
Sreekanth Iyer (Ushta Te Consultancy Services)
committed
Corrected the link suggestins
1 parent 5ec0a24 commit a9c2978

File tree

3 files changed

+4
-4
lines changed

3 files changed

+4
-4
lines changed

articles/hdinsight/hadoop/apache-hadoop-deep-dive-advanced-analytics.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@ HDInsight has several machine learning options for an advanced analytics workflo
5757
There are three scalable machine learning libraries that bring algorithmic modeling capabilities to this distributed environment:
5858

5959
* [**MLlib**](https://spark.apache.org/docs/latest/ml-guide.html) - MLlib contains the original API built on top of Spark RDDs.
60-
* [**SparkML**](https://spark.apache.org/docs/1.2.2/ml-guide.html) - SparkML is a newer package that provides a higher-level API built on top of Spark DataFrames for constructing ML pipelines.
60+
* SparkML- SparkML is a newer package that provides a higher-level API built on top of Spark DataFrames for constructing ML pipelines.
6161
* [**MMLSpark**](https://github.com/Azure/mmlspark) - The Microsoft Machine Learning library for Apache Spark (MMLSpark) is designed to make data scientists more productive on Spark, to increase the rate of experimentation, and to leverage cutting-edge machine learning techniques, including deep learning, on large datasets. The MMLSpark library simplifies common modeling tasks for building models in PySpark.
6262

6363
### Azure Machine Learning and Apache Hive

articles/hdinsight/spark/apache-spark-ipython-notebook-machine-learning.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ The application uses the sample **HVAC.csv** data that is available on all clust
3333

3434
## Develop a Spark machine learning application using Spark MLlib
3535

36-
This application uses a Spark [ML pipeline](https://spark.apache.org/docs/2.2.0/ml-pipeline.html) to do a document classification. ML Pipelines provide a uniform set of high-level APIs built on top of DataFrames. The DataFrames help users create and tune practical machine learning pipelines. In the pipeline, you split the document into words, convert the words into a numerical feature vector, and finally build a prediction model using the feature vectors and labels. Do the following steps to create the application.
36+
This application uses a Spark [ML pipeline](https://downloads.apache.org/spark/docs/3.3.1/ml-pipeline.html) to do a document classification. ML Pipelines provide a uniform set of high-level APIs built on top of DataFrames. The DataFrames help users create and tune practical machine learning pipelines. In the pipeline, you split the document into words, convert the words into a numerical feature vector, and finally build a prediction model using the feature vectors and labels. Do the following steps to create the application.
3737

3838
1. Create a Jupyter Notebook using the PySpark kernel. For the instructions, see [Create a Jupyter Notebook file](./apache-spark-jupyter-spark-sql.md#create-a-jupyter-notebook-file).
3939

articles/hdinsight/spark/apache-spark-structured-streaming-overview.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ This query yields results similar to the following:
126126
|{u'start': u'2016-07-26T07:00:00.000Z', u'end'... |95 | 96.980971 | 99 |
127127
|{u'start': u'2016-07-26T08:00:00.000Z', u'end'... |95 | 96.965997 | 99 |
128128

129-
For details on the Spark Structured Stream API, along with the input data sources, operations, and output sinks it supports, see [Apache Spark Structured Streaming Programming Guide](https://spark.apache.org/docs/2.1.0/structured-streaming-programming-guide.html).
129+
For details on the Spark Structured Stream API, along with the input data sources, operations, and output sinks it supports, see [Apache Spark Structured Streaming Programming Guide](https://spark.apache.org/docs/latest/ss-migration-guide.html).
130130

131131
## Checkpointing and write-ahead logs
132132

@@ -143,5 +143,5 @@ The status of all applications can also be checked with a GET request against a
143143
## Next steps
144144

145145
* [Create an Apache Spark cluster in HDInsight](../hdinsight-hadoop-create-linux-clusters-portal.md)
146-
* [Apache Spark Structured Streaming Programming Guide](https://spark.apache.org/docs/2.1.0/structured-streaming-programming-guide.html)
146+
* [Apache Spark Structured Streaming Programming Guide](https://spark.apache.org/docs/latest/ss-migration-guide.html)
147147
* [Launch Apache Spark jobs remotely with Apache LIVY](apache-spark-livy-rest-interface.md)

0 commit comments

Comments
 (0)