Skip to content

Commit 599761e

Browse files
scientific appls / discipls removed
1 parent 50350a9 commit 599761e

File tree

2 files changed

+4
-5
lines changed

2 files changed

+4
-5
lines changed

clustering.Rmd

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1098,4 +1098,4 @@ please follow the instructions for computer setup needed to run the worksheets
10981098
found in Chapter \@ref(move-to-your-own-machine).
10991099

11001100
## Additional resources
1101-
- Chapter 10 of [*An Introduction to Statistical Learning*](https://www.statlearning.com/) [-@james2013introduction] provides a great next stop in the process of learning about clustering and unsupervised learning in general. In the realm of clustering specifically, it provides a great companion introduction to K-means, but also covers *hierarchical* clustering for when you expect there to be subgroups, and then subgroups within subgroups, etc., in your data. In the realm of more general unsupervised learning, it covers *principal components analysis (PCA)*, which is a very popular technique in scientific applications for reducing the number of predictors in a dataset.
1101+
- Chapter 10 of [*An Introduction to Statistical Learning*](https://www.statlearning.com/) [-@james2013introduction] provides a great next stop in the process of learning about clustering and unsupervised learning in general. In the realm of clustering specifically, it provides a great companion introduction to K-means, but also covers *hierarchical* clustering for when you expect there to be subgroups, and then subgroups within subgroups, etc., in your data. In the realm of more general unsupervised learning, it covers *principal components analysis (PCA)*, which is a very popular technique for reducing the number of predictors in a dataset.

regression2.Rmd

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -29,8 +29,7 @@ print_tidymodels <- function(tidymodels_object) {
2929
## Overview
3030
Up to this point, we have solved all of our predictive problems---both classification
3131
and regression---using K-nearest neighbors (KNN)-based approaches. In the context of regression,
32-
there is another method commonly used
33-
in scientific disciplines known as *linear regression*. This chapter provides an introduction
32+
there is another commonly used method known as *linear regression*. This chapter provides an introduction
3433
to the basic concept of linear regression, shows how to use `tidymodels` to perform linear regression in R,
3534
and characterizes its strengths and weaknesses compared to KNN regression. The focus is, as usual,
3635
on the case where there is a single predictor and single response variable of interest; but the chapter
@@ -53,8 +52,8 @@ predict well beyond the range of the predictors in the training data, and
5352
the method gets significantly slower as the training data set grows.
5453
\index{regression!linear}
5554
Fortunately, there is an alternative to KNN regression---*linear regression*---that addresses
56-
both of these limitations. Linear regression is also much more commonly used in practice, especially
57-
in scientific applications, because it provides an interpretable mathematical equation that describes
55+
both of these limitations. Linear regression is also very commonly
56+
used in practice because it provides an interpretable mathematical equation that describes
5857
the relationship between the predictor and response variables. In this first part of the chapter, we will focus on *simple* linear regression,
5958
which involves only one predictor variable and one response variable; later on, we will consider
6059
*multivariable* linear regression, which involves multiple predictor variables.

0 commit comments

Comments
 (0)