You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+10-4Lines changed: 10 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,11 +6,17 @@
6
6
7
7
This is a package gathering functionalities to solve a number of generalised linear regression/classification problems which, inherently, correspond to an optimisation problem of the form
8
8
9
-
```
10
-
L(y, Xθ) + P(θ)
11
-
```
9
+
$$
10
+
L(y, X\theta) + P(\theta)
11
+
$$
12
+
13
+
where:
14
+
15
+
- $L$ is a loss function
16
+
- $X$ is the $n \times p$ matrix of training observations, where $n$ is the number of _observations_ (sample size) and $p$ is the number of _features_ (dimension)
17
+
- $\theta$ the length $p$ vector of weights to be optimized
18
+
- $P$ is a penalty function
12
19
13
-
where `L` is a loss function and `P` is a penalty function (both of those can be scaled or composed).
14
20
Additional regression/classification methods which do not directly correspond to this formulation may be added in the future.
Copy file name to clipboardExpand all lines: docs/src/index.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,7 +11,7 @@ where:
11
11
*``y`` is the **target** or **response**, a vector of length ``n`` either of real values (_regression_) or integers (_classification_),
12
12
*``X`` is the **design** or **feature** matrix, a matrix of real values of size ``n \times p`` where ``p`` is the number of _features_ or _dimensions_,\
13
13
*``\theta`` is a vector of ``p`` real valued coefficients to determine,
14
-
*``L`` is a **loss function**, a pre-determined function of ``\mathbb R^n`` to ``\mathbb R^+`` penalising the amplitude of the _residuals_ in a specific way,
14
+
*``L`` is a **loss function**, a pre-determined function of ``\mathbb R^n \times \mathbb R^n`` to ``\mathbb R^+`` penalising the amplitude of the _residuals_ in a specific way,
15
15
*``P`` is a **penalty function**, a pre-determined function of ``\mathbb R^n`` to ``\mathbb R^+`` penalising the amplitude of the _coefficients_ in a specific way.
16
16
17
17
A well known example is the [Ridge regression](https://en.wikipedia.org/wiki/Tikhonov_regularization) where the objective is to minimise:
@@ -51,17 +65,44 @@ Newton CG solver. This is the same as the Newton solver except that instead
51
65
of solving systems of the form `H\\b` where `H` is the full Hessian, it uses
52
66
a matrix-free conjugate gradient approach to solving that system. This should
53
67
generally be preferred for larger scale cases.
68
+
69
+
`optim_options` are the [general Optim Options](https://julianlsolvers.github.io/Optim.jl/stable/#user/config/).
70
+
`newtoncg_options` are the [options of Krylov Trust Region method](https://github.com/JuliaNLSolvers/Optim.jl/blob/master/src/multivariate/solvers/second_order/krylov_trust_region.jl)
0 commit comments