Skip to content

Commit 41016af

Browse files
9.9.0 updated documentation
1 parent 9cb5177 commit 41016af

File tree

5 files changed

+6
-6
lines changed

5 files changed

+6
-6
lines changed

API_REFERENCE_FOR_CLASSIFICATION.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ Specifies a penalty in the range [0.0, 1.0] on terms that are not linear effects
5959
Specifies a penalty in the range [0.0, 1.0] on interaction terms. A higher value increases model interpretability but can hurt predictiveness. Values outside of the [0.0, 1.0] range are rounded to the nearest boundary within the range.
6060

6161
#### max_terms (default = 0)
62-
Restricts the maximum number of terms in any of the underlying models trained to ***max_terms***. The default value of 0 means no limit. After the limit is reached, the remaining boosting steps are used to further update the coefficients of already included terms. A potential tuning objective could be to find the lowest positive value of ***max_terms*** that does not increase the prediction error significantly. Low positive values can speed up the training process significantly.
62+
Restricts the maximum number of terms in any of the underlying models trained to ***max_terms***. The default value of 0 means no limit. After the limit is reached, the remaining boosting steps are used to further update the coefficients of already included terms. An optional tuning objective could be to find the lowest positive value of ***max_terms*** that does not increase the prediction error significantly. Low positive values can speed up the training process significantly. Setting a limit with ***max_terms*** may require a higher learning rate for best results.
6363

6464

6565
## Method: fit(X:npt.ArrayLike, y:List[str], sample_weight:npt.ArrayLike = np.empty(0), X_names:List[str]=[], cv_observations: npt.ArrayLike = np.empty([0, 0]), prioritized_predictors_indexes:List[int]=[], monotonic_constraints:List[int]=[], interaction_constraints:List[List[int]]=[], predictor_learning_rates: List[float] = [], predictor_penalties_for_non_linearity: List[float] = [], predictor_penalties_for_interactions: List[float] = [])

API_REFERENCE_FOR_REGRESSION.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -127,7 +127,7 @@ Specifies a penalty in the range [0.0, 1.0] on terms that are not linear effects
127127
Specifies a penalty in the range [0.0, 1.0] on interaction terms. A higher value increases model interpretability but can hurt predictiveness. Values outside of the [0.0, 1.0] range are rounded to the nearest boundary within the range.
128128

129129
#### max_terms (default = 0)
130-
Restricts the maximum number of terms in any of the underlying models trained to ***max_terms***. The default value of 0 means no limit. After the limit is reached, the remaining boosting steps are used to further update the coefficients of already included terms. A potential tuning objective could be to find the lowest positive value of ***max_terms*** that does not increase the prediction error significantly. Low positive values can speed up the training process significantly.
130+
Restricts the maximum number of terms in any of the underlying models trained to ***max_terms***. The default value of 0 means no limit. After the limit is reached, the remaining boosting steps are used to further update the coefficients of already included terms. An optional tuning objective could be to find the lowest positive value of ***max_terms*** that does not increase the prediction error significantly. Low positive values can speed up the training process significantly. Setting a limit with ***max_terms*** may require a higher learning rate for best results.
131131

132132

133133
## Method: fit(X:npt.ArrayLike, y:npt.ArrayLike, sample_weight:npt.ArrayLike = np.empty(0), X_names:List[str]=[], cv_observations: npt.ArrayLike = np.empty([0, 0]), prioritized_predictors_indexes:List[int]=[], monotonic_constraints:List[int]=[], group:npt.ArrayLike = np.empty(0), interaction_constraints:List[List[int]]=[], other_data: npt.ArrayLike = np.empty([0, 0]), predictor_learning_rates: List[float] = [], predictor_penalties_for_non_linearity: List[float] = [], predictor_penalties_for_interactions: List[float] = [])

documentation/APLR 9.9.0.pdf

309 Bytes
Binary file not shown.

examples/train_aplr_classification.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@
4949
verbosity=2,
5050
m=3000,
5151
v=0.1,
52-
# max_terms=5, # max terms in each underlying model. Tune this to find a trade-off between interpretability and predictiveness.
52+
# max_terms=5, # Optionally tune this to find a trade-off between interpretability and predictiveness. May require a higher learning rate for best results.
5353
**params
5454
)
5555
model.fit(

examples/train_aplr_regression.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -40,9 +40,9 @@
4040
}
4141
)
4242
best_model = None
43-
loss_function = "mse" # other available loss functions are binomial, poisson, gamma, tweedie, group_mse, mae, quantile, negative_binomial, cauchy, weibull and custom_function.
43+
loss_function = "mse" # Other available loss functions are binomial, poisson, gamma, tweedie, group_mse, mae, quantile, negative_binomial, cauchy, weibull and custom_function.
4444
link_function = (
45-
"identity" # other available link functions are logit, log and custom_function.
45+
"identity" # Other available link functions are logit, log and custom_function.
4646
)
4747
for params in param_grid:
4848
model = APLRRegressor(
@@ -52,7 +52,7 @@
5252
v=0.1,
5353
loss_function=loss_function,
5454
link_function=link_function,
55-
# max_terms=10, # max terms in each underlying model. Tune this to find a trade-off between interpretability and predictiveness.
55+
# max_terms=10, # Optionally tune this to find a trade-off between interpretability and predictiveness. May require a higher learning rate for best results.
5656
**params
5757
)
5858
model.fit(

0 commit comments

Comments
 (0)