Skip to content

Commit ac3e276

Browse files
authored
Merge pull request #121123 from arun-rajora/many-models-ga-update
Remove preview for many models
2 parents 5ff2cde + bcf09f8 commit ac3e276

File tree

3 files changed

+7
-7
lines changed

3 files changed

+7
-7
lines changed

articles/machine-learning/concept-automl-forecasting-at-scale.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,7 @@ ms.date: 08/01/2023
1414
show_latex: true
1515
---
1616

17-
# Forecasting at scale: many models and distributed training (preview)
18-
19-
[!INCLUDE [machine-learning-preview-generic-disclaimer](./includes/machine-learning-preview-generic-disclaimer.md)]
17+
# Forecasting at scale: many models and distributed training
2018

2119
This article is about training forecasting models on large quantities of historical data. Instructions and examples for training forecasting models in AutoML can be found in our [set up AutoML for time series forecasting](./how-to-auto-train-forecast.md) article.
2220

@@ -28,7 +26,7 @@ The many models [components](concept-component.md) in AutoML enable you to train
2826

2927
:::image type="content" source="./media/how-to-auto-train-forecast/many-models.svg" alt-text="Diagram showing the AutoML many models workflow.":::
3028

31-
The many models training component applies AutoML's [model sweeping and selection](concept-automl-forecasting-sweeping.md) independently to each store in this example. This model independence aids scalability and can benefit model accuracy especially when the stores have diverging sales dynamics. However, a single model approach may yield more accurate forecasts when there are common sales dynamics. See the [distributed DNN training](#distributed-dnn-training) section for more details on that case.
29+
The many models training component applies AutoML's [model sweeping and selection](concept-automl-forecasting-sweeping.md) independently to each store in this example. This model independence aids scalability and can benefit model accuracy especially when the stores have diverging sales dynamics. However, a single model approach may yield more accurate forecasts when there are common sales dynamics. See the [distributed DNN training](#distributed-dnn-training-preview) section for more details on that case.
3230

3331
You can configure the data partitioning, the [AutoML settings](how-to-auto-train-forecast.md#configure-experiment) for the models, and the degree of parallelism for many models training jobs. For examples, see our guide section on [many models components](how-to-auto-train-forecast.md#forecasting-at-scale-many-models).
3432

@@ -53,7 +51,9 @@ AutoML supports the following features for hierarchical time series (HTS):
5351
HTS components in AutoML are built on top of [many models](#many-models), so HTS shares the scalable properties of many models.
5452
For examples, see our guide section on [HTS components](how-to-auto-train-forecast.md#forecasting-at-scale-hierarchical-time-series).
5553

56-
## Distributed DNN training
54+
## Distributed DNN training (preview)
55+
56+
[!INCLUDE [machine-learning-preview-generic-disclaimer](./includes/machine-learning-preview-generic-disclaimer.md)]
5757

5858
Data scenarios featuring large amounts of historical observations and/or large numbers of related time series may benefit from a scalable, single model approach. Accordingly, **AutoML supports distributed training and model search on temporal convolutional network (TCN) models**, which are a type of deep neural network (DNN) for time series data. For more information on AutoML's TCN model class, see our [DNN article](concept-automl-forecasting-deep-learning.md).
5959

articles/machine-learning/how-to-auto-train-forecast.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1585,7 +1585,7 @@ Also see the [demand forecasting with hierarchical time series notebook](https:/
15851585

15861586
## Forecasting at scale: distributed DNN training
15871587

1588-
* To learn how distributed training works for forecasting tasks, see our [forecasting at scale article](concept-automl-forecasting-at-scale.md#distributed-dnn-training).
1588+
* To learn how distributed training works for forecasting tasks, see our [forecasting at scale article](concept-automl-forecasting-at-scale.md#distributed-dnn-training-preview).
15891589
* See our [setup distributed training for tabular data](how-to-configure-auto-train.md#automl-at-scale-distributed-training) article section for code samples.
15901590

15911591
## Example notebooks

articles/machine-learning/how-to-configure-auto-train.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -698,7 +698,7 @@ limits:
698698

699699
### Distributed training for forecasting
700700

701-
To learn how distributed training works for forecasting tasks, see our [forecasting at scale](concept-automl-forecasting-at-scale.md#distributed-dnn-training) article. To use distributed training for forecasting, you need to set the `training_mode`, `enable_dnn_training`, `max_nodes`, and optionally the `max_concurrent_trials` properties of the job object.
701+
To learn how distributed training works for forecasting tasks, see our [forecasting at scale](concept-automl-forecasting-at-scale.md#distributed-dnn-training-preview) article. To use distributed training for forecasting, you need to set the `training_mode`, `enable_dnn_training`, `max_nodes`, and optionally the `max_concurrent_trials` properties of the job object.
702702

703703
Property | Description
704704
-- | --

0 commit comments

Comments
 (0)