Skip to content

Commit 9360e32

Browse files
htahir1claudestrickvl
authored
Deprecate model deployer in docs (#4076)
* Add deprecation notice to Model Deployer docs Add a prominent deprecation notice to the Model Deployer documentation explaining that it has been deprecated in favor of the more flexible Deployer component and Pipeline Deployments feature. The notice explains: - Why Model Deployer is deprecated (focused on single-model serving vs. modern multi-step pipeline needs) - Benefits of the new Pipeline Deployment approach (unified, flexible, simpler, more extensible) - Clear migration path for users - Links to relevant documentation This aligns with the evolution toward pipeline deployments as described in the recent blog post about real-time AI pipelines. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <[email protected]> * Use absolute URLs in Model Deployer deprecation notice Update the deprecation notice to use absolute docs.zenml.io URLs instead of relative file paths: - https://docs.zenml.io/component-guide/deployers - https://docs.zenml.io/how-to/deployment This ensures the links work correctly when viewed on the documentation website. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <[email protected]> * Add Model Deployer deprecation guidance to tutorial and LLMOps docs Users following the hyperparameter tuning tutorial and LLMOps finetuning guide were being directed toward the deprecated Model Deployer component without awareness of the newer Pipeline Deployments approach. Added minimal, focused notices at decision points where users choose deployment strategies, guiding them toward Pipeline Deployments while acknowledging Model Deployer remains available for backward compatibility. * Update docs/book/component-guide/model-deployers/README.md * Apply suggestions from code review --------- Co-authored-by: Claude <[email protected]> Co-authored-by: Alex Strick van Linschoten <[email protected]>
1 parent 9e57686 commit 9360e32

File tree

3 files changed

+25
-2
lines changed

3 files changed

+25
-2
lines changed

docs/book/component-guide/model-deployers/README.md

Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,23 @@ description: Deploying your models and serve real-time predictions.
55

66
# Model Deployers
77

8+
{% hint style="warning" %}
9+
**DEPRECATION NOTICE**
10+
11+
The Model Deployer stack component is deprecated in favor of the more flexible [**Deployer**](https://docs.zenml.io/stacks/stack-components/deployers) component and [**Pipeline Deployments**](https://docs.zenml.io/concepts/deployment).
12+
13+
The Model Deployer abstraction focused exclusively on single-model serving, but modern ML workflows often require multi-step pipelines with preprocessing, tool integration, and custom business logic. The new Pipeline Deployment paradigm provides:
14+
15+
- **Unified approach**: Deploy any pipeline—classical ML inference, agentic workflows, or hybrid systems—as a long-running HTTP service
16+
- **Greater flexibility**: Customize your deployment with full FastAPI control, add middleware, custom routes, and even frontend interfaces
17+
- **Simpler mental model**: One primitive for all deployment scenarios instead of separate abstractions for models vs. pipelines
18+
- **Better extensibility**: Deploy to Docker, AWS App Runner, GCP Cloud Run, and other platforms with consistent patterns
19+
20+
**Migration Path**: Instead of using Model Deployer-specific steps, wrap your model inference logic in a regular ZenML pipeline and deploy it using `zenml pipeline deploy`. See the [Pipeline Deployment guide](https://docs.zenml.io/concepts/deployment) for examples of deploying ML models as HTTP services.
21+
22+
While Model Deployer integrations remain available for backward compatibility, we strongly recommend migrating to Pipeline Deployments for new projects.
23+
{% endhint %}
24+
825
Model Deployment is the process of making a machine learning model available to make predictions and decisions on
926
real-world data. Getting predictions from trained models can be done in different ways depending on the use case, a
1027
batch prediction is used to generate predictions for a large amount of data at once, while a real-time prediction is

docs/book/user-guide/llmops-guide/finetuning-llms/deploying-finetuned-models.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,10 @@ When choosing a deployment option, consider factors such as your team's expertis
4040

4141
## Deployment with vLLM and ZenML
4242

43+
{% hint style="info" %}
44+
**Note**: The example below uses the Model Deployer component, which is maintained for backward compatibility. For new projects, consider using [Pipeline Deployments](https://docs.zenml.io/concepts/deployment) which offer greater flexibility for deploying LLM inference workflows with custom preprocessing and business logic.
45+
{% endhint %}
46+
4347
[vLLM](https://github.com/vllm-project/vllm) is a fast and easy-to-use library
4448
for running large language models (LLMs) at high throughputs and low latency.
4549
ZenML comes with a [vLLM integration](https://docs.zenml.io/stacks/model-deployers/vllm)

docs/book/user-guide/tutorial/hyper-parameter-tuning.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -118,5 +118,7 @@ For a deeper exploration of how to query past pipeline runs, see the [Inspecting
118118
## Next steps
119119

120120
* Replace the simple grid‑search with a more sophisticated tuner (e.g. `sklearn.model_selection.GridSearchCV` or [Optuna](https://optuna.org/)).
121-
* Serve the winning model via a [Model Deployer](https://docs.zenml.io/stacks/model-deployers) to serve it right away.
122-
* Move the pipeline to a [remote orchestrator](https://docs.zenml.io/stacks/orchestrators) to scale out the search.
121+
* Deploy the winning model as an HTTP service using [Pipeline Deployments](https://docs.zenml.io/concepts/deployment) (recommended) or via the legacy [Model Deployer](https://docs.zenml.io/stacks/stack-components/model-deployers).
122+
* Move the pipeline to a [remote
123+
orchestrator](https://docs.zenml.io/stacks/orchestrators) to scale out the
124+
search.

0 commit comments

Comments
 (0)