Skip to content

Commit 0238f4e

Browse files
authored
Add intro, features
1 parent b29d19f commit 0238f4e

File tree

1 file changed

+12
-49
lines changed

1 file changed

+12
-49
lines changed

README.md

Lines changed: 12 additions & 49 deletions
Original file line numberDiff line numberDiff line change
@@ -1,22 +1,19 @@
1-
# About
1+
# Core 2: Deploy Modular, Data-centric AI applications at scale
22

3-
Seldon Core 2 provides a state of the art solution for machine learning inference which can be run locally on a laptop as well as on Kubernetes for production.
3+
## 📖 About
4+
Seldon Core 2 is an MLOps and LLMOps framework for deploying, managing and scaling AI systems in Kubernetes - from singular models, to modular, data-centric applications. With Core 2 you can deploy across a wide range of model types, on-prem or in any cloud, in a standardized way that is production-ready out of the box.
45

56
[![Introductory Youtube Video](./docs-gb/images/Core-intro-thumbnail.png)](https://www.youtube.com/watch?v=ar5lSG_idh4)
67

78
## Features
89

9-
* A single platform for inference of wide range of standard and custom artifacts.
10-
* Deploy locally in Docker during development and testing of models.
11-
* Deploy at scale on Kubernetes for production.
12-
* Deploy single models to multi-step pipelines.
13-
* Save infrastructure costs by deploying multiple models transparently in inference servers.
14-
* Overcommit on resources to deploy more models than available memory.
15-
* Dynamically extended models with pipelines with a data-centric perspective backed by Kafka.
16-
* Explain individual models and pipelines with state of the art explanation techniques.
17-
* Deploy drift and outlier detectors alongside models.
18-
* Kubernetes Service mesh agnostic - use the service mesh of your choice.
19-
10+
* **Pipelines**: Deploy composable AI pipelines, leveraging Kafka for realtime data streaming between components
11+
* **Autoscaling** for models and application components based on native or custom logic
12+
* **Multi-Model Serving**: Save infrastructure costs by consolidating multiple models on shared inference servers
13+
* **Overcommit**: Deploy more models than available memory allows, saving infrastructure costs for unused models
14+
* **Experiments**: Route data between candidate models or pipeline, with support for A/B tests and shadow deployments
15+
* **Custom Components**: Implement custom logic, drift & outlier detection, LLMs and more through plug-and-play integrate with the rest of Seldon's ecosytem of ML/AI products!
16+
2017
## Publication
2118

2219
These features are influenced by our position paper on the next generation of ML model serving frameworks:
@@ -26,45 +23,11 @@ These features are influenced by our position paper on the next generation of ML
2623
*Workshop*: Challenges in deploying and monitoring ML systems workshop - NeurIPS 2022
2724

2825

29-
## Getting started
30-
31-
### Kubernetes quick-start via `KinD`
32-
33-
Install Seldon ansible collection
34-
35-
```
36-
pip install ansible openshift docker passlib
37-
ansible-galaxy collection install git+https://github.com/SeldonIO/ansible-k8s-collection.git
38-
```
39-
40-
Create a KinD cluster and install dependencies:
41-
42-
```
43-
cd ansible
44-
ansible-playbook playbooks/kind-cluster.yaml
45-
ansible-playbook playbooks/setup-ecosystem.yaml
46-
```
47-
48-
Deploy Seldon Core 2
49-
50-
```
51-
cd ..
52-
make deploy-k8s
53-
```
54-
55-
Run [k8s-examples.ipynb](samples/k8s-examples.ipynb)
56-
57-
Undeploy Seldon Core 2
58-
59-
```
60-
make undeploy-k8s
61-
```
26+
## ⚡️ Quickstart
6227

6328

6429
## Documentation
6530

6631
[Seldon Core 2 docs](https://docs.seldon.ai/seldon-core-2)
6732

68-
## License
69-
70-
[License](LICENSE)
33+
## 📜 License

0 commit comments

Comments
 (0)