You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/server-async/README.md
+49-14Lines changed: 49 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,24 +5,24 @@
5
5
6
6
## ⚠️ IMPORTANT
7
7
8
-
* The server and inference harness live in this repo: `https://github.com/F4k3r22/DiffusersServer`.
9
-
The example demonstrates how to run pipelines like `StableDiffusion3-3.5` and `Flux.1` concurrently while keeping a single copy of the heavy model parameters on GPU.
8
+
* The example demonstrates how to run pipelines like `StableDiffusion3-3.5` and `Flux.1` concurrently while keeping a single copy of the heavy model parameters on GPU.
10
9
11
10
## Necessary components
12
11
13
-
All the components needed to create the inference server are in `DiffusersServer/`
12
+
All the components needed to create the inference server are in the current directory:
14
13
15
14
```
16
-
DiffusersServer/
15
+
server-async/
17
16
├── utils/
18
17
├─────── __init__.py
19
-
├─────── scheduler.py # BaseAsyncScheduler wrapper and async_retrieve_timesteps for secure inferences
20
-
├─────── requestscopedpipeline.py # RequestScoped Pipeline for inference with a single in-memory model
21
-
├── __init__.py
22
-
├── create_server.py # helper script to build/run the app programmatically
0 commit comments