|
36 | 36 | "configuration_utils": ["ConfigMixin"], |
37 | 37 | "guiders": [], |
38 | 38 | "hooks": [], |
39 | | - "image_processor": [ |
40 | | - "VaeImageProcessor", |
41 | | - "VaeImageProcessorLDM3D", |
42 | | - "PixArtImageProcessor", |
43 | | - "IPAdapterMaskProcessor", |
44 | | - ], |
45 | 39 | "video_processor": [ |
46 | 40 | "VideoProcessor", |
47 | 41 | ], |
|
169 | 163 | "apply_pyramid_attention_broadcast", |
170 | 164 | ] |
171 | 165 | ) |
| 166 | + _import_structure["image_processor"] = [ |
| 167 | + "IPAdapterMaskProcessor", |
| 168 | + "PixArtImageProcessor", |
| 169 | + "VaeImageProcessor", |
| 170 | + "VaeImageProcessorLDM3D", |
| 171 | + ] |
172 | 172 | _import_structure["models"].extend( |
173 | 173 | [ |
174 | 174 | "AllegroTransformer3DModel", |
|
834 | 834 | apply_layer_skip, |
835 | 835 | apply_pyramid_attention_broadcast, |
836 | 836 | ) |
| 837 | + from .image_processor import ( |
| 838 | + IPAdapterMaskProcessor, |
| 839 | + PixArtImageProcessor, |
| 840 | + VaeImageProcessor, |
| 841 | + VaeImageProcessorLDM3D, |
| 842 | + ) |
837 | 843 | from .models import ( |
838 | 844 | AllegroTransformer3DModel, |
839 | 845 | AsymmetricAutoencoderKL, |
|
949 | 955 | ScoreSdeVePipeline, |
950 | 956 | StableDiffusionMixin, |
951 | 957 | ) |
952 | | - from .image_processor import ( |
953 | | - VaeImageProcessor, |
954 | | - VaeImageProcessorLDM3D, |
955 | | - PixArtImageProcessor, |
956 | | - IPAdapterMaskProcessor, |
957 | | - ) |
958 | | - from .video_processor import VideoProcessor |
959 | 958 | from .quantizers import DiffusersQuantizer |
960 | 959 | from .schedulers import ( |
961 | 960 | AmusedScheduler, |
|
997 | 996 | VQDiffusionScheduler, |
998 | 997 | ) |
999 | 998 | from .training_utils import EMAModel |
| 999 | + from .video_processor import VideoProcessor |
1000 | 1000 |
|
1001 | 1001 | try: |
1002 | 1002 | if not (is_torch_available() and is_scipy_available()): |
|
0 commit comments