|
17 | 17 | from huggingface_hub.utils import validate_hf_hub_args |
18 | 18 |
|
19 | 19 | from ..configuration_utils import ConfigMixin |
20 | | -from ..pipelines.audioldm2.modeling_audioldm2 import AudioLDM2ProjectionModel, AudioLDM2UNet2DConditionModel |
21 | | -from ..pipelines.deepfloyd_if.watermark import IFWatermarker |
22 | | -from ..pipelines.flux.modeling_flux import ReduxImageEncoder |
23 | | -from ..pipelines.shap_e.renderer import MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer |
24 | | -from ..pipelines.stable_audio.modeling_stable_audio import StableAudioProjectionModel |
25 | | -from ..pipelines.stable_diffusion.clip_image_project_model import CLIPImageProjection |
26 | | -from ..pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer |
27 | | -from ..pipelines.unclip.text_proj import UnCLIPTextProjModel |
28 | | -from ..pipelines.unidiffuser.modeling_uvit import UniDiffuserModel, UTransformer2DModel |
29 | | -from ..pipelines.wuerstchen.modeling_paella_vq_model import PaellaVQModel |
30 | | -from ..pipelines.wuerstchen.modeling_wuerstchen_diffnext import WuerstchenDiffNeXt |
31 | | -from ..pipelines.wuerstchen.modeling_wuerstchen_prior import WuerstchenPrior |
32 | | -from ..utils.import_utils import is_transformers_available |
33 | 20 | from .adapter import MultiAdapter, T2IAdapter |
34 | 21 | from .autoencoders.autoencoder_asym_kl import AsymmetricAutoencoderKL |
35 | 22 | from .autoencoders.autoencoder_dc import AutoencoderDC |
|
90 | 77 | from .unets.uvit_2d import UVit2DModel |
91 | 78 |
|
92 | 79 |
|
93 | | -if is_transformers_available(): |
94 | | - from ..pipelines.unidiffuser.modeling_text_decoder import UniDiffuserTextDecoder |
95 | | - |
96 | 80 | AUTO_MODEL_MAPPING = OrderedDict( |
97 | 81 | [ |
98 | 82 | ("multi-adapter", MultiAdapter), |
|
159 | 143 | ("unet-spatio-temporal", UNetSpatioTemporalConditionModel), |
160 | 144 | ("stable-cascade-unet", StableCascadeUNet), |
161 | 145 | ("uvit-2d", UVit2DModel), |
162 | | - ("audioldm2-projection", AudioLDM2ProjectionModel), |
163 | | - ("audioldm2-unet-2d", AudioLDM2UNet2DConditionModel), |
164 | | - ("if-watermarker", IFWatermarker), |
165 | | - ("redux-image-encoder", ReduxImageEncoder), |
166 | | - ("mlp-nerstf", MLPNeRSTFModel), |
167 | | - ("shap-e-params-proj", ShapEParamsProjModel), |
168 | | - ("shap-e-renderer", ShapERenderer), |
169 | | - ("stable-audio-projection", StableAudioProjectionModel), |
170 | | - ("clip-image-projection", CLIPImageProjection), |
171 | | - ("stable-unclip-image-normalizer", StableUnCLIPImageNormalizer), |
172 | | - ("unclip-text-proj", UnCLIPTextProjModel), |
173 | | - ("utransformer-2d", UTransformer2DModel), |
174 | | - ("unidiffuser", UniDiffuserModel), |
175 | | - ("paella-vq", PaellaVQModel), |
176 | | - ("wuerstchen-diffnext", WuerstchenDiffNeXt), |
177 | | - ("wuerstchen-prior", WuerstchenPrior), |
178 | 146 | ] |
179 | 147 | ) |
180 | 148 |
|
181 | | - |
182 | | -if is_transformers_available(): |
183 | | - AUTO_MODEL_MAPPING["unidiffuser-text-decoder"] = UniDiffuserTextDecoder |
184 | | - |
185 | | - |
186 | 149 | SUPPORTED_TASKS_MAPPINGS = [AUTO_MODEL_MAPPING] |
187 | 150 |
|
188 | 151 |
|
|
0 commit comments