@@ -111,6 +111,7 @@ def _register_attention_processors_metadata():
111111 from ..models .transformers .transformer_hunyuanimage import HunyuanImageAttnProcessor
112112 from ..models .transformers .transformer_qwenimage import QwenDoubleStreamAttnProcessor2_0
113113 from ..models .transformers .transformer_wan import WanAttnProcessor2_0
114+ from ..models .transformers .transformer_z_image import ZSingleStreamAttnProcessor
114115
115116 # AttnProcessor2_0
116117 AttentionProcessorRegistry .register (
@@ -158,6 +159,14 @@ def _register_attention_processors_metadata():
158159 ),
159160 )
160161
162+ # ZSingleStreamAttnProcessor
163+ AttentionProcessorRegistry .register (
164+ model_class = ZSingleStreamAttnProcessor ,
165+ metadata = AttentionProcessorMetadata (
166+ skip_processor_output_fn = _skip_proc_output_fn_Attention_ZSingleStreamAttnProcessor ,
167+ ),
168+ )
169+
161170
162171def _register_transformer_blocks_metadata ():
163172 from ..models .attention import BasicTransformerBlock
@@ -179,6 +188,7 @@ def _register_transformer_blocks_metadata():
179188 from ..models .transformers .transformer_mochi import MochiTransformerBlock
180189 from ..models .transformers .transformer_qwenimage import QwenImageTransformerBlock
181190 from ..models .transformers .transformer_wan import WanTransformerBlock
191+ from ..models .transformers .transformer_z_image import ZImageTransformerBlock
182192
183193 # BasicTransformerBlock
184194 TransformerBlockRegistry .register (
@@ -312,6 +322,15 @@ def _register_transformer_blocks_metadata():
312322 ),
313323 )
314324
325+ # ZImage
326+ TransformerBlockRegistry .register (
327+ model_class = ZImageTransformerBlock ,
328+ metadata = TransformerBlockMetadata (
329+ return_hidden_states_index = 0 ,
330+ return_encoder_hidden_states_index = None ,
331+ ),
332+ )
333+
315334
316335# fmt: off
317336def _skip_attention___ret___hidden_states (self , * args , ** kwargs ):
@@ -338,4 +357,5 @@ def _skip_attention___ret___hidden_states___encoder_hidden_states(self, *args, *
338357_skip_proc_output_fn_Attention_FluxAttnProcessor = _skip_attention___ret___hidden_states
339358_skip_proc_output_fn_Attention_QwenDoubleStreamAttnProcessor2_0 = _skip_attention___ret___hidden_states
340359_skip_proc_output_fn_Attention_HunyuanImageAttnProcessor = _skip_attention___ret___hidden_states
360+ _skip_proc_output_fn_Attention_ZSingleStreamAttnProcessor = _skip_attention___ret___hidden_states
341361# fmt: on
0 commit comments