Skip to content

Commit 795c990

Browse files
zhang-cheng09chanchzhang
authored andcommitted
add unit tests and functional tests
Signed-off-by: chanchzhang <chanchzhang@tencent.com>
1 parent d523b4d commit 795c990

File tree

15 files changed

+2401
-308
lines changed

15 files changed

+2401
-308
lines changed

src/megatron/bridge/models/qwen_omni/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
from megatron.bridge.models.qwen_omni.modelling_qwen3_omni import Qwen3OmniMoeModel
15+
from megatron.bridge.models.qwen_omni.modeling_qwen3_omni.model import Qwen3OmniMoeModel
1616
from megatron.bridge.models.qwen_omni.qwen3_omni_bridge import Qwen3OmniMoeBridge
1717
from megatron.bridge.models.qwen_omni.qwen3_omni_provider import Qwen3OmniMoeModelProvider
1818

src/megatron/bridge/models/qwen_omni/context_parallel_utils.py

Lines changed: 0 additions & 227 deletions
This file was deleted.

src/megatron/bridge/models/qwen_omni/modelling_qwen3_omni.py renamed to src/megatron/bridge/models/qwen_omni/modeling_qwen3_omni/model.py

Lines changed: 12 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -26,8 +26,8 @@
2626
Qwen3OmniMoeCode2WavConfig,
2727
)
2828

29-
from megatron.bridge.models.qwen_omni.thinker_model import Qwen3OmniMoeThinkerModel
30-
from megatron.bridge.models.qwen_omni.transformer_config import Qwen3OmniTransformerConfig
29+
from megatron.bridge.models.qwen_omni.modeling_qwen3_omni.thinker_model import Qwen3OmniMoeThinkerModel
30+
from megatron.bridge.models.qwen_omni.modeling_qwen3_omni.transformer_config import Qwen3OmniTransformerConfig
3131

3232

3333
class Qwen3OmniMoeModel(MegatronModule):
@@ -73,23 +73,26 @@ def set_input_tensor(self, input_tensor) -> None:
7373

7474
def freeze(
7575
self,
76-
freeze_language_model: bool,
77-
freeze_vision_model: bool,
78-
freeze_vision_projection: bool,
76+
freeze_language_model: bool=False,
77+
freeze_vision_model: bool=False,
78+
freeze_vision_projection: bool=False,
79+
freeze_audio_model: bool=False,
7980
):
8081
"""Freeze model modules.
8182
8283
Make specific modules non-trainable by setting requires_grad to False.
8384
8485
Args:
8586
freeze_language_model (bool): Freeze the language model module.
86-
freeze_vision_model (bool): Freeze the vision model module (patch_embed, blocks, pos_embed).
87+
freeze_vision_model (bool): Freeze the vision model module.
8788
freeze_vision_projection (bool): Freeze the vision projection modules (merger and deepstack_merger_list).
89+
freeze_audio_model (bool): Freeze the audio model module.
8890
"""
8991
return self.thinker.freeze(
9092
freeze_language_model,
9193
freeze_vision_model,
92-
freeze_vision_projection
94+
freeze_vision_projection,
95+
freeze_audio_model,
9396
)
9497

9598
def forward(
@@ -113,6 +116,7 @@ def forward(
113116
feature_attention_mask=None,
114117
audio_feature_lengths=None,
115118
cp_img_num: list[int] = None,
119+
images_padded: list[bool] = None,
116120
use_audio_in_video=None,
117121
video_second_per_grid=None,
118122
**kwargs,
@@ -136,6 +140,7 @@ def forward(
136140
feature_attention_mask=feature_attention_mask,
137141
audio_feature_lengths=audio_feature_lengths,
138142
cp_img_num=cp_img_num,
143+
images_padded=images_padded,
139144
use_audio_in_video=use_audio_in_video,
140145
video_second_per_grid=video_second_per_grid,
141146
**kwargs,

src/megatron/bridge/models/qwen_omni/utils.py renamed to src/megatron/bridge/models/qwen_omni/modeling_qwen3_omni/rope.py

File renamed without changes.

0 commit comments

Comments
 (0)