Skip to content

Commit 9774858

Browse files
committed
Fix Gemma3N audio notebooks to use reentrant checkpointing
Non-reentrant gradient checkpointing (use_reentrant=False) causes AOT autograd stride assertion failures during backward pass with Gemma3N audio conformer due to variable-length audio tensors: AssertionError: expected size 2==2, stride 1928==1936 at dim=0 Switch to use_reentrant=True which avoids AOT autograd compilation of the backward pass entirely. Companion to unslothai/unsloth#4629 which adds a server-side guard.
1 parent 7b8ebad commit 9774858

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

nb/Gemma3N_(4B)-Audio.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -646,8 +646,8 @@
646646
" args = SFTConfig(\n",
647647
" per_device_train_batch_size = 2,\n",
648648
" gradient_accumulation_steps = 1,\n",
649-
" # use reentrant checkpointing\n",
650-
" gradient_checkpointing_kwargs = {\"use_reentrant\": False},\n",
649+
" # Gemma3N audio requires reentrant checkpointing\n",
650+
" gradient_checkpointing_kwargs = {\"use_reentrant\": True},\n",
651651
" warmup_ratio = 0.1,\n",
652652
" #max_steps = 60,\n",
653653
" num_train_epochs = 1, # Set this instead of max_steps for full training runs\n",

nb/Kaggle-Gemma3N_(4B)-Audio.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -646,8 +646,8 @@
646646
" args = SFTConfig(\n",
647647
" per_device_train_batch_size = 2,\n",
648648
" gradient_accumulation_steps = 1,\n",
649-
" # use reentrant checkpointing\n",
650-
" gradient_checkpointing_kwargs = {\"use_reentrant\": False},\n",
649+
" # Gemma3N audio requires reentrant checkpointing\n",
650+
" gradient_checkpointing_kwargs = {\"use_reentrant\": True},\n",
651651
" warmup_ratio = 0.1,\n",
652652
" #max_steps = 60,\n",
653653
" num_train_epochs = 1, # Set this instead of max_steps for full training runs\n",

0 commit comments

Comments
 (0)