Skip to content

Commit d83c633

Browse files
authored
[hotfix] Fix examples no pad token & auto parallel codegen bug; (#5606)
* fix no pad token bug * fixed some auto parallel codegen bug, but might not run on torch 2.1 --------- Co-authored-by: Edenzzzz <[email protected]>
1 parent a0ad587 commit d83c633

File tree

6 files changed

+8
-5
lines changed

6 files changed

+8
-5
lines changed

colossalai/_analyzer/fx/codegen.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -246,7 +246,7 @@ def emit_code_with_activation_checkpoint(body, ckpt_func, nodes, emit_node_func,
246246

247247
@compatibility(is_backward_compatible=True)
248248
class ActivationCheckpointCodeGen(CodeGen):
249-
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace) -> PythonCode:
249+
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace, verbose=None) -> PythonCode:
250250
free_vars: List[str] = []
251251
body: List[str] = []
252252
globals_: Dict[str, Any] = {}

colossalai/auto_parallel/offload/base_offload_module.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
import torch.nn as nn
66

77
from colossalai.utils import _cast_float
8-
from colossalai.zero.legacy.gemini.tensor_utils import free_storage
8+
from colossalai.utils.common import free_storage
99

1010
from .region_manager import RegionManager
1111
from .util import GlobalRuntimeInfo

colossalai/auto_parallel/offload/region.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,8 @@
33
import torch
44
from torch.fx import Node
55

6-
from colossalai.zero.legacy.gemini.tensor_utils import alloc_storage, free_storage
6+
from colossalai.utils.common import free_storage
7+
from colossalai.zero.gemini.chunk.chunk import alloc_storage
78

89

910
class Region:

colossalai/autochunk/autochunk_codegen.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -372,7 +372,7 @@ def __init__(
372372
if print_progress:
373373
get_logger().info("AutoChunk start codegen")
374374

375-
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace) -> PythonCode:
375+
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace, verbose=None) -> PythonCode:
376376
free_vars: List[str] = []
377377
body: List[str] = []
378378
globals_: Dict[str, Any] = {}

colossalai/fx/codegen/activation_checkpoint_codegen.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -625,7 +625,7 @@ def emit_code_with_activation_checkpoint(body, ckpt_func, nodes, emit_node_func,
625625
if CODEGEN_AVAILABLE:
626626

627627
class ActivationCheckpointCodeGen(CodeGen):
628-
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace) -> PythonCode:
628+
def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace, verbose=None) -> PythonCode:
629629
free_vars: List[str] = []
630630
body: List[str] = []
631631
globals_: Dict[str, Any] = {}

examples/language/gpt/hybridparallelism/data.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -62,6 +62,8 @@ def __init__(
6262
self.text_fields = self.task_text_field_map[task_name]
6363
self.num_labels = self.glue_task_num_labels[task_name]
6464
self.tokenizer: PreTrainedTokenizer = AutoTokenizer.from_pretrained(self.model_name_or_path, use_fast=True)
65+
if not getattr(self.tokenizer, "pad_token", None):
66+
self.tokenizer.pad_token = self.tokenizer._eos_token
6567
self.setup()
6668

6769
def setup(self):

0 commit comments

Comments
 (0)