File tree Expand file tree Collapse file tree 9 files changed +84
-52
lines changed
model_zoo/gpt-3/ppfleetx/models/language_model Expand file tree Collapse file tree 9 files changed +84
-52
lines changed Original file line number Diff line number Diff line change 48
48
MinLengthLogitsProcessor ,
49
49
RepetitionPenaltyLogitsProcessor ,
50
50
)
51
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
52
- ColumnSequenceParallelLinear ,
53
- GatherOp ,
54
- RowSequenceParallelLinear ,
55
- ScatterOp ,
56
- mark_as_sequence_parallel_parameter ,
57
- )
51
+ try :
52
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
53
+ ColumnSequenceParallelLinear ,
54
+ GatherOp ,
55
+ RowSequenceParallelLinear ,
56
+ ScatterOp ,
57
+ mark_as_sequence_parallel_parameter ,
58
+ )
59
+ except :
60
+ pass
58
61
59
62
from paddlenlp .transformers .segment_parallel_utils import ReshardLayer
60
63
Original file line number Diff line number Diff line change 24
24
from ppfleetx .core .module .basic_module import BasicModule
25
25
from ppfleetx .data .tokenizers import GPTTokenizer
26
26
from ppfleetx .distributed .apis import env
27
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
28
- register_sequence_parallel_allreduce_hooks ,
29
- )
27
+ try :
28
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
29
+ register_sequence_parallel_allreduce_hooks ,
30
+ )
31
+ except :
32
+ pass
30
33
from ppfleetx .utils .log import logger
31
34
32
35
# TODO(haohongxiang): to solve the problem of cross-reference
Original file line number Diff line number Diff line change 29
29
from .feature_extraction_utils import BatchFeature , FeatureExtractionMixin
30
30
from .image_processing_utils import ImageProcessingMixin
31
31
from .attention_utils import create_bigbird_rand_mask_idx_list
32
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
33
- GatherOp ,
34
- ScatterOp ,
35
- AllGatherOp ,
36
- ReduceScatterOp ,
37
- ColumnSequenceParallelLinear ,
38
- RowSequenceParallelLinear ,
39
- mark_as_sequence_parallel_parameter ,
40
- register_sequence_parallel_allreduce_hooks ,
41
- )
32
+
33
+ try :
34
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
35
+ GatherOp ,
36
+ ScatterOp ,
37
+ AllGatherOp ,
38
+ ReduceScatterOp ,
39
+ ColumnSequenceParallelLinear ,
40
+ RowSequenceParallelLinear ,
41
+ mark_as_sequence_parallel_parameter ,
42
+ register_sequence_parallel_allreduce_hooks ,
43
+ )
44
+ except :
45
+ pass
42
46
from .export import export_model
43
47
44
48
# isort: split
Original file line number Diff line number Diff line change 29
29
from paddle .distributed import fleet
30
30
from paddle .distributed .fleet .meta_parallel import get_rng_state_tracker
31
31
from paddle .distributed .fleet .utils import recompute
32
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
33
- ColumnSequenceParallelLinear ,
34
- GatherOp ,
35
- RowSequenceParallelLinear ,
36
- ScatterOp ,
37
- mark_as_sequence_parallel_parameter ,
38
- )
32
+
33
+ try :
34
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
35
+ ColumnSequenceParallelLinear ,
36
+ GatherOp ,
37
+ RowSequenceParallelLinear ,
38
+ ScatterOp ,
39
+ mark_as_sequence_parallel_parameter ,
40
+ )
41
+ except :
42
+ pass
39
43
from paddle .nn import BCEWithLogitsLoss , CrossEntropyLoss , MSELoss
40
44
from paddle .utils import try_import
41
45
Original file line number Diff line number Diff line change 30
30
from paddle .distributed import fleet
31
31
from paddle .distributed .fleet .meta_parallel import get_rng_state_tracker
32
32
from paddle .distributed .fleet .utils import recompute
33
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
34
- ScatterOp ,
35
- mark_as_sequence_parallel_parameter ,
36
- )
33
+
34
+ try :
35
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
36
+ ScatterOp ,
37
+ mark_as_sequence_parallel_parameter ,
38
+ )
39
+ except :
40
+ pass
37
41
38
42
from ...utils .converter import StateDictNameMapping
39
43
from .. import PretrainedModel , register_base_model
Original file line number Diff line number Diff line change 19
19
SharedLayerDesc ,
20
20
)
21
21
from paddle .distributed .fleet .utils import recompute
22
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
23
- mark_as_sequence_parallel_parameter ,
24
- )
22
+
23
+ try :
24
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
25
+ mark_as_sequence_parallel_parameter ,
26
+ )
27
+ except :
28
+ pass
25
29
26
30
from paddlenlp .transformers .model_utils import PipelinePretrainedModel
27
31
Original file line number Diff line number Diff line change @@ -45,13 +45,16 @@ def swiglu(x, y=None):
45
45
return F .silu (x ) * y
46
46
47
47
48
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
49
- ColumnSequenceParallelLinear ,
50
- GatherOp ,
51
- RowSequenceParallelLinear ,
52
- ScatterOp ,
53
- mark_as_sequence_parallel_parameter ,
54
- )
48
+ try :
49
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
50
+ ColumnSequenceParallelLinear ,
51
+ GatherOp ,
52
+ RowSequenceParallelLinear ,
53
+ ScatterOp ,
54
+ mark_as_sequence_parallel_parameter ,
55
+ )
56
+ except :
57
+ pass
55
58
from paddle .utils import try_import
56
59
57
60
from paddlenlp .transformers .conversion_utils import (
Original file line number Diff line number Diff line change 23
23
24
24
from paddle import distributed as dist
25
25
from paddle .autograd import PyLayer
26
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
27
- ColumnSequenceParallelLinear ,
28
- RowSequenceParallelLinear ,
29
- )
26
+
27
+ try :
28
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
29
+ ColumnSequenceParallelLinear ,
30
+ RowSequenceParallelLinear ,
31
+ )
32
+ except :
33
+ pass
30
34
31
35
__all_gather_recomputation__ = False
32
36
if int (os .getenv ("MC2_Recompute" , 0 )):
Original file line number Diff line number Diff line change 33
33
except ImportError :
34
34
fused_rotary_position_embedding = None
35
35
36
- from paddle .distributed .fleet .utils .sequence_parallel_utils import (
37
- ColumnSequenceParallelLinear ,
38
- GatherOp ,
39
- RowSequenceParallelLinear ,
40
- ScatterOp ,
41
- mark_as_sequence_parallel_parameter ,
42
- )
36
+ try :
37
+ from paddle .distributed .fleet .utils .sequence_parallel_utils import (
38
+ ColumnSequenceParallelLinear ,
39
+ GatherOp ,
40
+ RowSequenceParallelLinear ,
41
+ ScatterOp ,
42
+ mark_as_sequence_parallel_parameter ,
43
+ )
44
+ except :
45
+ pass
43
46
44
47
from paddlenlp .transformers .conversion_utils import (
45
48
StateDictNameMapping ,
You can’t perform that action at this time.
0 commit comments