@@ -65,15 +65,11 @@ def rewrite_constant_fold(g, ops):
65
65
if set (op .output ) & set (g .outputs ): continue
66
66
try :
67
67
inputs = []
68
- skip = False
69
68
for node in op .inputs :
70
69
if not node .is_const ():
71
- skip = True
72
70
break
73
71
inputs .append (node .get_tensor_value (as_list = False ))
74
72
75
- if skip : continue
76
-
77
73
logger .debug ("op name %s, %s, %s" , op .name , len (op .input ), len (inputs ))
78
74
if inputs and len (op .input ) == len (inputs ):
79
75
logger .info ("folding node type=%s, name=%s" % (op .type , op .name ))
@@ -453,14 +449,14 @@ def compat_handler(ctx, node, **kwargs):
453
449
454
450
# pre-processing graph rewrites
455
451
# bi-directional re-writer should be placed after single directional re-writer
456
- rewriters = [rewrite_quantize_and_dequantize , rewrite_transpose , rewrite_flatten , rewrite_gemm ,
457
- rewrite_random_uniform , rewrite_random_uniform_fold_const ,
452
+ rewriters = [rewrite_constant_fold , rewrite_quantize_and_dequantize , rewrite_transpose , rewrite_flatten ,
453
+ rewrite_gemm , rewrite_random_uniform , rewrite_random_uniform_fold_const ,
458
454
rewrite_random_normal , rewrite_dropout , rewrite_eye ,
459
455
rewrite_leakyrelu , rewrite_thresholded_relu , rewrite_conv2d_with_pad ,
460
456
rewrite_single_direction_lstm , rewrite_bi_direction_lstm ,
461
457
rewrite_single_direction_gru , rewrite_bi_direction_gru ,
462
458
rewrite_custom_rnn_cell , rewrite_generic_loop , rewrite_cond ,
463
- rewrite_biasadd_with_conv2d , rewrite_constant_fold
459
+ rewrite_biasadd_with_conv2d ,
464
460
]
465
461
466
462
if custom_rewriter is not None :
0 commit comments