@@ -95,10 +95,11 @@ def copy(self):
9595 def _update_members (self , node = None , attr = None ):
9696 "Updates internal members."
9797 if node is None or attr is None :
98- if len (self .nodes_attr ) == 0 :
99- self .size_ = self .dim
100- else :
101- self .size_ = max (d ["output" ] for d in self .nodes_attr ) + 1
98+ self .size_ = (
99+ self .dim
100+ if not self .nodes_attr
101+ else (max (d ["output" ] for d in self .nodes_attr ) + 1 )
102+ )
102103 self .output_to_node_ = {}
103104 self .input_to_node_ = {}
104105 for node2 , attr2 in zip (self .nodes , self .nodes_attr ):
@@ -150,7 +151,7 @@ def append(self, node, inputs):
150151 self .nodes .append (node )
151152 first_coef = (
152153 0
153- if len ( self .nodes_attr ) == 0
154+ if not self .nodes_attr
154155 else self .nodes_attr [- 1 ]["first_coef" ]
155156 + self .nodes_attr [- 1 ]["coef_size" ]
156157 )
@@ -173,7 +174,7 @@ def append(self, node, inputs):
173174 self .nodes .append (node )
174175 first_coef = (
175176 0
176- if len ( self .nodes_attr ) == 0
177+ if not self .nodes_attr
177178 else self .nodes_attr [- 1 ]["first_coef" ]
178179 + self .nodes_attr [- 1 ]["coef_size" ]
179180 )
@@ -702,7 +703,7 @@ def gradient_backward(self, graddx, X, inputs=False, cache=None):
702703
703704 whole_gradx = numpy .zeros (pred .shape , dtype = numpy .float64 )
704705 whole_gradw = numpy .zeros (shape , dtype = numpy .float64 )
705- if len ( graddx .shape ) == 0 :
706+ if not graddx .shape :
706707 whole_gradx [- 1 ] = graddx
707708 else :
708709 whole_gradx [- graddx .shape [0 ] :] = graddx
0 commit comments