@@ -221,10 +221,10 @@ public void DecideAction()
221221 runner . AddInput ( graph [ graphScope + ObservationPlaceholderName [ obs_number ] ] [ 0 ] , observationMatrixList [ obs_number ] ) ;
222222 }
223223
224- TFTensor [ ] runned ;
224+ TFTensor [ ] networkOutput ;
225225 try
226226 {
227- runned = runner . Run ( ) ;
227+ networkOutput = runner . Run ( ) ;
228228 }
229229 catch ( TFException e )
230230 {
@@ -249,7 +249,7 @@ public void DecideAction()
249249
250250 runner . AddInput ( graph [ graphScope + RecurrentInPlaceholderName ] [ 0 ] , inputOldMemories ) ;
251251 runner . Fetch ( graph [ graphScope + RecurrentOutPlaceholderName ] [ 0 ] ) ;
252- float [ , ] recurrent_tensor = runned [ 1 ] . GetValue ( ) as float [ , ] ;
252+ float [ , ] recurrent_tensor = networkOutput [ 1 ] . GetValue ( ) as float [ , ] ;
253253
254254 int i = 0 ;
255255 foreach ( int k in agentKeys )
@@ -270,7 +270,7 @@ public void DecideAction()
270270
271271 if ( brain . brainParameters . actionSpaceType == StateType . continuous )
272272 {
273- float [ , ] output = runned [ 0 ] . GetValue ( ) as float [ , ] ;
273+ float [ , ] output = networkOutput [ 0 ] . GetValue ( ) as float [ , ] ;
274274 int i = 0 ;
275275 foreach ( int k in agentKeys )
276276 {
@@ -285,7 +285,7 @@ public void DecideAction()
285285 }
286286 else if ( brain . brainParameters . actionSpaceType == StateType . discrete )
287287 {
288- long [ , ] output = runned [ 0 ] . GetValue ( ) as long [ , ] ;
288+ long [ , ] output = networkOutput [ 0 ] . GetValue ( ) as long [ , ] ;
289289 int i = 0 ;
290290 foreach ( int k in agentKeys )
291291 {
0 commit comments