1313
1414class TestEvaluator (AbstractEvaluator ):
1515
16- def __init__ (self , Datamanager , output_dir ,
16+ def __init__ (self , Datamanager , backend ,
1717 configuration = None ,
1818 with_predictions = False ,
1919 all_scoring_functions = False ,
2020 seed = 1 ):
2121 super (TestEvaluator , self ).__init__ (
22- Datamanager , output_dir , configuration ,
22+ Datamanager , backend , configuration ,
2323 with_predictions = with_predictions ,
2424 all_scoring_functions = all_scoring_functions ,
2525 seed = seed ,
@@ -71,16 +71,19 @@ def predict_and_loss(self, train=False):
7171
7272# create closure for evaluating an algorithm
7373# Has a stupid name so nosetests doesn't regard it as a test
74- def eval_t (queue , config , data , tmp_dir , seed , num_run , subsample ,
74+ def eval_t (queue , config , data , backend , seed , num_run , subsample ,
7575 with_predictions , all_scoring_functions ,
7676 output_y_test ):
77- evaluator = TestEvaluator (data , tmp_dir , config ,
78- seed = seed , with_predictions = with_predictions ,
77+ evaluator = TestEvaluator (Datamanager = data , configuration = config ,
78+ backend = backend , seed = seed ,
79+ with_predictions = with_predictions ,
7980 all_scoring_functions = all_scoring_functions )
8081
8182 loss , opt_pred , valid_pred , test_pred = evaluator .fit_predict_and_loss ()
8283 duration , result , seed , run_info = evaluator .finish_up (
8384 loss , opt_pred , valid_pred , test_pred , file_output = False )
8485
8586 status = StatusType .SUCCESS
86- queue .put ((duration , result , seed , run_info , status ))
87+ queue .put ((duration , result , seed , run_info , status ))
88+
89+
0 commit comments