66class TestEvaluationFunctions (TestBase ):
77 _multiprocess_can_split_ = True
88
9- def _check_list_evaluation_setups (self , size , ** kwargs ):
9+ def _check_list_evaluation_setups (self , ** kwargs ):
1010 evals_setups = openml .evaluations .list_evaluations_setups ("predictive_accuracy" ,
11- ** kwargs , size = size ,
11+ ** kwargs ,
1212 sort_order = 'desc' ,
1313 output_format = 'dataframe' )
1414 evals = openml .evaluations .list_evaluations ("predictive_accuracy" ,
15- ** kwargs , size = size ,
15+ ** kwargs ,
1616 sort_order = 'desc' ,
1717 output_format = 'dataframe' )
1818
1919 # Check if list is non-empty
2020 self .assertGreater (len (evals_setups ), 0 )
21+ # Check if length is accurate
22+ self .assertEqual (len (evals_setups ), len (evals ))
2123 # Check if output from sort is sorted in the right order
2224 self .assertSequenceEqual (sorted (evals_setups ['value' ].tolist (), reverse = True ),
2325 evals_setups ['value' ].tolist ())
@@ -176,7 +178,7 @@ def test_list_evaluations_setups_filter_flow(self):
176178 openml .config .server = self .production_server
177179 flow_id = [405 ]
178180 size = 100
179- evals = self ._check_list_evaluation_setups (size , flow = flow_id )
181+ evals = self ._check_list_evaluation_setups (flow = flow_id , size = size )
180182 # check if parameters in separate columns works
181183 evals_cols = openml .evaluations .list_evaluations_setups ("predictive_accuracy" ,
182184 flow = flow_id , size = size ,
@@ -191,5 +193,5 @@ def test_list_evaluations_setups_filter_flow(self):
191193 def test_list_evaluations_setups_filter_task (self ):
192194 openml .config .server = self .production_server
193195 task_id = [6 ]
194- size = 100
195- self ._check_list_evaluation_setups (size , task = task_id )
196+ size = 121
197+ self ._check_list_evaluation_setups (task = task_id , size = size )
0 commit comments