@@ -398,6 +398,7 @@ def _check_sample_evaluations(
398398 assert evaluation < max_time_allowed
399399
400400 @pytest .mark .sklearn ()
401+ @pytest .mark .xfail (reason = "failures_issue_1544" )
401402 def test_run_regression_on_classif_task (self ):
402403 task_id = 259 # collins; crossvalidation; has numeric targets
403404
@@ -414,6 +415,7 @@ def test_run_regression_on_classif_task(self):
414415 )
415416
416417 @pytest .mark .sklearn ()
418+ @pytest .mark .xfail (reason = "failures_issue_1544" )
417419 def test_check_erronous_sklearn_flow_fails (self ):
418420 task_id = 115 # diabetes; crossvalidation
419421 task = openml .tasks .get_task (task_id )
@@ -627,6 +629,7 @@ def _run_and_upload_regression(
627629
628630 @pytest .mark .skip (reason = "failures_issue_1544" )
629631 @pytest .mark .sklearn ()
632+ @pytest .mark .xfail (reason = "failures_issue_1544" )
630633 def test_run_and_upload_logistic_regression (self ):
631634 lr = LogisticRegression (solver = "lbfgs" , max_iter = 1000 )
632635 task_id = self .TEST_SERVER_TASK_SIMPLE ["task_id" ]
@@ -636,6 +639,7 @@ def test_run_and_upload_logistic_regression(self):
636639
637640 @pytest .mark .skip (reason = "failures_issue_1544" )
638641 @pytest .mark .sklearn ()
642+ @pytest .mark .xfail (reason = "failures_issue_1544" )
639643 def test_run_and_upload_linear_regression (self ):
640644 lr = LinearRegression ()
641645 task_id = self .TEST_SERVER_TASK_REGRESSION ["task_id" ]
@@ -667,6 +671,7 @@ def test_run_and_upload_linear_regression(self):
667671
668672 @pytest .mark .skip (reason = "failures_issue_1544" )
669673 @pytest .mark .sklearn ()
674+ @pytest .mark .xfail (reason = "failures_issue_1544" )
670675 def test_run_and_upload_pipeline_dummy_pipeline (self ):
671676 pipeline1 = Pipeline (
672677 steps = [
@@ -681,6 +686,7 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
681686
682687 @pytest .mark .skip (reason = "failures_issue_1544" )
683688 @pytest .mark .sklearn ()
689+ @pytest .mark .xfail (reason = "failures_issue_1544" )
684690 @unittest .skipIf (
685691 Version (sklearn .__version__ ) < Version ("0.20" ),
686692 reason = "columntransformer introduction in 0.20.0" ,
@@ -799,6 +805,7 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
799805
800806 @pytest .mark .skip (reason = "failures_issue_1544" )
801807 @pytest .mark .sklearn ()
808+ @pytest .mark .xfail (reason = "failures_issue_1544" )
802809 def test_run_and_upload_gridsearch (self ):
803810 estimator_name = (
804811 "base_estimator" if Version (sklearn .__version__ ) < Version ("1.4" ) else "estimator"
@@ -821,7 +828,7 @@ def test_run_and_upload_gridsearch(self):
821828 assert len (run .trace .trace_iterations ) == 9
822829
823830 @pytest .mark .sklearn ()
824- @pytest .mark .skip (reason = "failures_issue_1544" )
831+ @pytest .mark .xfail (reason = "failures_issue_1544" )
825832 def test_run_and_upload_randomsearch (self ):
826833 randomsearch = RandomizedSearchCV (
827834 RandomForestClassifier (n_estimators = 5 ),
@@ -855,6 +862,7 @@ def test_run_and_upload_randomsearch(self):
855862
856863 @pytest .mark .skip (reason = "failures_issue_1544" )
857864 @pytest .mark .sklearn ()
865+ @pytest .mark .xfail (reason = "failures_issue_1544" )
858866 def test_run_and_upload_maskedarrays (self ):
859867 # This testcase is important for 2 reasons:
860868 # 1) it verifies the correct handling of masked arrays (not all
@@ -882,6 +890,7 @@ def test_run_and_upload_maskedarrays(self):
882890 ##########################################################################
883891
884892 @pytest .mark .sklearn ()
893+ @pytest .mark .xfail (reason = "failures_issue_1544" )
885894 def test_learning_curve_task_1 (self ):
886895 task_id = 801 # diabates dataset
887896 num_test_instances = 6144 # for learning curve
@@ -906,6 +915,7 @@ def test_learning_curve_task_1(self):
906915 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
907916
908917 @pytest .mark .sklearn ()
918+ @pytest .mark .xfail (reason = "failures_issue_1544" )
909919 def test_learning_curve_task_2 (self ):
910920 task_id = 801 # diabates dataset
911921 num_test_instances = 6144 # for learning curve
@@ -942,6 +952,7 @@ def test_learning_curve_task_2(self):
942952 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
943953
944954 @pytest .mark .sklearn ()
955+ @pytest .mark .xfail (reason = "failures_issue_1544" )
945956 @unittest .skipIf (
946957 Version (sklearn .__version__ ) < Version ("0.21" ),
947958 reason = "Pipelines don't support indexing (used for the assert check)" ,
@@ -1020,6 +1031,7 @@ def _test_local_evaluations(self, run):
10201031 assert alt_scores [idx ] <= 1
10211032
10221033 @pytest .mark .sklearn ()
1034+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10231035 def test_local_run_swapped_parameter_order_model (self ):
10241036 clf = DecisionTreeClassifier ()
10251037 australian_task = 595 # Australian; crossvalidation
@@ -1035,6 +1047,7 @@ def test_local_run_swapped_parameter_order_model(self):
10351047 self ._test_local_evaluations (run )
10361048
10371049 @pytest .mark .sklearn ()
1050+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10381051 @unittest .skipIf (
10391052 Version (sklearn .__version__ ) < Version ("0.20" ),
10401053 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1063,6 +1076,7 @@ def test_local_run_swapped_parameter_order_flow(self):
10631076 self ._test_local_evaluations (run )
10641077
10651078 @pytest .mark .sklearn ()
1079+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10661080 @unittest .skipIf (
10671081 Version (sklearn .__version__ ) < Version ("0.20" ),
10681082 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1100,6 +1114,7 @@ def test_online_run_metric_score(self):
11001114 self ._test_local_evaluations (run )
11011115
11021116 @pytest .mark .sklearn ()
1117+ @pytest .mark .xfail (reason = "failures_issue_1544" )
11031118 @unittest .skipIf (
11041119 Version (sklearn .__version__ ) < Version ("0.20" ),
11051120 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1165,6 +1180,7 @@ def test_initialize_model_from_run(self):
11651180 Version (sklearn .__version__ ) < Version ("0.20" ),
11661181 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
11671182 )
1183+ @pytest .mark .xfail (reason = "failures_issue_1544" )
11681184 def test__run_exists (self ):
11691185 # would be better to not sentinel these clfs,
11701186 # so we do not have to perform the actual runs
@@ -1220,6 +1236,7 @@ def test__run_exists(self):
12201236 assert run_ids , (run_ids , clf )
12211237
12221238 @pytest .mark .sklearn ()
1239+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12231240 def test_run_with_illegal_flow_id (self ):
12241241 # check the case where the user adds an illegal flow id to a
12251242 # non-existing flo
@@ -1239,6 +1256,7 @@ def test_run_with_illegal_flow_id(self):
12391256 )
12401257
12411258 @pytest .mark .sklearn ()
1259+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12421260 def test_run_with_illegal_flow_id_after_load (self ):
12431261 # Same as `test_run_with_illegal_flow_id`, but test this error is also
12441262 # caught if the run is stored to and loaded from disk first.
@@ -1270,6 +1288,7 @@ def test_run_with_illegal_flow_id_after_load(self):
12701288 TestBase .logger .info (f"collected from test_run_functions: { loaded_run .run_id } " )
12711289
12721290 @pytest .mark .sklearn ()
1291+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12731292 def test_run_with_illegal_flow_id_1 (self ):
12741293 # Check the case where the user adds an illegal flow id to an existing
12751294 # flow. Comes to a different value error than the previous test
@@ -1295,6 +1314,7 @@ def test_run_with_illegal_flow_id_1(self):
12951314 )
12961315
12971316 @pytest .mark .sklearn ()
1317+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12981318 def test_run_with_illegal_flow_id_1_after_load (self ):
12991319 # Same as `test_run_with_illegal_flow_id_1`, but test this error is
13001320 # also caught if the run is stored to and loaded from disk first.
@@ -1333,6 +1353,7 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13331353 )
13341354
13351355 @pytest .mark .sklearn ()
1356+ @pytest .mark .xfail (reason = "failures_issue_1544" )
13361357 @unittest .skipIf (
13371358 Version (sklearn .__version__ ) < Version ("0.20" ),
13381359 reason = "OneHotEncoder cannot handle mixed type DataFrame as input" ,
@@ -1560,6 +1581,7 @@ def test_get_runs_list_by_tag(self):
15601581 assert len (runs ) >= 1
15611582
15621583 @pytest .mark .sklearn ()
1584+ @pytest .mark .xfail (reason = "failures_issue_1544" )
15631585 @unittest .skipIf (
15641586 Version (sklearn .__version__ ) < Version ("0.20" ),
15651587 reason = "columntransformer introduction in 0.20.0" ,
@@ -1596,6 +1618,7 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
15961618 assert len (row ) == 12
15971619
15981620 @pytest .mark .sklearn ()
1621+ @pytest .mark .xfail (reason = "failures_issue_1544" )
15991622 @unittest .skipIf (
16001623 Version (sklearn .__version__ ) < Version ("0.20" ),
16011624 reason = "columntransformer introduction in 0.20.0" ,
@@ -1648,6 +1671,7 @@ def test_get_uncached_run(self):
16481671 openml .runs .functions ._get_cached_run (10 )
16491672
16501673 @pytest .mark .sklearn ()
1674+ @pytest .mark .xfail (reason = "failures_issue_1544" )
16511675 def test_run_flow_on_task_downloaded_flow (self ):
16521676 model = sklearn .ensemble .RandomForestClassifier (n_estimators = 33 )
16531677 flow = self .extension .model_to_flow (model )
@@ -1748,6 +1772,7 @@ def test_format_prediction_task_regression(self):
17481772 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
17491773 )
17501774 @pytest .mark .sklearn ()
1775+ @pytest .mark .xfail (reason = "failures_issue_1544" )
17511776 def test_delete_run (self ):
17521777 rs = np .random .randint (1 , 2 ** 31 - 1 )
17531778 clf = sklearn .pipeline .Pipeline (
@@ -1843,6 +1868,7 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18431868
18441869
18451870@pytest .mark .sklearn ()
1871+ @pytest .mark .xfail (reason = "failures_issue_1544" )
18461872@unittest .skipIf (
18471873 Version (sklearn .__version__ ) < Version ("0.21" ),
18481874 reason = "couldn't perform local tests successfully w/o bloating RAM" ,
@@ -1938,6 +1964,7 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19381964 (- 1 , "threading" , 10 ), # the threading backend does preserve mocks even with parallelizing
19391965 ]
19401966)
1967+ @pytest .mark .xfail (reason = "failures_issue_1544" )
19411968def test_joblib_backends (parallel_mock , n_jobs , backend , call_count ):
19421969 """Tests evaluation of a run using various joblib backends and n_jobs."""
19431970 if backend is None :
0 commit comments