@@ -398,7 +398,6 @@ def _check_sample_evaluations(
398398 assert evaluation < max_time_allowed
399399
400400 @pytest .mark .sklearn ()
401- @pytest .mark .xfail (reason = "failures_issue_1544" )
402401 def test_run_regression_on_classif_task (self ):
403402 task_id = 259 # collins; crossvalidation; has numeric targets
404403
@@ -415,7 +414,6 @@ def test_run_regression_on_classif_task(self):
415414 )
416415
417416 @pytest .mark .sklearn ()
418- @pytest .mark .xfail (reason = "failures_issue_1544" )
419417 def test_check_erronous_sklearn_flow_fails (self ):
420418 task_id = 115 # diabetes; crossvalidation
421419 task = openml .tasks .get_task (task_id )
@@ -628,7 +626,6 @@ def _run_and_upload_regression(
628626 )
629627
630628 @pytest .mark .sklearn ()
631- @pytest .mark .xfail (reason = "failures_issue_1544" )
632629 def test_run_and_upload_logistic_regression (self ):
633630 lr = LogisticRegression (solver = "lbfgs" , max_iter = 1000 )
634631 task_id = self .TEST_SERVER_TASK_SIMPLE ["task_id" ]
@@ -637,7 +634,6 @@ def test_run_and_upload_logistic_regression(self):
637634 self ._run_and_upload_classification (lr , task_id , n_missing_vals , n_test_obs , "62501" )
638635
639636 @pytest .mark .sklearn ()
640- @pytest .mark .xfail (reason = "failures_issue_1544" )
641637 def test_run_and_upload_linear_regression (self ):
642638 lr = LinearRegression ()
643639 task_id = self .TEST_SERVER_TASK_REGRESSION ["task_id" ]
@@ -668,7 +664,6 @@ def test_run_and_upload_linear_regression(self):
668664 self ._run_and_upload_regression (lr , task_id , n_missing_vals , n_test_obs , "62501" )
669665
670666 @pytest .mark .sklearn ()
671- @pytest .mark .xfail (reason = "failures_issue_1544" )
672667 def test_run_and_upload_pipeline_dummy_pipeline (self ):
673668 pipeline1 = Pipeline (
674669 steps = [
@@ -682,7 +677,6 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
682677 self ._run_and_upload_classification (pipeline1 , task_id , n_missing_vals , n_test_obs , "62501" )
683678
684679 @pytest .mark .sklearn ()
685- @pytest .mark .xfail (reason = "failures_issue_1544" )
686680 @unittest .skipIf (
687681 Version (sklearn .__version__ ) < Version ("0.20" ),
688682 reason = "columntransformer introduction in 0.20.0" ,
@@ -799,7 +793,6 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
799793 assert call_count == 3
800794
801795 @pytest .mark .sklearn ()
802- @pytest .mark .xfail (reason = "failures_issue_1544" )
803796 def test_run_and_upload_gridsearch (self ):
804797 estimator_name = (
805798 "base_estimator" if Version (sklearn .__version__ ) < Version ("1.4" ) else "estimator"
@@ -822,7 +815,6 @@ def test_run_and_upload_gridsearch(self):
822815 assert len (run .trace .trace_iterations ) == 9
823816
824817 @pytest .mark .sklearn ()
825- @pytest .mark .xfail (reason = "failures_issue_1544" )
826818 def test_run_and_upload_randomsearch (self ):
827819 randomsearch = RandomizedSearchCV (
828820 RandomForestClassifier (n_estimators = 5 ),
@@ -855,7 +847,6 @@ def test_run_and_upload_randomsearch(self):
855847 assert len (trace .trace_iterations ) == 5
856848
857849 @pytest .mark .sklearn ()
858- @pytest .mark .xfail (reason = "failures_issue_1544" )
859850 def test_run_and_upload_maskedarrays (self ):
860851 # This testcase is important for 2 reasons:
861852 # 1) it verifies the correct handling of masked arrays (not all
@@ -883,7 +874,6 @@ def test_run_and_upload_maskedarrays(self):
883874 ##########################################################################
884875
885876 @pytest .mark .sklearn ()
886- @pytest .mark .xfail (reason = "failures_issue_1544" )
887877 def test_learning_curve_task_1 (self ):
888878 task_id = 801 # diabates dataset
889879 num_test_instances = 6144 # for learning curve
@@ -908,7 +898,6 @@ def test_learning_curve_task_1(self):
908898 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
909899
910900 @pytest .mark .sklearn ()
911- @pytest .mark .xfail (reason = "failures_issue_1544" )
912901 def test_learning_curve_task_2 (self ):
913902 task_id = 801 # diabates dataset
914903 num_test_instances = 6144 # for learning curve
@@ -945,7 +934,6 @@ def test_learning_curve_task_2(self):
945934 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
946935
947936 @pytest .mark .sklearn ()
948- @pytest .mark .xfail (reason = "failures_issue_1544" )
949937 @unittest .skipIf (
950938 Version (sklearn .__version__ ) < Version ("0.21" ),
951939 reason = "Pipelines don't support indexing (used for the assert check)" ,
@@ -1024,7 +1012,6 @@ def _test_local_evaluations(self, run):
10241012 assert alt_scores [idx ] <= 1
10251013
10261014 @pytest .mark .sklearn ()
1027- @pytest .mark .xfail (reason = "failures_issue_1544" )
10281015 def test_local_run_swapped_parameter_order_model (self ):
10291016 clf = DecisionTreeClassifier ()
10301017 australian_task = 595 # Australian; crossvalidation
@@ -1040,7 +1027,6 @@ def test_local_run_swapped_parameter_order_model(self):
10401027 self ._test_local_evaluations (run )
10411028
10421029 @pytest .mark .sklearn ()
1043- @pytest .mark .xfail (reason = "failures_issue_1544" )
10441030 @unittest .skipIf (
10451031 Version (sklearn .__version__ ) < Version ("0.20" ),
10461032 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1069,7 +1055,6 @@ def test_local_run_swapped_parameter_order_flow(self):
10691055 self ._test_local_evaluations (run )
10701056
10711057 @pytest .mark .sklearn ()
1072- @pytest .mark .xfail (reason = "failures_issue_1544" )
10731058 @unittest .skipIf (
10741059 Version (sklearn .__version__ ) < Version ("0.20" ),
10751060 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1107,7 +1092,6 @@ def test_online_run_metric_score(self):
11071092 self ._test_local_evaluations (run )
11081093
11091094 @pytest .mark .sklearn ()
1110- @pytest .mark .xfail (reason = "failures_issue_1544" )
11111095 @unittest .skipIf (
11121096 Version (sklearn .__version__ ) < Version ("0.20" ),
11131097 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1173,7 +1157,6 @@ def test_initialize_model_from_run(self):
11731157 Version (sklearn .__version__ ) < Version ("0.20" ),
11741158 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
11751159 )
1176- @pytest .mark .xfail (reason = "failures_issue_1544" )
11771160 def test__run_exists (self ):
11781161 # would be better to not sentinel these clfs,
11791162 # so we do not have to perform the actual runs
@@ -1229,7 +1212,6 @@ def test__run_exists(self):
12291212 assert run_ids , (run_ids , clf )
12301213
12311214 @pytest .mark .sklearn ()
1232- @pytest .mark .xfail (reason = "failures_issue_1544" )
12331215 def test_run_with_illegal_flow_id (self ):
12341216 # check the case where the user adds an illegal flow id to a
12351217 # non-existing flo
@@ -1249,7 +1231,6 @@ def test_run_with_illegal_flow_id(self):
12491231 )
12501232
12511233 @pytest .mark .sklearn ()
1252- @pytest .mark .xfail (reason = "failures_issue_1544" )
12531234 def test_run_with_illegal_flow_id_after_load (self ):
12541235 # Same as `test_run_with_illegal_flow_id`, but test this error is also
12551236 # caught if the run is stored to and loaded from disk first.
@@ -1281,7 +1262,6 @@ def test_run_with_illegal_flow_id_after_load(self):
12811262 TestBase .logger .info (f"collected from test_run_functions: { loaded_run .run_id } " )
12821263
12831264 @pytest .mark .sklearn ()
1284- @pytest .mark .xfail (reason = "failures_issue_1544" )
12851265 def test_run_with_illegal_flow_id_1 (self ):
12861266 # Check the case where the user adds an illegal flow id to an existing
12871267 # flow. Comes to a different value error than the previous test
@@ -1307,7 +1287,6 @@ def test_run_with_illegal_flow_id_1(self):
13071287 )
13081288
13091289 @pytest .mark .sklearn ()
1310- @pytest .mark .xfail (reason = "failures_issue_1544" )
13111290 def test_run_with_illegal_flow_id_1_after_load (self ):
13121291 # Same as `test_run_with_illegal_flow_id_1`, but test this error is
13131292 # also caught if the run is stored to and loaded from disk first.
@@ -1346,7 +1325,6 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13461325 )
13471326
13481327 @pytest .mark .sklearn ()
1349- @pytest .mark .xfail (reason = "failures_issue_1544" )
13501328 @unittest .skipIf (
13511329 Version (sklearn .__version__ ) < Version ("0.20" ),
13521330 reason = "OneHotEncoder cannot handle mixed type DataFrame as input" ,
@@ -1574,7 +1552,6 @@ def test_get_runs_list_by_tag(self):
15741552 assert len (runs ) >= 1
15751553
15761554 @pytest .mark .sklearn ()
1577- @pytest .mark .xfail (reason = "failures_issue_1544" )
15781555 @unittest .skipIf (
15791556 Version (sklearn .__version__ ) < Version ("0.20" ),
15801557 reason = "columntransformer introduction in 0.20.0" ,
@@ -1611,7 +1588,6 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
16111588 assert len (row ) == 12
16121589
16131590 @pytest .mark .sklearn ()
1614- @pytest .mark .xfail (reason = "failures_issue_1544" )
16151591 @unittest .skipIf (
16161592 Version (sklearn .__version__ ) < Version ("0.20" ),
16171593 reason = "columntransformer introduction in 0.20.0" ,
@@ -1664,7 +1640,6 @@ def test_get_uncached_run(self):
16641640 openml .runs .functions ._get_cached_run (10 )
16651641
16661642 @pytest .mark .sklearn ()
1667- @pytest .mark .xfail (reason = "failures_issue_1544" )
16681643 def test_run_flow_on_task_downloaded_flow (self ):
16691644 model = sklearn .ensemble .RandomForestClassifier (n_estimators = 33 )
16701645 flow = self .extension .model_to_flow (model )
@@ -1765,7 +1740,6 @@ def test_format_prediction_task_regression(self):
17651740 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
17661741 )
17671742 @pytest .mark .sklearn ()
1768- @pytest .mark .xfail (reason = "failures_issue_1544" )
17691743 def test_delete_run (self ):
17701744 rs = np .random .randint (1 , 2 ** 31 - 1 )
17711745 clf = sklearn .pipeline .Pipeline (
@@ -1861,7 +1835,6 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18611835
18621836
18631837@pytest .mark .sklearn ()
1864- @pytest .mark .xfail (reason = "failures_issue_1544" )
18651838@unittest .skipIf (
18661839 Version (sklearn .__version__ ) < Version ("0.21" ),
18671840 reason = "couldn't perform local tests successfully w/o bloating RAM" ,
@@ -1957,7 +1930,6 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19571930 (- 1 , "threading" , 10 ), # the threading backend does preserve mocks even with parallelizing
19581931 ]
19591932)
1960- @pytest .mark .xfail (reason = "failures_issue_1544" )
19611933def test_joblib_backends (parallel_mock , n_jobs , backend , call_count ):
19621934 """Tests evaluation of a run using various joblib backends and n_jobs."""
19631935 if backend is None :
0 commit comments