Skip to content

Commit e148979

Browse files
committed
Revert "[MNT] Update xfail for new test server state (openml#1585)"
This reverts commit 6043686.
1 parent f9f77ba commit e148979

4 files changed

Lines changed: 37 additions & 3 deletions

File tree

tests/test_datasets/test_dataset.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -294,7 +294,6 @@ def test_tagging():
294294
datasets = openml.datasets.list_datasets(tag=tag)
295295
assert datasets.empty
296296

297-
@pytest.mark.xfail(reason="failures_issue_1544")
298297
def test_get_feature_with_ontology_data_id_11():
299298
# test on car dataset, which has built-in ontology references
300299
dataset = openml.datasets.get_dataset(11)
@@ -471,4 +470,4 @@ def test__check_qualities():
471470

472471
qualities = [{"oml:name": "a", "oml:value": None}]
473472
qualities = openml.datasets.dataset._check_qualities(qualities)
474-
assert qualities["a"] != qualities["a"]
473+
assert qualities["a"] != qualities["a"]

tests/test_runs/test_run.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -118,6 +118,7 @@ def _check_array(array, type_):
118118
assert run_prime_trace_content is None
119119

120120
@pytest.mark.sklearn()
121+
@pytest.mark.xfail(reason="failures_issue_1544")
121122
def test_to_from_filesystem_vanilla(self):
122123
model = Pipeline(
123124
[
@@ -153,6 +154,7 @@ def test_to_from_filesystem_vanilla(self):
153154

154155
@pytest.mark.sklearn()
155156
@pytest.mark.flaky()
157+
@pytest.mark.xfail(reason="failures_issue_1544")
156158
def test_to_from_filesystem_search(self):
157159
model = Pipeline(
158160
[
@@ -187,6 +189,7 @@ def test_to_from_filesystem_search(self):
187189
)
188190

189191
@pytest.mark.sklearn()
192+
@pytest.mark.xfail(reason="failures_issue_1544")
190193
def test_to_from_filesystem_no_model(self):
191194
model = Pipeline(
192195
[("imputer", SimpleImputer(strategy="mean")), ("classifier", DummyClassifier())],
@@ -292,6 +295,7 @@ def assert_run_prediction_data(task, run, model):
292295
assert_method(y_test, saved_y_test)
293296

294297
@pytest.mark.sklearn()
298+
@pytest.mark.xfail(reason="failures_issue_1544")
295299
def test_publish_with_local_loaded_flow(self):
296300
"""
297301
Publish a run tied to a local flow after it has first been saved to
@@ -335,6 +339,7 @@ def test_publish_with_local_loaded_flow(self):
335339
openml.runs.get_run(loaded_run.run_id)
336340

337341
@pytest.mark.sklearn()
342+
@pytest.mark.xfail(reason="failures_issue_1544")
338343
def test_offline_and_online_run_identical(self):
339344
extension = SklearnExtension()
340345

tests/test_runs/test_run_functions.py

Lines changed: 28 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -398,6 +398,7 @@ def _check_sample_evaluations(
398398
assert evaluation < max_time_allowed
399399

400400
@pytest.mark.sklearn()
401+
@pytest.mark.xfail(reason="failures_issue_1544")
401402
def test_run_regression_on_classif_task(self):
402403
task_id = 259 # collins; crossvalidation; has numeric targets
403404

@@ -414,6 +415,7 @@ def test_run_regression_on_classif_task(self):
414415
)
415416

416417
@pytest.mark.sklearn()
418+
@pytest.mark.xfail(reason="failures_issue_1544")
417419
def test_check_erronous_sklearn_flow_fails(self):
418420
task_id = 115 # diabetes; crossvalidation
419421
task = openml.tasks.get_task(task_id)
@@ -627,6 +629,7 @@ def _run_and_upload_regression(
627629

628630
@pytest.mark.skip(reason="failures_issue_1544")
629631
@pytest.mark.sklearn()
632+
@pytest.mark.xfail(reason="failures_issue_1544")
630633
def test_run_and_upload_logistic_regression(self):
631634
lr = LogisticRegression(solver="lbfgs", max_iter=1000)
632635
task_id = self.TEST_SERVER_TASK_SIMPLE["task_id"]
@@ -636,6 +639,7 @@ def test_run_and_upload_logistic_regression(self):
636639

637640
@pytest.mark.skip(reason="failures_issue_1544")
638641
@pytest.mark.sklearn()
642+
@pytest.mark.xfail(reason="failures_issue_1544")
639643
def test_run_and_upload_linear_regression(self):
640644
lr = LinearRegression()
641645
task_id = self.TEST_SERVER_TASK_REGRESSION["task_id"]
@@ -667,6 +671,7 @@ def test_run_and_upload_linear_regression(self):
667671

668672
@pytest.mark.skip(reason="failures_issue_1544")
669673
@pytest.mark.sklearn()
674+
@pytest.mark.xfail(reason="failures_issue_1544")
670675
def test_run_and_upload_pipeline_dummy_pipeline(self):
671676
pipeline1 = Pipeline(
672677
steps=[
@@ -681,6 +686,7 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
681686

682687
@pytest.mark.skip(reason="failures_issue_1544")
683688
@pytest.mark.sklearn()
689+
@pytest.mark.xfail(reason="failures_issue_1544")
684690
@unittest.skipIf(
685691
Version(sklearn.__version__) < Version("0.20"),
686692
reason="columntransformer introduction in 0.20.0",
@@ -799,6 +805,7 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
799805

800806
@pytest.mark.skip(reason="failures_issue_1544")
801807
@pytest.mark.sklearn()
808+
@pytest.mark.xfail(reason="failures_issue_1544")
802809
def test_run_and_upload_gridsearch(self):
803810
estimator_name = (
804811
"base_estimator" if Version(sklearn.__version__) < Version("1.4") else "estimator"
@@ -821,7 +828,7 @@ def test_run_and_upload_gridsearch(self):
821828
assert len(run.trace.trace_iterations) == 9
822829

823830
@pytest.mark.sklearn()
824-
@pytest.mark.skip(reason="failures_issue_1544")
831+
@pytest.mark.xfail(reason="failures_issue_1544")
825832
def test_run_and_upload_randomsearch(self):
826833
randomsearch = RandomizedSearchCV(
827834
RandomForestClassifier(n_estimators=5),
@@ -855,6 +862,7 @@ def test_run_and_upload_randomsearch(self):
855862

856863
@pytest.mark.skip(reason="failures_issue_1544")
857864
@pytest.mark.sklearn()
865+
@pytest.mark.xfail(reason="failures_issue_1544")
858866
def test_run_and_upload_maskedarrays(self):
859867
# This testcase is important for 2 reasons:
860868
# 1) it verifies the correct handling of masked arrays (not all
@@ -882,6 +890,7 @@ def test_run_and_upload_maskedarrays(self):
882890
##########################################################################
883891

884892
@pytest.mark.sklearn()
893+
@pytest.mark.xfail(reason="failures_issue_1544")
885894
def test_learning_curve_task_1(self):
886895
task_id = 801 # diabates dataset
887896
num_test_instances = 6144 # for learning curve
@@ -906,6 +915,7 @@ def test_learning_curve_task_1(self):
906915
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
907916

908917
@pytest.mark.sklearn()
918+
@pytest.mark.xfail(reason="failures_issue_1544")
909919
def test_learning_curve_task_2(self):
910920
task_id = 801 # diabates dataset
911921
num_test_instances = 6144 # for learning curve
@@ -942,6 +952,7 @@ def test_learning_curve_task_2(self):
942952
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
943953

944954
@pytest.mark.sklearn()
955+
@pytest.mark.xfail(reason="failures_issue_1544")
945956
@unittest.skipIf(
946957
Version(sklearn.__version__) < Version("0.21"),
947958
reason="Pipelines don't support indexing (used for the assert check)",
@@ -1020,6 +1031,7 @@ def _test_local_evaluations(self, run):
10201031
assert alt_scores[idx] <= 1
10211032

10221033
@pytest.mark.sklearn()
1034+
@pytest.mark.xfail(reason="failures_issue_1544")
10231035
def test_local_run_swapped_parameter_order_model(self):
10241036
clf = DecisionTreeClassifier()
10251037
australian_task = 595 # Australian; crossvalidation
@@ -1035,6 +1047,7 @@ def test_local_run_swapped_parameter_order_model(self):
10351047
self._test_local_evaluations(run)
10361048

10371049
@pytest.mark.sklearn()
1050+
@pytest.mark.xfail(reason="failures_issue_1544")
10381051
@unittest.skipIf(
10391052
Version(sklearn.__version__) < Version("0.20"),
10401053
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1063,6 +1076,7 @@ def test_local_run_swapped_parameter_order_flow(self):
10631076
self._test_local_evaluations(run)
10641077

10651078
@pytest.mark.sklearn()
1079+
@pytest.mark.xfail(reason="failures_issue_1544")
10661080
@unittest.skipIf(
10671081
Version(sklearn.__version__) < Version("0.20"),
10681082
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1100,6 +1114,7 @@ def test_online_run_metric_score(self):
11001114
self._test_local_evaluations(run)
11011115

11021116
@pytest.mark.sklearn()
1117+
@pytest.mark.xfail(reason="failures_issue_1544")
11031118
@unittest.skipIf(
11041119
Version(sklearn.__version__) < Version("0.20"),
11051120
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1165,6 +1180,7 @@ def test_initialize_model_from_run(self):
11651180
Version(sklearn.__version__) < Version("0.20"),
11661181
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
11671182
)
1183+
@pytest.mark.xfail(reason="failures_issue_1544")
11681184
def test__run_exists(self):
11691185
# would be better to not sentinel these clfs,
11701186
# so we do not have to perform the actual runs
@@ -1220,6 +1236,7 @@ def test__run_exists(self):
12201236
assert run_ids, (run_ids, clf)
12211237

12221238
@pytest.mark.sklearn()
1239+
@pytest.mark.xfail(reason="failures_issue_1544")
12231240
def test_run_with_illegal_flow_id(self):
12241241
# check the case where the user adds an illegal flow id to a
12251242
# non-existing flo
@@ -1239,6 +1256,7 @@ def test_run_with_illegal_flow_id(self):
12391256
)
12401257

12411258
@pytest.mark.sklearn()
1259+
@pytest.mark.xfail(reason="failures_issue_1544")
12421260
def test_run_with_illegal_flow_id_after_load(self):
12431261
# Same as `test_run_with_illegal_flow_id`, but test this error is also
12441262
# caught if the run is stored to and loaded from disk first.
@@ -1270,6 +1288,7 @@ def test_run_with_illegal_flow_id_after_load(self):
12701288
TestBase.logger.info(f"collected from test_run_functions: {loaded_run.run_id}")
12711289

12721290
@pytest.mark.sklearn()
1291+
@pytest.mark.xfail(reason="failures_issue_1544")
12731292
def test_run_with_illegal_flow_id_1(self):
12741293
# Check the case where the user adds an illegal flow id to an existing
12751294
# flow. Comes to a different value error than the previous test
@@ -1295,6 +1314,7 @@ def test_run_with_illegal_flow_id_1(self):
12951314
)
12961315

12971316
@pytest.mark.sklearn()
1317+
@pytest.mark.xfail(reason="failures_issue_1544")
12981318
def test_run_with_illegal_flow_id_1_after_load(self):
12991319
# Same as `test_run_with_illegal_flow_id_1`, but test this error is
13001320
# also caught if the run is stored to and loaded from disk first.
@@ -1333,6 +1353,7 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13331353
)
13341354

13351355
@pytest.mark.sklearn()
1356+
@pytest.mark.xfail(reason="failures_issue_1544")
13361357
@unittest.skipIf(
13371358
Version(sklearn.__version__) < Version("0.20"),
13381359
reason="OneHotEncoder cannot handle mixed type DataFrame as input",
@@ -1560,6 +1581,7 @@ def test_get_runs_list_by_tag(self):
15601581
assert len(runs) >= 1
15611582

15621583
@pytest.mark.sklearn()
1584+
@pytest.mark.xfail(reason="failures_issue_1544")
15631585
@unittest.skipIf(
15641586
Version(sklearn.__version__) < Version("0.20"),
15651587
reason="columntransformer introduction in 0.20.0",
@@ -1596,6 +1618,7 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
15961618
assert len(row) == 12
15971619

15981620
@pytest.mark.sklearn()
1621+
@pytest.mark.xfail(reason="failures_issue_1544")
15991622
@unittest.skipIf(
16001623
Version(sklearn.__version__) < Version("0.20"),
16011624
reason="columntransformer introduction in 0.20.0",
@@ -1648,6 +1671,7 @@ def test_get_uncached_run(self):
16481671
openml.runs.functions._get_cached_run(10)
16491672

16501673
@pytest.mark.sklearn()
1674+
@pytest.mark.xfail(reason="failures_issue_1544")
16511675
def test_run_flow_on_task_downloaded_flow(self):
16521676
model = sklearn.ensemble.RandomForestClassifier(n_estimators=33)
16531677
flow = self.extension.model_to_flow(model)
@@ -1748,6 +1772,7 @@ def test_format_prediction_task_regression(self):
17481772
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
17491773
)
17501774
@pytest.mark.sklearn()
1775+
@pytest.mark.xfail(reason="failures_issue_1544")
17511776
def test_delete_run(self):
17521777
rs = np.random.randint(1, 2**31 - 1)
17531778
clf = sklearn.pipeline.Pipeline(
@@ -1843,6 +1868,7 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18431868

18441869

18451870
@pytest.mark.sklearn()
1871+
@pytest.mark.xfail(reason="failures_issue_1544")
18461872
@unittest.skipIf(
18471873
Version(sklearn.__version__) < Version("0.21"),
18481874
reason="couldn't perform local tests successfully w/o bloating RAM",
@@ -1938,6 +1964,7 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19381964
(-1, "threading", 10), # the threading backend does preserve mocks even with parallelizing
19391965
]
19401966
)
1967+
@pytest.mark.xfail(reason="failures_issue_1544")
19411968
def test_joblib_backends(parallel_mock, n_jobs, backend, call_count):
19421969
"""Tests evaluation of a run using various joblib backends and n_jobs."""
19431970
if backend is None:

tests/test_setups/test_setup_functions.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -82,6 +82,7 @@ def _existing_setup_exists(self, classif):
8282
assert setup_id == run.setup_id
8383

8484
@pytest.mark.sklearn()
85+
@pytest.mark.xfail(reason="failures_issue_1544")
8586
def test_existing_setup_exists_1(self):
8687
def side_effect(self):
8788
self.var_smoothing = 1e-9
@@ -97,11 +98,13 @@ def side_effect(self):
9798
self._existing_setup_exists(nb)
9899

99100
@pytest.mark.sklearn()
101+
@pytest.mark.xfail(reason="failures_issue_1544")
100102
def test_exisiting_setup_exists_2(self):
101103
# Check a flow with one hyperparameter
102104
self._existing_setup_exists(sklearn.naive_bayes.GaussianNB())
103105

104106
@pytest.mark.sklearn()
107+
@pytest.mark.xfail(reason="failures_issue_1544")
105108
def test_existing_setup_exists_3(self):
106109
# Check a flow with many hyperparameters
107110
self._existing_setup_exists(

0 commit comments

Comments
 (0)