Skip to content

Commit 6043686

Browse files
authored
[MNT] Update xfail for new test server state (#1585)
#### Metadata * Reference Issue: #1544 * New Tests Added: No * Documentation Updated: No #### Details I investigated the failures and the root cause was incorrect test server state. This still remains an issue for one test, but I can look into that later (after I return from my vacation).
1 parent 7975eb5 commit 6043686

4 files changed

Lines changed: 2 additions & 37 deletions

File tree

tests/test_datasets/test_dataset.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -294,6 +294,7 @@ def test_tagging():
294294
datasets = openml.datasets.list_datasets(tag=tag)
295295
assert datasets.empty
296296

297+
@pytest.mark.xfail(reason="failures_issue_1544")
297298
def test_get_feature_with_ontology_data_id_11():
298299
# test on car dataset, which has built-in ontology references
299300
dataset = openml.datasets.get_dataset(11)
@@ -470,4 +471,4 @@ def test__check_qualities():
470471

471472
qualities = [{"oml:name": "a", "oml:value": None}]
472473
qualities = openml.datasets.dataset._check_qualities(qualities)
473-
assert qualities["a"] != qualities["a"]
474+
assert qualities["a"] != qualities["a"]

tests/test_runs/test_run.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -118,7 +118,6 @@ def _check_array(array, type_):
118118
assert run_prime_trace_content is None
119119

120120
@pytest.mark.sklearn()
121-
@pytest.mark.xfail(reason="failures_issue_1544")
122121
def test_to_from_filesystem_vanilla(self):
123122
model = Pipeline(
124123
[
@@ -154,7 +153,6 @@ def test_to_from_filesystem_vanilla(self):
154153

155154
@pytest.mark.sklearn()
156155
@pytest.mark.flaky()
157-
@pytest.mark.xfail(reason="failures_issue_1544")
158156
def test_to_from_filesystem_search(self):
159157
model = Pipeline(
160158
[
@@ -189,7 +187,6 @@ def test_to_from_filesystem_search(self):
189187
)
190188

191189
@pytest.mark.sklearn()
192-
@pytest.mark.xfail(reason="failures_issue_1544")
193190
def test_to_from_filesystem_no_model(self):
194191
model = Pipeline(
195192
[("imputer", SimpleImputer(strategy="mean")), ("classifier", DummyClassifier())],
@@ -295,7 +292,6 @@ def assert_run_prediction_data(task, run, model):
295292
assert_method(y_test, saved_y_test)
296293

297294
@pytest.mark.sklearn()
298-
@pytest.mark.xfail(reason="failures_issue_1544")
299295
def test_publish_with_local_loaded_flow(self):
300296
"""
301297
Publish a run tied to a local flow after it has first been saved to
@@ -339,7 +335,6 @@ def test_publish_with_local_loaded_flow(self):
339335
openml.runs.get_run(loaded_run.run_id)
340336

341337
@pytest.mark.sklearn()
342-
@pytest.mark.xfail(reason="failures_issue_1544")
343338
def test_offline_and_online_run_identical(self):
344339
extension = SklearnExtension()
345340

tests/test_runs/test_run_functions.py

Lines changed: 0 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -398,7 +398,6 @@ def _check_sample_evaluations(
398398
assert evaluation < max_time_allowed
399399

400400
@pytest.mark.sklearn()
401-
@pytest.mark.xfail(reason="failures_issue_1544")
402401
def test_run_regression_on_classif_task(self):
403402
task_id = 259 # collins; crossvalidation; has numeric targets
404403

@@ -415,7 +414,6 @@ def test_run_regression_on_classif_task(self):
415414
)
416415

417416
@pytest.mark.sklearn()
418-
@pytest.mark.xfail(reason="failures_issue_1544")
419417
def test_check_erronous_sklearn_flow_fails(self):
420418
task_id = 115 # diabetes; crossvalidation
421419
task = openml.tasks.get_task(task_id)
@@ -628,7 +626,6 @@ def _run_and_upload_regression(
628626
)
629627

630628
@pytest.mark.sklearn()
631-
@pytest.mark.xfail(reason="failures_issue_1544")
632629
def test_run_and_upload_logistic_regression(self):
633630
lr = LogisticRegression(solver="lbfgs", max_iter=1000)
634631
task_id = self.TEST_SERVER_TASK_SIMPLE["task_id"]
@@ -637,7 +634,6 @@ def test_run_and_upload_logistic_regression(self):
637634
self._run_and_upload_classification(lr, task_id, n_missing_vals, n_test_obs, "62501")
638635

639636
@pytest.mark.sklearn()
640-
@pytest.mark.xfail(reason="failures_issue_1544")
641637
def test_run_and_upload_linear_regression(self):
642638
lr = LinearRegression()
643639
task_id = self.TEST_SERVER_TASK_REGRESSION["task_id"]
@@ -668,7 +664,6 @@ def test_run_and_upload_linear_regression(self):
668664
self._run_and_upload_regression(lr, task_id, n_missing_vals, n_test_obs, "62501")
669665

670666
@pytest.mark.sklearn()
671-
@pytest.mark.xfail(reason="failures_issue_1544")
672667
def test_run_and_upload_pipeline_dummy_pipeline(self):
673668
pipeline1 = Pipeline(
674669
steps=[
@@ -682,7 +677,6 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
682677
self._run_and_upload_classification(pipeline1, task_id, n_missing_vals, n_test_obs, "62501")
683678

684679
@pytest.mark.sklearn()
685-
@pytest.mark.xfail(reason="failures_issue_1544")
686680
@unittest.skipIf(
687681
Version(sklearn.__version__) < Version("0.20"),
688682
reason="columntransformer introduction in 0.20.0",
@@ -799,7 +793,6 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
799793
assert call_count == 3
800794

801795
@pytest.mark.sklearn()
802-
@pytest.mark.xfail(reason="failures_issue_1544")
803796
def test_run_and_upload_gridsearch(self):
804797
estimator_name = (
805798
"base_estimator" if Version(sklearn.__version__) < Version("1.4") else "estimator"
@@ -822,7 +815,6 @@ def test_run_and_upload_gridsearch(self):
822815
assert len(run.trace.trace_iterations) == 9
823816

824817
@pytest.mark.sklearn()
825-
@pytest.mark.xfail(reason="failures_issue_1544")
826818
def test_run_and_upload_randomsearch(self):
827819
randomsearch = RandomizedSearchCV(
828820
RandomForestClassifier(n_estimators=5),
@@ -855,7 +847,6 @@ def test_run_and_upload_randomsearch(self):
855847
assert len(trace.trace_iterations) == 5
856848

857849
@pytest.mark.sklearn()
858-
@pytest.mark.xfail(reason="failures_issue_1544")
859850
def test_run_and_upload_maskedarrays(self):
860851
# This testcase is important for 2 reasons:
861852
# 1) it verifies the correct handling of masked arrays (not all
@@ -883,7 +874,6 @@ def test_run_and_upload_maskedarrays(self):
883874
##########################################################################
884875

885876
@pytest.mark.sklearn()
886-
@pytest.mark.xfail(reason="failures_issue_1544")
887877
def test_learning_curve_task_1(self):
888878
task_id = 801 # diabates dataset
889879
num_test_instances = 6144 # for learning curve
@@ -908,7 +898,6 @@ def test_learning_curve_task_1(self):
908898
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
909899

910900
@pytest.mark.sklearn()
911-
@pytest.mark.xfail(reason="failures_issue_1544")
912901
def test_learning_curve_task_2(self):
913902
task_id = 801 # diabates dataset
914903
num_test_instances = 6144 # for learning curve
@@ -945,7 +934,6 @@ def test_learning_curve_task_2(self):
945934
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
946935

947936
@pytest.mark.sklearn()
948-
@pytest.mark.xfail(reason="failures_issue_1544")
949937
@unittest.skipIf(
950938
Version(sklearn.__version__) < Version("0.21"),
951939
reason="Pipelines don't support indexing (used for the assert check)",
@@ -1024,7 +1012,6 @@ def _test_local_evaluations(self, run):
10241012
assert alt_scores[idx] <= 1
10251013

10261014
@pytest.mark.sklearn()
1027-
@pytest.mark.xfail(reason="failures_issue_1544")
10281015
def test_local_run_swapped_parameter_order_model(self):
10291016
clf = DecisionTreeClassifier()
10301017
australian_task = 595 # Australian; crossvalidation
@@ -1040,7 +1027,6 @@ def test_local_run_swapped_parameter_order_model(self):
10401027
self._test_local_evaluations(run)
10411028

10421029
@pytest.mark.sklearn()
1043-
@pytest.mark.xfail(reason="failures_issue_1544")
10441030
@unittest.skipIf(
10451031
Version(sklearn.__version__) < Version("0.20"),
10461032
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1069,7 +1055,6 @@ def test_local_run_swapped_parameter_order_flow(self):
10691055
self._test_local_evaluations(run)
10701056

10711057
@pytest.mark.sklearn()
1072-
@pytest.mark.xfail(reason="failures_issue_1544")
10731058
@unittest.skipIf(
10741059
Version(sklearn.__version__) < Version("0.20"),
10751060
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1107,7 +1092,6 @@ def test_online_run_metric_score(self):
11071092
self._test_local_evaluations(run)
11081093

11091094
@pytest.mark.sklearn()
1110-
@pytest.mark.xfail(reason="failures_issue_1544")
11111095
@unittest.skipIf(
11121096
Version(sklearn.__version__) < Version("0.20"),
11131097
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1173,7 +1157,6 @@ def test_initialize_model_from_run(self):
11731157
Version(sklearn.__version__) < Version("0.20"),
11741158
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
11751159
)
1176-
@pytest.mark.xfail(reason="failures_issue_1544")
11771160
def test__run_exists(self):
11781161
# would be better to not sentinel these clfs,
11791162
# so we do not have to perform the actual runs
@@ -1229,7 +1212,6 @@ def test__run_exists(self):
12291212
assert run_ids, (run_ids, clf)
12301213

12311214
@pytest.mark.sklearn()
1232-
@pytest.mark.xfail(reason="failures_issue_1544")
12331215
def test_run_with_illegal_flow_id(self):
12341216
# check the case where the user adds an illegal flow id to a
12351217
# non-existing flo
@@ -1249,7 +1231,6 @@ def test_run_with_illegal_flow_id(self):
12491231
)
12501232

12511233
@pytest.mark.sklearn()
1252-
@pytest.mark.xfail(reason="failures_issue_1544")
12531234
def test_run_with_illegal_flow_id_after_load(self):
12541235
# Same as `test_run_with_illegal_flow_id`, but test this error is also
12551236
# caught if the run is stored to and loaded from disk first.
@@ -1281,7 +1262,6 @@ def test_run_with_illegal_flow_id_after_load(self):
12811262
TestBase.logger.info(f"collected from test_run_functions: {loaded_run.run_id}")
12821263

12831264
@pytest.mark.sklearn()
1284-
@pytest.mark.xfail(reason="failures_issue_1544")
12851265
def test_run_with_illegal_flow_id_1(self):
12861266
# Check the case where the user adds an illegal flow id to an existing
12871267
# flow. Comes to a different value error than the previous test
@@ -1307,7 +1287,6 @@ def test_run_with_illegal_flow_id_1(self):
13071287
)
13081288

13091289
@pytest.mark.sklearn()
1310-
@pytest.mark.xfail(reason="failures_issue_1544")
13111290
def test_run_with_illegal_flow_id_1_after_load(self):
13121291
# Same as `test_run_with_illegal_flow_id_1`, but test this error is
13131292
# also caught if the run is stored to and loaded from disk first.
@@ -1346,7 +1325,6 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13461325
)
13471326

13481327
@pytest.mark.sklearn()
1349-
@pytest.mark.xfail(reason="failures_issue_1544")
13501328
@unittest.skipIf(
13511329
Version(sklearn.__version__) < Version("0.20"),
13521330
reason="OneHotEncoder cannot handle mixed type DataFrame as input",
@@ -1574,7 +1552,6 @@ def test_get_runs_list_by_tag(self):
15741552
assert len(runs) >= 1
15751553

15761554
@pytest.mark.sklearn()
1577-
@pytest.mark.xfail(reason="failures_issue_1544")
15781555
@unittest.skipIf(
15791556
Version(sklearn.__version__) < Version("0.20"),
15801557
reason="columntransformer introduction in 0.20.0",
@@ -1611,7 +1588,6 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
16111588
assert len(row) == 12
16121589

16131590
@pytest.mark.sklearn()
1614-
@pytest.mark.xfail(reason="failures_issue_1544")
16151591
@unittest.skipIf(
16161592
Version(sklearn.__version__) < Version("0.20"),
16171593
reason="columntransformer introduction in 0.20.0",
@@ -1664,7 +1640,6 @@ def test_get_uncached_run(self):
16641640
openml.runs.functions._get_cached_run(10)
16651641

16661642
@pytest.mark.sklearn()
1667-
@pytest.mark.xfail(reason="failures_issue_1544")
16681643
def test_run_flow_on_task_downloaded_flow(self):
16691644
model = sklearn.ensemble.RandomForestClassifier(n_estimators=33)
16701645
flow = self.extension.model_to_flow(model)
@@ -1765,7 +1740,6 @@ def test_format_prediction_task_regression(self):
17651740
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
17661741
)
17671742
@pytest.mark.sklearn()
1768-
@pytest.mark.xfail(reason="failures_issue_1544")
17691743
def test_delete_run(self):
17701744
rs = np.random.randint(1, 2**31 - 1)
17711745
clf = sklearn.pipeline.Pipeline(
@@ -1861,7 +1835,6 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18611835

18621836

18631837
@pytest.mark.sklearn()
1864-
@pytest.mark.xfail(reason="failures_issue_1544")
18651838
@unittest.skipIf(
18661839
Version(sklearn.__version__) < Version("0.21"),
18671840
reason="couldn't perform local tests successfully w/o bloating RAM",
@@ -1957,7 +1930,6 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19571930
(-1, "threading", 10), # the threading backend does preserve mocks even with parallelizing
19581931
]
19591932
)
1960-
@pytest.mark.xfail(reason="failures_issue_1544")
19611933
def test_joblib_backends(parallel_mock, n_jobs, backend, call_count):
19621934
"""Tests evaluation of a run using various joblib backends and n_jobs."""
19631935
if backend is None:

tests/test_setups/test_setup_functions.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,6 @@ def _existing_setup_exists(self, classif):
8282
assert setup_id == run.setup_id
8383

8484
@pytest.mark.sklearn()
85-
@pytest.mark.xfail(reason="failures_issue_1544")
8685
def test_existing_setup_exists_1(self):
8786
def side_effect(self):
8887
self.var_smoothing = 1e-9
@@ -98,13 +97,11 @@ def side_effect(self):
9897
self._existing_setup_exists(nb)
9998

10099
@pytest.mark.sklearn()
101-
@pytest.mark.xfail(reason="failures_issue_1544")
102100
def test_exisiting_setup_exists_2(self):
103101
# Check a flow with one hyperparameter
104102
self._existing_setup_exists(sklearn.naive_bayes.GaussianNB())
105103

106104
@pytest.mark.sklearn()
107-
@pytest.mark.xfail(reason="failures_issue_1544")
108105
def test_existing_setup_exists_3(self):
109106
# Check a flow with many hyperparameters
110107
self._existing_setup_exists(

0 commit comments

Comments
 (0)