From 70085154538d7741225787acf25d87a4b2a24319 Mon Sep 17 00:00:00 2001 From: Gael Varoquaux Date: Sat, 19 Jul 2014 20:40:19 +0200 Subject: [PATCH] TST: fix tests on numpy 1.9.b2 --- sklearn/metrics/tests/test_ranking.py | 4 ++-- sklearn/metrics/tests/test_score_objects.py | 14 ++++++-------- 2 files changed, 8 insertions(+), 10 deletions(-) diff --git a/sklearn/metrics/tests/test_ranking.py b/sklearn/metrics/tests/test_ranking.py index ccea148ca6fd4..f53bbc53dc550 100644 --- a/sklearn/metrics/tests/test_ranking.py +++ b/sklearn/metrics/tests/test_ranking.py @@ -867,14 +867,14 @@ def check_alternative_lrap_implementation(lrap_score, n_classes=5, y_score = y_score.toarray() score_lrap = label_ranking_average_precision_score(y_true, y_score) score_my_lrap = _my_lrap(y_true, y_score) - assert_equal(score_lrap, score_my_lrap) + assert_almost_equal(score_lrap, score_my_lrap) # Uniform score random_state = check_random_state(random_state) y_score = random_state.uniform(size=(n_samples, n_classes)) score_lrap = label_ranking_average_precision_score(y_true, y_score) score_my_lrap = _my_lrap(y_true, y_score) - assert_equal(score_lrap, score_my_lrap) + assert_almost_equal(score_lrap, score_my_lrap) def test_label_ranking_avp(): diff --git a/sklearn/metrics/tests/test_score_objects.py b/sklearn/metrics/tests/test_score_objects.py index ea1e05bb6334d..0845b5e5a5709 100644 --- a/sklearn/metrics/tests/test_score_objects.py +++ b/sklearn/metrics/tests/test_score_objects.py @@ -7,7 +7,6 @@ from sklearn.utils.testing import assert_raises_regexp from sklearn.utils.testing import assert_true from sklearn.utils.testing import ignore_warnings -from sklearn.utils.testing import assert_equal from sklearn.utils.testing import assert_not_equal from sklearn.metrics import (f1_score, r2_score, roc_auc_score, fbeta_score, @@ -270,13 +269,12 @@ def test_scorer_sample_weight(): ignored = scorer(estimator[name], X_test[10:], y_test[10:]) unweighted = scorer(estimator[name], X_test, y_test) assert_not_equal(weighted, unweighted, - "scorer {0} behaves identically when called with " - "sample weights: {1} vs {2}".format(name, - weighted, - unweighted)) - assert_equal(weighted, ignored, - "scorer {0} behaves differently when ignoring " - "samples and setting sample_weight to 0: " + msg="scorer {0} behaves identically when " + "called with sample weights: {1} vs " + "{2}".format(name, weighted, unweighted)) + assert_almost_equal(weighted, ignored, + err_msg="scorer {0} behaves differently when " + "ignoring samples and setting sample_weight to 0: " "{1} vs {2}".format(name, weighted, ignored)) except TypeError as e: