Skip to content

Enhance ROC Curve Display Tests for Improved Clarity and Maintainability #31264

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 6 commits into from

Conversation

NEREUScode
Copy link
Contributor

PR Description:

Summary of Changes:

This PR refactors the data_binary fixture in the test_roc_curve_display.py file. The previous fixture filtered a multiclass dataset (Iris) to create a binary classification task. However, this approach resulted in AUC values consistently reaching 1.0, which does not reflect real-world challenges.

The new fixture utilizes make_classification from sklearn.datasets to generate a synthetic binary classification dataset with the following characteristics:

  • 200 samples and 20 features.
  • 5 informative features and 2 redundant features.
  • 10% label noise (flip_y=0.1) to simulate real-world imperfections in the data.
  • Class separation (class_sep=0.8) set to avoid perfect separation.

These changes provide a more complex and representative dataset for testing the roc_curve_display function and other related metrics, thereby improving the robustness of tests.

Reference Issues/PRs:


For Reviewers:

  • This change ensures that the dataset used for testing is more reflective of real-world data, particularly in classification tasks that may involve noise and less clear separation between classes.

Replaced the `data_binary` fixture that filtered classes from a multiclass dataset 
with a new fixture generating a synthetic binary classification dataset using 
`make_classification`. This ensures consistent data characteristics, introduces 
label noise, and better simulates real-world classification challenges.
Copy link

github-actions bot commented Apr 28, 2025

✔️ Linting Passed

All linting checks passed. Your pull request is in excellent shape! ☀️

Generated for commit: 57bc822. Link to the linter CI: here

@NEREUScode
Copy link
Contributor Author

@lucyleeow I removed data() as suggested, but I'm still unsure why the Linux check is failing

@@ -1,11 +1,11 @@
import numpy as np
import pytest
from numpy.testing import assert_allclose
from scipy.integrate import trapezoid
from scipy.integrate import trapz as trapezoid
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why this change?
The CI failure seems to be due to this

E ImportError: cannot import name 'trapz' from 'scipy.integrate' (/usr/share/miniconda/envs/testvenv/lib/python3.13/site-packages/scipy/integrate/init.py)

you can see the test failure details by clicking through 'details' eg. https://dev.azure.com/scikit-learn/scikit-learn/_build/results?buildId=76020&view=logs&j=dde5042c-7464-5d47-9507-31bdd2ee0a3a&t=4bd2dad8-62b3-5bf9-08a5-a9880c530c94&l=918

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thanks a lot i'll fix it

@NEREUScode NEREUScode closed this Apr 28, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Use more complex data in test_roc_curve_display.py
2 participants