Skip to content

GitLab

  • Projects
  • Groups
  • Snippets
  • Help
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in / Register
D dollvenue
  • Project overview
    • Project overview
    • Details
    • Activity
  • Issues 8
    • Issues 8
    • List
    • Boards
    • Labels
    • Service Desk
    • Milestones
  • Merge requests 0
    • Merge requests 0
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Operations
    • Operations
    • Incidents
    • Environments
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Container Registry
  • Analytics
    • Analytics
    • Value Stream
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Members
    • Members
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
Collapse sidebar
  • Aretha Prater
  • dollvenue
  • Issues
  • #5

Closed
Open
Created Feb 10, 2025 by Aretha Prater@arethaprater22Maintainer

Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy


Machine-learning models can fail when they attempt to make predictions for people who were underrepresented in the datasets they were trained on.

For instance, a design that anticipates the best treatment choice for somebody with a persistent disease may be trained utilizing a dataset that contains mainly male clients. That design may make inaccurate predictions for female patients when released in a healthcare facility.

To improve outcomes, engineers can try balancing the training dataset by getting rid of information points till all subgroups are represented similarly. While dataset balancing is promising, it typically requires getting rid of big quantity of data, harming the model's overall efficiency.

MIT scientists developed a new strategy that identifies and eliminates specific points in a training dataset that contribute most to a design's failures on minority subgroups. By eliminating far fewer datapoints than other methods, this technique maintains the total precision of the design while improving its efficiency concerning underrepresented groups.

In addition, the method can identify surprise sources of predisposition in a training dataset that does not have labels. Unlabeled information are even more common than labeled information for numerous applications.

This technique might likewise be combined with other techniques to improve the fairness of machine-learning designs released in high-stakes situations. For instance, it may one day help make sure underrepresented clients aren't misdiagnosed due to a prejudiced AI model.

"Many other algorithms that attempt to address this concern assume each datapoint matters as much as every other datapoint. In this paper, we are revealing that presumption is not true. There specify points in our dataset that are contributing to this predisposition, and we can find those data points, eliminate them, and get much better efficiency," says Kimia Hamidieh, an electrical engineering and computer system science (EECS) graduate trainee at MIT and co-lead author of a paper on this method.

She wrote the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev; Andrew Ilyas MEng '18, PhD '23, a Stein Fellow at Stanford University; and senior authors Marzyeh Ghassemi, an associate teacher in EECS and a member of the Institute of Medical Engineering Sciences and the Laboratory for Details and Decision Systems, and Aleksander Madry, the Cadence Design Systems Professor at MIT. The research study will exist at the Conference on Neural Details Processing Systems.

Removing bad examples

Often, machine-learning designs are trained using huge datasets gathered from many sources across the internet. These datasets are far too big to be carefully curated by hand, so they may contain bad examples that harm model performance.

Scientists also know that some data points impact a design's efficiency on certain downstream jobs more than others.

The MIT researchers integrated these 2 ideas into a technique that determines and gets rid of these troublesome datapoints. They look for to fix a problem understood as worst-group error, which occurs when a model underperforms on minority subgroups in a training dataset.

The researchers' brand-new method is driven by prior work in which they introduced a technique, called TRAK, that recognizes the most essential training examples for a specific design output.

For this brand-new method, they take inaccurate predictions the model made about minority subgroups and utilize TRAK to identify which training examples contributed the most to that incorrect prediction.

"By aggregating this details across bad test predictions in the proper way, we have the ability to find the specific parts of the training that are driving worst-group precision down in general," Ilyas explains.

Then they remove those specific samples and retrain the model on the .

Since having more information generally yields much better general efficiency, removing just the samples that drive worst-group failures maintains the model's overall precision while increasing its performance on minority subgroups.

A more available approach

Across 3 machine-learning datasets, their technique surpassed multiple methods. In one circumstances, fakenews.win it enhanced worst-group precision while removing about 20,000 fewer training samples than a standard data balancing approach. Their method also attained higher accuracy than techniques that require making changes to the inner functions of a model.

Because the MIT technique includes changing a dataset rather, it would be simpler for a practitioner to use and can be applied to lots of kinds of designs.

It can likewise be made use of when predisposition is unidentified because subgroups in a training dataset are not labeled. By recognizing datapoints that contribute most to a feature the model is discovering, they can understand the variables it is utilizing to make a prediction.

"This is a tool anyone can use when they are training a machine-learning design. They can look at those datapoints and see whether they are aligned with the capability they are trying to teach the design," says Hamidieh.

Using the strategy to spot unidentified subgroup bias would need intuition about which groups to try to find, so the researchers hope to validate it and explore it more completely through future human research studies.

They likewise wish to enhance the performance and dependability of their method and ensure the technique is available and user friendly for professionals who could someday release it in real-world environments.

"When you have tools that let you critically look at the data and figure out which datapoints are going to result in bias or other unfavorable behavior, it provides you an initial step towards building designs that are going to be more fair and more trustworthy," Ilyas states.

This work is moneyed, in part, by the National Science Foundation and the U.S. Defense Advanced Research Projects Agency.

Assignee
Assign to
None
Milestone
None
Assign milestone
Time tracking