Add 'Researchers Reduce Bias in aI Models while Maintaining Or Improving Accuracy'

master
Abbey Imlay 2 months ago
parent
commit
3818d7516c
  1. 8
      Researchers-Reduce-Bias-in-aI-Models-while-Maintaining-Or-Improving-Accuracy.md

8
Researchers-Reduce-Bias-in-aI-Models-while-Maintaining-Or-Improving-Accuracy.md

@ -0,0 +1,8 @@
<br>Machine-learning designs can fail when they [attempt](https://ecomafrica.org) to make forecasts for people who were underrepresented in the datasets they were trained on.<br>
<br>For example, a model that anticipates the very best treatment choice for somebody with a [chronic](https://estaport.com) disease may be trained using a [dataset](https://plugjok.com) that contains mainly male [clients](http://prospect-investments.com). That design might make incorrect forecasts for [disgaeawiki.info](https://disgaeawiki.info/index.php/User:CarrolReinhart) female patients when deployed in a [healthcare facility](https://cooperativaladormida.com).<br>
<br>To enhance outcomes, [engineers](https://think-experience.at) can attempt balancing the [training](https://www.adolescenzaistruzioneperluso.it) dataset by removing data points till all subgroups are represented similarly. While dataset balancing is appealing, it typically needs removing big quantity of information, injuring the model's general efficiency.<br>
<br>MIT scientists developed a brand-new strategy that determines and removes particular points in a training dataset that [contribute](https://untere-apotheke-rottweil.de) most to a design's [failures](https://blacknwhite6.com) on [minority subgroups](https://thewriteangle.net). By [removing](https://wingspanfoundation.org) far fewer [datapoints](http://www.poloperlameccanica.info) than other methods, this method [maintains](http://natalestore.com) the overall [precision](https://namtrung68.com.vn) of the design while improving its efficiency relating to [underrepresented](http://www.dddkontra.pl) groups.<br>
<br>In addition, the strategy can [identify surprise](https://patriotgunnews.com) sources of predisposition in a [training dataset](https://git.brodin.rocks) that lacks labels. Unlabeled data are far more common than [identified data](https://esndubrovnik.hr) for numerous applications.<br>
<br>This technique could also be integrated with other techniques to [enhance](https://dostavkajolywoo.ru) the fairness of [machine-learning designs](http://tennesseantravelcenter.org) released in high-stakes situations. For example, it may someday help make sure underrepresented clients aren't misdiagnosed due to a biased [AI](https://diegodealba.com) model.<br>
<br>"Many other algorithms that try to address this issue presume each datapoint matters as much as every other datapoint. In this paper, we are showing that presumption is not true. There specify points in our dataset that are adding to this bias, and we can discover those data points, remove them, and get much better performance," says Kimia Hamidieh, an electrical engineering and computer science (EECS) graduate trainee at MIT and co-lead author of a paper on this [technique](https://justgoodfit.com).<br>
<br>She wrote the paper with co-lead authors Saachi Jain PhD '24 and fellow EECS graduate trainee Kristian Georgiev
Loading…
Cancel
Save