Abstract
It is widely recognized that machine learning algorithms may be biased in the sense that they perform worse on some demographic groups than others. This motivates algorithmic development to remove algorithmic bias, which in turn might lead to a hope—even an expectation—that algorithmic bias can be mitigated or removed (1). In this short comment, we make three points to qualify Wang et al.’s suggestion: 1) It may not be possible for algorithms to perform equally well across groups on all measures, 2) which inequalities count as morally unacceptable bias is an ethical question, and 3) the answer to the ethical question will vary across decision contexts.
Original language | English |
---|---|
Article number | e2304710120 |
Journal | Proceedings of the National Academy of Sciences of the United States of America |
Volume | 120 |
Issue number | 23 |
Number of pages | 1 |
ISSN | 0027-8424 |
DOIs | |
Publication status | Published - 2023 |
Keywords
- Bias