Abstract
Recent studies have suggested that weight pruning, e.g. using lottery ticket extraction techniques (Frankle and Carbin, 2018), comes at the risk of compromising the group fairness of machine learning models (Paganini, 2020; Hooker et al., 2020), but to the best of our knowledge, no one has empirically evaluated this hypothesis at scale in the context of natural language processing. We present experiments with two text classification datasets annotated with demographic information: the Trustpilot Corpus (sentiment) and CivilComments (toxicity). We evaluate the fairness of lottery ticket extraction through layer-wise and global weight pruning across three languages and two tasks. Our results suggest that there is a small increase in group disparity, which is most pronounced at high pruning rates and correlates with instability. The fairness of models trained with distributionally robust optimization objectives is sometimes less sensitive to pruning, but results are not consistent. The code for our experiments is available at https://github.com/vpetren/fairness_lottery.
Originalsprog | Engelsk |
---|---|
Titel | Findings of the Association for Computational Linguistics : ACL-IJCNLP 2021 |
Redaktører | Chengqing Zong, Fei Xia, Wenjie Li, Roberto Navigli |
Forlag | Association for Computational Linguistics |
Publikationsdato | 2021 |
Sider | 3214-3224 |
ISBN (Elektronisk) | 9781954085541 |
DOI | |
Status | Udgivet - 2021 |
Begivenhed | Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 - Virtual, Online Varighed: 1 aug. 2021 → 6 aug. 2021 |
Konference
Konference | Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 |
---|---|
By | Virtual, Online |
Periode | 01/08/2021 → 06/08/2021 |
Bibliografisk note
Publisher Copyright:© 2021 Association for Computational Linguistics