Abstract
Recent studies have suggested that weight pruning, e.g. using lottery ticket extraction techniques (Frankle and Carbin, 2018), comes at the risk of compromising the group fairness of machine learning models (Paganini, 2020; Hooker et al., 2020), but to the best of our knowledge, no one has empirically evaluated this hypothesis at scale in the context of natural language processing. We present experiments with two text classification datasets annotated with demographic information: the Trustpilot Corpus (sentiment) and CivilComments (toxicity). We evaluate the fairness of lottery ticket extraction through layer-wise and global weight pruning across three languages and two tasks. Our results suggest that there is a small increase in group disparity, which is most pronounced at high pruning rates and correlates with instability. The fairness of models trained with distributionally robust optimization objectives is sometimes less sensitive to pruning, but results are not consistent. The code for our experiments is available at https://github.com/vpetren/fairness_lottery.
Original language | English |
---|---|
Title of host publication | Findings of the Association for Computational Linguistics : ACL-IJCNLP 2021 |
Editors | Chengqing Zong, Fei Xia, Wenjie Li, Roberto Navigli |
Publisher | Association for Computational Linguistics |
Publication date | 2021 |
Pages | 3214-3224 |
ISBN (Electronic) | 9781954085541 |
DOIs | |
Publication status | Published - 2021 |
Event | Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 - Virtual, Online Duration: 1 Aug 2021 → 6 Aug 2021 |
Conference
Conference | Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 |
---|---|
City | Virtual, Online |
Period | 01/08/2021 → 06/08/2021 |
Bibliographical note
Publisher Copyright:© 2021 Association for Computational Linguistics