Sara Hooker, Nyalleng Moorosi, Gregory Clark, Samy
Bengio, and Emily Denton. 2020. Characterising
Bias in Compressed Models. CoRR, abs/2010.03058.
Masahiro Kaneko and Danushka Bollegala. 2021. De-
biasing Pre-trained Contextualised Embeddings. In
EACL, pages 1256–1266. Association for Computa-
tional Linguistics.
Masahiro Kaneko and Danushka Bollegala. 2022. Un-
masking the Mask - Evaluating Social Biases in
Masked Language Models. In AAAI, pages 11954–
11962. AAAI Press.
Masahiro Kaneko, Danushka Bollegala, and Naoaki
Okazaki. 2022. Debiasing Isn’t Enough! - on the
Effectiveness of Debiasing MLMs and Their Social
Biases in Downstream Tasks. In COLING, pages
1299–1310. International Committee on Computa-
tional Linguistics.
Sneha Kudugunta, Yanping Huang, Ankur Bapna,
Maxim Krikun, Dmitry Lepikhin, Minh-Thang Lu-
ong, and Orhan Firat. 2021. Beyond Distillation:
Task-level Mixture-of-Experts for Efficient Inference.
In EMNLP (Findings), pages 3577–3599. Associa-
tion for Computational Linguistics.
Keita Kurita, Nidhi Vyas, Ayush Pareek, Alan W.
Black, and Yulia Tsvetkov. 2019. Measuring Bias
in Contextualized Word Representations. CoRR,
abs/1906.07337.
Faisal Ladhak, Esin Durmus, Mirac Suzgun, Tianyi
Zhang, Dan Jurafsky, Kathleen R. McKeown, and
Tatsunori Hashimoto. 2023. When Do Pre-Training
Biases Propagate to Downstream Tasks? A Case
Study in Text Summarization. In EACL, pages 3198–
3211. Association for Computational Linguistics.
Paul Pu Liang, Irene Mengze Li, Emily Zheng,
Yao Chong Lim, Ruslan Salakhutdinov, and Louis-
Philippe Morency. 2020. Towards Debiasing Sen-
tence Representations. In ACL, pages 5502–5515.
Association for Computational Linguistics.
Lucas Liebenwein, Cenk Baykal, Brandon Carter, David
Gifford, and Daniela Rus. 2021. Lost in Pruning:
The Effects of Pruning Neural Networks beyond Test
Accuracy. In MLSys. mlsys.org.
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man-
dar Joshi, Danqi Chen, Omer Levy, Mike Lewis,
Luke Zettlemoyer, and Veselin Stoyanov. 2019.
RoBERTa: A Robustly Optimized BERT Pretrain-
ing Approach. CoRR, abs/1907.11692.
Nicholas Meade, Elinor Poole-Dayan, and Siva Reddy.
2022. An Empirical Survey of the Effectiveness of
Debiasing Techniques for Pre-trained Language Mod-
els. In ACL (1), pages 1878–1898. Association for
Computational Linguistics.
Moin Nadeem, Anna Bethke, and Siva Reddy. 2021.
StereoSet: Measuring stereotypical bias in pretrained
language models. In ACL/IJCNLP (1), pages 5356–
5371. Association for Computational Linguistics.
Nikita Nangia, Clara Vania, Rasika Bhalerao, and
Samuel R. Bowman. 2020. CrowS-Pairs: A Chal-
lenge Dataset for Measuring Social Biases in Masked
Language Models. In EMNLP (1), pages 1953–1967.
Association for Computational Linguistics.