skip to main content
research-article
Open access

Theories of “Gender” in NLP Bias Research

Published: 20 June 2022 Publication History

Abstract

The rise of concern around Natural Language Processing (NLP) technologies containing and perpetuating social biases has led to a rich and rapidly growing area of research. Gender bias is one of the central biases being analyzed, but to date there is no comprehensive analysis of how “gender” is theorized in the field. We survey nearly 200 articles concerning gender bias in NLP to discover how the field conceptualizes gender both explicitly (e.g. through definitions of terms) and implicitly (e.g. through how gender is operationalized in practice). In order to get a better idea of emerging trajectories of thought, we split these articles into two sections by time.
We find that the majority of the articles do not make their theorization of gender explicit, even if they clearly define “bias.” Almost none use a model of gender that is intersectional or inclusive of nonbinary genders; and many conflate sex characteristics, social gender, and linguistic gender in ways that disregard the existence and experience of trans, nonbinary, and intersex people. There is an increase between the two time-sections in statements acknowledging that gender is a complicated reality, however, very few articles manage to put this acknowledgment into practice. In addition to analyzing these findings, we provide specific recommendations to facilitate interdisciplinary work, and to incorporate theory and methodology from Gender Studies. Our hope is that this will produce more inclusive gender bias research in NLP.

References

[1]
Lauren Ackerman. 2019. Syntactic and cognitive issues in investigating gendered coreference. Glossa: a journal of general linguistics 4, 1 (oct 2019), 27 pages. https://doi.org/10.5334/gjgl.721
[2]
Su Lin Blodgett, Solon Barocas, Hal Daumé Iii, and Hanna Wallach. 2020. Language (Technology) is Power: A Critical Survey of ”Bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5454–5476. https://aclanthology.org/2020.acl-main.485/
[3]
Tolga Bolukbasi, Kai Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In Advances in Neural Information Processing Systems. NeurIPS, Barcelona, Spain, 4356–4364. arxiv:1607.06520http://papers.nips.cc/paper/6228-man-is-to-computer-programmer-as-woman-is-to-homemaker-d
[4]
Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency(Proceedings of Machine Learning Research, Vol. 81), Sorelle A. Friedler and Christo Wilson (Eds.). PMLR, New York, New York, USA, 77–91. https://proceedings.mlr.press/v81/buolamwini18a.html
[5]
Judith Butler. 1999. Gender Trouble. Routledge, New York.
[6]
Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases.Science (New York, N.Y.) 356, 6334 (apr 2017), 183–186. https://doi.org/10.1126/science.aal4230
[7]
Yang Trista Cao and Hal Daumé III. 2020. Toward Gender-Inclusive Coreference Resolution. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 4568–4595. https://doi.org/10.18653/v1/2020.acl-main.418
[8]
Serina Chang and Kathleen McKeown. 2019. Automatically Inferring Gender Associations from Language. In Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Hong Kong, 5746–5752. https://doi.org/10.18653/v1/D19-1579
[9]
Raewyn W. Connell and Rebecca Pearse. 2015. Gender: In World Perspective. Polity Press, Cambridge, UK.
[10]
Kirby Conrod. 2020. How to do things with gender. https://www.youtube.com/watch?v=jVr8NJwcMH4 Talk given for English Language and Linguistics at the University of Kent.
[11]
Greville G. Corbett. 2013. Number of Genders. In The World Atlas of Language Structures Online, Matthew S. Dryer and Martin Haspelmath (Eds.). Max Planck Institute for Evolutionary Anthropology, Leipzig. https://wals.info/chapter/30
[12]
Marta R. Costa-jussà, Christian Hardmeier, Will Radford, and Kellie Webster (Eds.). 2020. Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Barcelona, Spain (Online). https://www.aclweb.org/anthology/2020.gebnlp-1.0
[13]
Kimberlé Crenshaw. 1991. Mapping the Margins: Intersectionality, Identity Politics, and Violence against Women of Color. Stanford Law Review 43, 6 (1991), 1241–1299.
[14]
Erenay Dayanık and Sebastian Padó. 2021. Disentangling Document Topic and Author Gender in Multiple Languages: Lessons for Adversarial Debiasing. In Proceedings of the Eleventh Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis. Association for Computational Linguistics, Online, 50–61. https://aclanthology.org/2021.wassa-1.6
[15]
Sunipa Dev, Masoud Monajatipoor, Anaelia Ovalle, Arjun Subramonian, Jeff M Phillips, and Kai-Wei Chang. 2021. Harms of Gender Exclusivity and Challenges in Non-Binary Representation in Language Technologies. In Proceedings ofthe 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, 1968–1994. https://aclanthology.org/2021.emnlp-main.150/
[16]
Hannah Devinney, Jenny Björklund, and Henrik Björklund. 2020. Semi-Supervised Topic Modeling for Gender Bias Discovery in English and Swedish. In Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Barcelona, Spain (Online), 79–92. https://www.aclweb.org/anthology/2020.gebnlp-1.8
[17]
Catherine D’Ignazio and Lauren F Klein. 2020. Data Feminism. The MIT Press, Cambridge, Massachusetts.
[18]
Emily Dinan, Angela Fan, Adina Williams, Jack Urbanek, Douwe Kiela, and Jason Weston. 2020. Queens are powerful too: Mitigating gender bias in dialogue generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, Online, 8173–8188. https://doi.org/10.18653/v1/2020.emnlp-main.656
[19]
Emily Dinan, Angela Fan, Ledell Wu, Jason Weston, Douwe Kiela, and Adina Williams. 2020. Multi-Dimensional Gender Bias Classification. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, Online, 314–331. https://doi.org/10.18653/v1/2020.emnlp-main.23
[20]
Agnieszka Falenska and Özlem Çetinoğlu. 2021. Assessing Gender Bias in Wikipedia: Inequalities in Article Titles. In Proceedings of the 3rd Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Online, 75–85. https://doi.org/10.18653/v1/2021.gebnlp-1.9
[21]
Anne Fausto-Sterling. 1992. Myths of Gender (2ed.). Basic Books, New York, New York.
[22]
Anne Fausto-Sterling. 2000. Sexing the Body (1ed.). Basic Books, New York, New York.
[23]
Nikhil Garg, Londa Schiebinger, Dan Jurafsky, and James Zou. 2018. Word embeddings quantify 100 years of gender and ethnic stereotypes. Proceedings of the National Academy of Sciences of the United States of America 115, 16 (apr 2018), E3635–E3644. https://doi.org/10.1073/pnas.1720347115 arXiv:1711.08412
[24]
Ben Green. 2019. ”Good” isn’t good enough. In NeurIPS Joint Workshop on AI for Social Good. NeurIPS, Vancouver, Canada, 7 pages. https://aiforsocialgood.github.io/neurips2019/accepted/track3/pdfs/67_aisg_neurips2019.pdf
[25]
Stuart Hall. 2013. The Work of Representation. In Representation, Stuart Hall, Jessica Evans, and Sean Nixon (Eds.). Sage, 1–59.
[26]
Saga Hansson, Konstantinos Mavromatakis, Yvonne Adesam, Gerlof Bouma, and Dana Dannélls. 2021. The Swedish Winogender Dataset. In Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa). Linköping University Electronic Press, Sweden, Reykjavik, Iceland (Online), 452–459. https://aclanthology.org/2021.nodalida-main.52
[27]
Donna Haraway. 1988. Situated Knowledges: The Science Question in Feminism and the Privilege of Partial Perspective. Feminist Studies 14, 3 (1988), 575–599. https://doi.org/10.2307/3178066
[28]
Myra J. Hird. 2000. Gender’s nature: Intersexuality, transsexualism, and the ‘sex’/‘gender’ binary. Feminist Theory 1, 3 (2000), 347–34.
[29]
Anna Lauren Hoffmann. 2021. Terms of inclusion: Data, discourse, violence. New Media & Society 23, 12 (2021), 3539–3556. https://doi.org/10.1177/1461444820958725
[30]
bell hooks. 2000. Feminism is for Everybody. Pluto Press, London, UK.
[31]
Human Rights Council. 2011. Discriminatory laws and practices and acts of violence against individuals based on their sexual orientation and gender identity: REport of the United Nations High Commissioner for Human Rights. Technical Report. United Nations. Report No. A/HRC/19/41.
[32]
Masahiro Kaneko and Danushka Bollegala. 2019. Gender-preserving Debiasing for Pre-trained Word Embeddings. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL), Florence, Italy, 1641–1650. https://doi.org/10.18653/v1/P19-1160 arXiv:1906.00742
[33]
Dongyeop Kang, Varun Gangal, and Eduard Hovy. 2019. (Male, Bachelor) and (Female, Ph.D) have different connotations: Parallelly Annotated Stylistic Language Dataset with Multiple Personas. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Langauge Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, 1696–1706. https://www.aclweb.org/anthology/D19-1179/
[34]
Os Keyes. 2018. The Misgendering Machines: Trans/HCI Implications of Automatic Gender Recognition. Proceedings of the ACM on Human-Computer Interaction 2 (2018), 22. https://doi.org/10.1145/3274357
[35]
Svetlana Kiritchenko and Saif M. Mohammad. 2018. Examining Gender and Race Bias in Two Hundred Sentiment Analysis Systems. In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics. Association for Computational Linguistics, New Orleans, Louisiana, USA, 43–53. https://www.aclweb.org/anthology/S18-2005/
[36]
Vid Kocijan, Oana-Maria Camburu, and Thomas Lukasiewicz. 2020. The Gap on GAP: Tackling the Problem of Differing Data Distributions in Bias-Measuring Datasets. In Proceedings of the 35th AAAI Conference on Artificial Intelligence‚ AAAI 2021. AAAI, Online, 9727–9736. arxiv:2011.01837https://arxiv.org/abs/2011.01837v3
[37]
Brian Larson. 2017. Gender as a Variable in Natural-Language Processing: Ethical Considerations. In Proceedings of the First ACL Workshop on Ethics in Natural Language Processing. Association for Computational Linguistics, Stroudsburg, PA, USA, 1–11. https://doi.org/10.18653/v1/W17-1601
[38]
Kaiji Lu, Piotr Mardziel, Fangjing Wu, Preetam Amancharla, and Anupam Datta. 2020. Gender Bias in Neural Natural Language Processing. In Logic, Language, and Security. Springer International Publishing, Cham, 189–202. https://doi.org/10.1007/978-3-030-62077-6_14
[39]
Helma Lutz, Maria Teresa Herrera Vivar, and Linda Supik (Eds.). 2011. Framing Intersectionality. Ashgate Publishing Limited, Farnham, England.
[40]
Anne Maass and Luciano Arcuri. 1996. Language and Stereotyping. In Stereotypes and Stereotyping, C. Niel Macra, Charles Strangor, and Miles Hewstone (Eds.). Guilford Press, New York, NY, Chapter 6, 193–225.
[41]
Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, and Aram Galstyan. 2019. A Survey on Bias and Fairness in Machine Learning. Technical Report. University of Southern California, Information Sciences Institute. arxiv:1908.09635http://arxiv.org/abs/1908.09635
[42]
Malin Mobjörk, Camilla Berglund, Mikael Granberg, Magnus Johansson, Margareta Dahlström, Jon Moen, Lars Nyberg, and Mariele Evers. 2019. Facilitating Doctoral Education in Cross-disciplinary Milieus: Experiences from PhD-candidates. Technical Report. Karlstads University.
[43]
Robert Munro and Alex (Carmen) Morrison. 2020. Detecting Independent Pronoun Bias with Partially-Synthetic Data Generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (ACL), Online, 2011–2017. https://doi.org/10.18653/V1/2020.EMNLP-MAIN.157
[44]
Krithika Ramesh, Gauri Gupta, and Sanjay Singh. 2021. Evaluating Gender Bias in Hindi-English Machine Translation. In Proceedings of the 3rd Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics (ACL), Online, 16–23. https://doi.org/10.18653/V1/2021.GEBNLP-1.3
[45]
Damien W Riggs, Carla A Pfeffer, Ruth Pearce, Sally Hines, and Francis Ray White. 2020. Men, trans/masculine, and non-binary people negotiating conception: Normative resistance and inventive pragmatism. International Journal of Transgender Health 22, 1-2 (sep 2020), 6–17. https://doi.org/10.1080/15532739.2020.1808554
[46]
Maarten Sap, Saadia Gabriel, Lianhui Qin, Dan Jurafsky, Noah A. Smith, and Yejin Choi. 2020. Social Bias Frames: Reasoning about Social and Power Implications of Language. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5477–5490. https://doi.org/10.18653/v1/2020.acl-main.486 arXiv:1911.03891
[47]
Beatrice Savoldi, Marco Gaido, Luisa Bentivogli, Matteo Negri, and Marco Turchi. 2021. Gender Bias in Machine Translation. Transactions of the Association for Computational Linguistics 9 (aug 2021), 845–874. https://doi.org/10.1162/TACL_A_00401 arxiv:2104.06001
[48]
Chase Strangio. 2018. Deadly Violence Against Transgender People Is on the Rise. The Government Isn’t Helping.https://www.aclu.org/blog/lgbt-rights/criminal-justice-reform-lgbt-people/deadly-violence-against-transgender-people-rise
[49]
Tony Sun, Andrew Gaut, Shirlyn Tang, Yuxin Huang, Mai ElSherief, Jieyu Zhao, Diba Mirza, Elizabeth Belding, Kai-Wei Chang, and William Yang Wang. 2019. Mitigating Gender Bias in Natural Language Processing: Literature Review. In ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference. Association for Computational Linguistics (ACL), Florence, Italy, 1630–1640. arxiv:1906.08976http://arxiv.org/abs/1906.08976
[50]
Nathaniel Swinger, Maria De-Arteaga, Neil Thomas Heffernan IV, Mark DM Leiserson, and Adam Tauman Kalai. 2019. What are the Biases in My Word Embedding?. In Artificial Intelligence, Ethics, and Society. Association for Computing Machinery, New York, NY, USA, 305–311. https://doi.org/10.1145/3306618.3314270
[51]
Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Yaron Singer, and Stuart Shieber. 2020. Investigating Gender Bias in Language Models Using Causal Mediation Analysis. In 34th Conference on Neural Information Processing Systems (NeurIPS 2020). NeurIPS, Vancouver, Canada, 14 pages. https://www.cs.technion.ac.il/~belinkov/assets/pdf/neurips2020.pdf
[52]
Catherine Yeo and Alyssa Chen. 2020. Defining and Evaluating Fair Natural Language Generation. In Proceedings of the The Fourth Widening Natural Language Processing Workshop at ACL. Association for Computational Linguistics, Seattle, USA, 107–109. https://doi.org/10.18653/v1/2020.winlp-1.27 arxiv:2008.01548
[53]
Guanhua Zhang, Bing Bai, Junqi Zhang, Kun Bai, Conghui Zhu, and Tiejun Zhao. 2020. Demographics Should Not Be the Reason of Toxicity: Mitigating Discrimination in Text Classifications with Instance Weighting. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 4134–4145. https://doi.org/10.18653/v1/2020.acl-main.380
[54]
Jieyu Zhao, Yichao Zhou, Zeyu Li, Wei Wang, and Kai-Wei Chang. 2018. Learning Gender-Neutral Word Embeddings. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Brussels, Belgium, 4847–4853. https://doi.org/10.18653/v1/D18-1521
[55]
Pei Zhou, Weijia Shi, Jieyu Zhao, Kuan-Hao Huang, Muhao Chen, Ryan Cotterell, and Kai-Wei Chang. 2019. Examining Gender Bias in Languages with Grammatical Gender. In EMNLP-IJCNLP 2019 - 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference. Association for Computational Linguistics, Hong Kong, 5276–5284. arxiv:1909.02224http://arxiv.org/abs/1909.02224

Cited By

View all
  • (2024)Assessing and Mitigating Bias in Artificial Intelligence: A ReviewRecent Advances in Computer Science and Communications10.2174/266625581666623052311442517:1Online publication date: Jan-2024
  • (2024)CIDER: Context-sensitive polarity measurement for short-form textPLOS ONE10.1371/journal.pone.029949019:4(e0299490)Online publication date: 18-Apr-2024
  • (2024)Bias and Fairness in Large Language Models: A SurveyComputational Linguistics10.1162/coli_a_0052450:3(1097-1179)Online publication date: 1-Sep-2024
  • Show More Cited By

Index Terms

  1. Theories of “Gender” in NLP Bias Research
    Index terms have been assigned to the content through auto-classification.

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    FAccT '22: Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency
    June 2022
    2351 pages
    ISBN:9781450393522
    DOI:10.1145/3531146
    This work is licensed under a Creative Commons Attribution-NonCommercial International 4.0 License.

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 20 June 2022

    Check for updates

    Author Tags

    1. gender bias
    2. gender studies
    3. natural language processing

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    FAccT '22
    Sponsor:

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)1,319
    • Downloads (Last 6 weeks)160
    Reflects downloads up to 22 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Assessing and Mitigating Bias in Artificial Intelligence: A ReviewRecent Advances in Computer Science and Communications10.2174/266625581666623052311442517:1Online publication date: Jan-2024
    • (2024)CIDER: Context-sensitive polarity measurement for short-form textPLOS ONE10.1371/journal.pone.029949019:4(e0299490)Online publication date: 18-Apr-2024
    • (2024)Bias and Fairness in Large Language Models: A SurveyComputational Linguistics10.1162/coli_a_0052450:3(1097-1179)Online publication date: 1-Sep-2024
    • (2024)Transforming Dutch: Debiasing Dutch Coreference Resolution Systems for Non-binary PronounsProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3659049(2470-2483)Online publication date: 3-Jun-2024
    • (2024)Akal Badi ya Bias: An Exploratory Study of Gender Bias in Hindi Language TechnologyProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3659017(1926-1939)Online publication date: 3-Jun-2024
    • (2024)Gender Bias Detection in Court Decisions: A Brazilian Case StudyProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658937(746-763)Online publication date: 3-Jun-2024
    • (2024)Data Feminism for AIProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658543(100-112)Online publication date: 3-Jun-2024
    • (2024)Unlocking Bias Detection: Leveraging Transformer-Based Models for Content AnalysisIEEE Transactions on Computational Social Systems10.1109/TCSS.2024.339246911:5(6422-6434)Online publication date: Oct-2024
    • (2024)FairDeDup: Detecting and Mitigating Vision-Language Fairness Disparities in Semantic Dataset Deduplication2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR52733.2024.01319(13905-13916)Online publication date: 16-Jun-2024
    • (2024)A Security Risk Taxonomy for Prompt-Based Interaction With Large Language ModelsIEEE Access10.1109/ACCESS.2024.345038812(126176-126187)Online publication date: 2024
    • Show More Cited By

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media