Gender-neutral Language Use in the Context of Gender Bias in Machine Translation (A Review Literature)

Authors

  • Aida Kostikova New Bulgarian University

DOI:

https://doi.org/10.33919/JCAL.23.1.5

Keywords:

gender bias, machine translation, NLP tools, gender-neutral language use, non-binary gender

Abstract

Gender bias has become one of the central issues analysed within natural language processing (NLP) research. A main concerns in this field relates to the fact that many NLP tools and automatic machine learning systems not only reflect, but also reinforce social disparities, including those related to gender, and language technology is one of the areas in which this issue is pronounced. This paper analyses the problem of gender-neutral language use from the standpoint of gender bias in machine translation (MT). We determine which types of harms can be caused by the failure to reflect gender-neutral language in translation, provide the general definition of gender bias in MT, describe its sources and provide an overview of existing mitigating strategies. One of the main contributions of this work is that it focuses not only on females, but also non-binary people, whose linguistic visibility has been receiving only limited attention from academia. This literature review provides a firm foundation for further research in this area aimed at addressing the problem of gender bias in machine translation, especially bias linked to representational harms.

References

Airton, L. (2018). The De/politicization of Pronouns: Implications of the No Big Deal Campaign for Gender-expansive Educational Policy and Practice. Gender and Education, 30(6), 790–810.

Alhafni, B., Habash, N. and Bouamor, H. (2020). Gender-aware Reinflection Using Linguistically Enhanced Neural Models. In: Proceedings of the Second Workshop on Gender Bias in Natural Language Processing, 139–150.

Bau, A., Belinkov, Y., Sajjad, H., Durrani, N., Dalvi, F. and Glass, J. (2019). Identifying and Controlling Important Neurons in Neural Machine Translation. In: Proceedings of the Seventh International Conference on Learning Representations (ICLR).

Barocas, S., Hardt, M. and Narayanan, A. (2017). Fairness in Machine Learning. Nips Tutorial, 1, 2.

Bender, E. M., Gebru, T., McMillan-Major, A. and Shmitchell, S. (2021). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 610–623.

Bolukbasi, T., Chang, K. W., Zou, J. Y., Saligrama, V. and Kalai, A. T. (2016). Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. In: Proceedings of the 30th International Conference on Neural Information Processing Systems (NIPS 2016). NY: Curran Associates Inc., 4356–4364.

Bonnin, J. E. and Coronel, A. A. (2021). Attitudes Toward Gender-Neutral Spanish: Acceptability and Adoptability. Frontiers in sociology, 6, 35.

Bourguignon, D., Yzerbyt, V. Y., Teixeira, C. P. and Herman, G. (2015). When Does it Hurt? Intergroup Permeability Moderates the Link Between Discrimination and Self‐esteem. European Journal of Social Psychology, 45(1), 3–9.

Caliskan, A., Bryson, J. J. and Narayanan, A. (2017). Semantics Derived Automatically from Language Corpora Contain Human-like Biases. Science, 356(6334), 183–186. Available at: http://opus.bath.ac.uk/55288/.

Cao, Y. T. and Daumé III, H. (2021). An Analysis of Gender and Bias Throughout the Machine Learning Lifecyle. Computational Linguistics, 47(3), 615–661. Available at: https://doi.org/10.1162/coli_a_00413.

Cordoba, S. (2020). Exploring non-binary genders: language and identity. [PhD diss.]. De Montfort University.

Crawford, K. (2017). The Trouble with Bias – NIPS 2017 Keynote – Kate Crawford #NIPS2017. YouTube. [Video]. Available at: https://www.youtube.com/watch?v=fMym_BKWQzk&t=10s.

Dev, S., Monajatipoor, M., Ovalle, A., Subramonian, A., Phillips, J. M. and Chang, K. W. (2021). Harms of Gender Exclusivity and Challenges in Non-binary Representation in Language Technologies. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 1968–1994.

Devinney, H. Björklund, J. and Björklund, H. (2020). Semi-Supervised Topic Modeling for Gender Bias Discovery in English and Swedish. In: Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, 79–92. Available at: https://aclanthology.org/2020.gebnlp-1.8/.

Dinan, E., Fan, A., Wu, L., Weston, J., Kiela, D. and Williams, A. (2020). Mul- tidimensional Gender Bias Classification. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 314–331.

Douglas, K. M. and Sutton, R. M. (2014). “A Giant Leap for Mankind” But What About Women? The Role of System-justifying Ideologies in Predicting Attitudes Toward Sexist Language. Journal of Language and Social Psychology, 33(6), 667–680.

EIGE – European Institute for Gener Equality. (2019). Toolkit on Gender-sensitive Communication. A resource for policymakers, legislators, media and anyone else with an interest in making their communication more inclusive. Publications Office of the European Union. Available at: https://eige.europa.eu/sites/default/files/20193925_mh0119609enn_pdf.pdf.

Fairclough, N. (2001) Language and Power. 2nd ed. Harlow: Pearson Education.

Ferrer, X., van Nuenen, T., Such, J. M., Coté, M. and Criado, N. (2021). Bias and Discrimination in AI: A Cross-Disciplinary Perspective. IEEE Technology and Society Magazine, 40(2), 72–80.

Font, J. E. and Costa-Jussa, M. R. (2019). Equalizing Gender Biases in Neural Machine Translation with Word Embeddings Techniques. In: Proceedings of the First Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, 147–154.

Friedman, B., and Nissenbaum, H. (1996). Bias in Computer Systems. ACM Transactions on Information Systems (TOIS), 14(3), 330–347.

Gaido, M., Savoldi, B., Bentivogli, L., Negri, M. and Turchi, M. (2020). Breeding Gender-aware Direct Speech Translation Systems. arXiv.org e-Print arXiv:2012.04955. Available at: https://arxiv.org/abs/2012.04955.

Garg, N., Schiebinger, L., Jurafsky, D. and Zou, J., 2018. Word Embeddings Quantify 100 years of Gender and Ethnic Stereotypes. In: Proceedings of the National Academy of Sciences, 115(16), E3635-E3644.

Gonen, H. and Goldberg, Y. (2019). Lipstick on a Pig: Debiasing Methods Cover up Systematic Gender Biases in Word Embeddings but do not Remove Them. arXiv.org e-Print arXiv:1903.03862. Available at: https://arxiv.org/abs/1903.03862.

Gustafsson Sendén, M., Bäck, E.A. and Lindqvist, A. (2015). Introducing a Gen- der-neutral Pronoun in a Natural Gender Language: The Influence of Time on Attitudes and Behavior. Frontiers in psychology, 6, 893.

Hamilton, M.C. (1991) Masculine Bias in the Attribution of Personhood: People = male, male = people. Psychology of Women Quarterly, 15(3), 393–402.

Harris, C. A., Biencowe, N. and Telem, D. A. (2017) What’s in a Pronoun? Why gender-fair Language Matters. Annals of Surgery, 266(6), 932.

Hausmann, R., Tyson, L. D., and Zahidi, S. (2009). The Global Gender Gap Re- port 2009. Geneva: World Economic Forum.

Hekanaho, Laura. (2020). Generic and nonbinary pronouns: usage, acceptability and attitudes. [PhD diss.]. Helsingfors University, Helsinki.

Hellinger, M., and Bußmann, H. (2001, 2002, 2003). Gender Across Languages: The Linguistic Representation of Women and Men, Vol. 1, 2, 3. John Benjamins Publishing Company.

Horvath, L. K., and Sczesny, S. (2016).Reducing Women’s Lack of Fit with Leadership? Effects of the Wording of Job Advertisements. European Journal of Work and Organizational Psychology, 25(2), 316–328.

Hovy, D. and Yang, D. (2021). The Importance of Modeling Social Factors of Language: Theory and Practice. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 588–602.

Hovy, D. and Spruit, S. L. (2016). The Social Impact of Natural Language Process- ing. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2 (Short Papers), 591–598.

Hovy, D., Bianchi, F. and Fornaciari, T. (2020). “You Sound Just Like Your Father” Commercial Machine Translation Systems Include Stylistic Biases. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Online: Association for Computational Linguistics, 1686–1690. Available at: https://aclanthology.org/2020.acl-main.154/.

Irmen, L. (2007). What’s in a (Role) Name? Formal and Conceptual Aspects of Comprehending Personal Nouns. Journal of Psycholinguistic Research, 36(6), 431–456.

Koch, B., Denton, E., Hanna, A. and Foster, J. G. (2021). Reduced, Reused and Recycled: The Life of a Dataset in Machine Learning Research. In: Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2). arXiv.org e-Print arXiv:2112.01716. Available at: https://arxiv.org/abs/2112.01716.

Koeser, S. and Sczesny, S. (2014). Promoting Gender-fair Language: The Impact of Arguments on Language Use, Attitudes, and Cognitions. Journal of Language and Social Psychology, 33(5), 548–560.

Levesque, R. J. (2011). Sex Roles and Gender Roles. In: Encyclopedia of Adolescence. Springer International Publishing, 2622–2623.

Lindqvist, A., Renström, E. A. and Gustafsson Sendén, M. (2019). Reducing a Male Bias in Language? Establishing the Efficiency of Three Different Gender-fair Language Strategies. Sex Roles, 81(1), 109–117.

López, Á (2020). Cuando el lenguaje excluye: consideraciones sobre el lenguaje no binario indirecto, Cuarenta naipes, (3), 295–312.

López, Á (2021). Direct and Indirect Non-binary Language in English to Spanish Translation. In: 27th Annual Lavender Languages and Linguistics Conference, Online, 21–23.

Maass, A., Suitner, C. and Merkel, E. M. (2013). Does Political Correctness Make (social) Sense? In: Forgas, J. P., Vincze, O. and László J., (eds.). Social Cognition and Communication. Psychology Press, 345–360.

María del Río-González, A. (2021) To Latinx or not to Latinx: A Question of Gender Inclusivity Versus Gender Neutrality. American Journal of Public Health, 111(6), 1018–1021.

Martindale, M.J. and Carpuat, M. (2018). Fluency Over Adequacy: A Pilot Study in Measuring User Trust in Imperfect MT. In: Proceedings of the 13th Conference of the Association for Machine Translation in the Americas (Volume 1: Research Track). Association for Machine Translation in the Americas, 13–25. Available at: https://aclanthology.org/W18-1803.pdf.

McGlashan, H. and Fitzpatrick, K. (2018). “I Use Any Pronouns, and I’m Ques- tioning Everything Else”: Transgender Youth and the Issue of Gender Pronouns. Sex Education, 18(3), 239–252.

Menegatti, M. and Rubini, M. (2017). Geneder Bias and Sexism in Language. In: Oxford Research Encyclopedia of Communication. Oxford University Press, 451–468.

Misiek, S. (2020) Misgendered in Translation? Genderqueerness Polish Translations of English-language Television Series. Anglica. An International Journal of English Studies, 29(2), 165–185.

Monti, J. (2020). Gender Issues in Machine Translation: An Unsolved Problem? In: von Flotow, L. and Hālah, K., (eds.). The Routledge Handbook of Translation, Feminism and Gender. Abingdon Oxon: Routledge, 457–468.

Moryossef, A., Aharoni, R. and Goldberg, Y. (2019). Filling Gender & Number Gaps in Neural Machine Translation with Black-box Context Injection. In: Proceedings of the First Workshop on Gender Bias in Natural Language Processing, 49–56. arXiv.org e-Print arXiv:1903.03467. Available at: https://arxiv.org/abs/1903.03467.

Habash, N., Bouamor, H. and Chung, C. (2019). Automatic Gender Identification and Reinflection in Arabic. In: Proceedings of the First Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, 155–165.

Olson, P. (2018). The Algorithm that Helped Google Translate Become Sexist. Forbes. [Online]. Available at: https://www.forbes.com/sites/parmyolson/2018/02/15/the-algorithm-that-helped-google-translate-become-sexist/?sh=d675b9c7daa2.

Papadimoulis, D. (2018). Gender-neutral Language in the European Parliament.Brussels: European Parliament.

Paullada, A., Raji, I. D., Bender, E. M., Denton, E. and Hanna, A. (2021). Data and its (Dis)contents: A Survey of Dataset Development and Use in Machine Learning Research. Patterns, 2(11), 100336.

Prates, M. O. R., Avelar, P. H. and Lamb, L. C. (2020). Assessing Gender Bias in Machine Translation: A Case Study with Google Translate. Neural Comput & Applic, 32, 6363– 6381. Available at: https://doi.org/10.1007/s00521-019-04144-6.

Reddy, S. and Knight, K. (2016). Obfuscating Gender in Social Media Writing. In: Proceedings of 2016 EMNLP Workshop on Natural Language Processing and Computational Social Science. Association for Computational Linguistics, 17–26.

Régner, I., Thinus-Blanc, C., Netter, A., Schmader, T. and Huguet, P. (2019). Committees with Implicit Biases Promote Fewer Women When they do not Believe Gender Bias Exists. Nature Human Behavior, 3(11), 1171–1179.

Richards, C. and Barker, M. J. (2015). The Palgrave Handbook of the Psychology of Sexuality and Gender. Palgrave Macmillan.

Rudinger, R., Naradowsky, J., Leonard, B., Van Durme, B. (2018). Gender Bias in Coreference Resolution. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2 (Short Papers). Association for Computational Linguistics, 8–14.

Sap, M., Gabriel, S., Qin, L., Jurafsky, D., Smith, N.A. and Choi, Y. (2019). So- cial Bias Frames: Reasoning about Social and Power Implications of Language. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 5477–5490. arXiv.org e-Print arXiv:1911.03891. Available at: https://arxiv.org/abs/1911.03891.

Saunders, D., Sallis, R., and Byrne, B. (2020). Neural Machine Translation doesn’t Translate Gender Coreference Right Unless You Make It. In: Proceedings of the Second Workshop on Gender Bias in Natural Language Processing, 35–43.

Savoldi, B., Gaido, M., Bentivogli, L., Negri, M., Turchi, M. (2021). Gender Bias in Machine Translation. In: Transactions of the Association for Computational Lin- guistics, 9, 845–874.

Schiebinger, L. (2014). Scientific Research Must Take Gender into Account. Nature, 507, 9.

Schnoebelen, T. (2017). Goal-oriented Design for Ethical Machine Learning and NLP. In: Proceedings of the First ACL Workshop on Ethics in Natural Language Processing, 88–93.

Sczesny, S., Formanowicz, M. and Moser, F. (2016). Can Gender-fair Language Reduce Gender Stereotyping and Discrimination? Frontiers in psychology, 7, 25.

Silveira, J. (1980). Generic Masculine Words and Thinking. Women‘s Studies In- ternational Quarterly, 3(2-3), 165–178.

Stahlberg, D., Braun, F., Irmen, L. and Sczesny, S. (2007). Representation of the Sexes in Language. In: Fiedler, K. (ed.). Social communication. Psychology Press, 163–187.

Stahlberg, D., Sczesny, S. and Braun, F. (2001). Name Your Favorite Musician: Effects of Masculine Generics and of Their Alternatives in German. Journal of Language and Social Psychology, 20(4), 464–469.

Stout, J. G. and Dasgupta, N. (2011). When He doesn’t Mean You: Gender-exclusive Language as Ostracism, Personality and Social Psychology Bulletin, 37(6), 757–769.

Sun, T., Webster, K., Shah, A., Wang, W. Y. and Johnson, M. (2021). They, Them, Theirs: Rewriting with Gender-neutral Еnglish. arXiv.org e-Print arXiv:2102.06788. Available at: https://arxiv.org/abs/2102.06788.

Switzer, J. Y. (1990). The Impact of Generic Word Choices: An Empirical Investigation of Age- and Sex-related Differences. Sex Roles, 22(172), 69–81.

Tallon, T. (2019). A Century of “shrill”: How Bias in Technology has Hurt Women’s Voices. The New Yorker. Available at: https://www.newyorker.com/culture/cultural-comment/a-century-of-shrill-how-bias-in-technology-has-hurt-womens-voices.

Trainer, T. (2021). The (non) Binary of Success and Failure: A Corpus-based Evaluation of the European Parliament‘s Commitment to Using Gender-neutral Language in Legislation Published in English and Portuguese. [Master's thesis]. University of Porto.

Turchi, M., Negri, M., Farajian, M. and Federico, M. (2017). Continuous Learning from Human Post-edits for Neural Machine Translation. The Prague Bulletin of Mathematical Linguistics, 108, 233–244.

Uppunda, A., Cochran, S.D., Foster, J. G., Arseniev-Koehler, A., Mays, V. M. and Chang, K. (2021). Adapting Coreference Resolution for Processing Violent Death Narratives. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 4553–4559. arXiv.org e-Print arXiv:2104.14703. Available at: https://arxiv.org/abs/2104.14703.

Vanmassenhove, E., Emmery, C. and Shterionov, D. (2021). NeuTral Rewriter: A Rule-Based and Neural Approach to Automatic Rewriting into Gender-Neutral Alternatives. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 8940–8948. arXiv.org e-Print arXiv:2109.06105. Available at: https://arxiv.org/abs/2109.06105.

Vanmassenhove, E., Hardmeier, C. and Way, A. (2018). Getting Gender Right in Neural Machine Translation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 3003–3008. arXiv.org preprint arXiv:1909.05088. Available at: https://arxiv.org/abs/1909.05088.

Waseem, Z. (2016). Are You a Racist or Am I Seeing Things? Annotator Influence on Hate Speech Detection on Twitter. In: Proceedings of the First Workshop on NLP and Computational Social Science. Association for Computational Linguistics, 138–142.

Wasserman, B. D., and Weseley, A. J. (2009). ¿Qué? Quoi? Do Languages with Grammatical Gender Promote Sexist Attitudes? Sex Roles, 61, 634–643.

Webster, K., Recasens, M., Axelrod, V. and Baldridge, J. (2018). Mind the GAP: A Balanced Corpus of Gendered Ambiguous Pronouns. Transactions of the Associa- tion for Computational Linguistics, 6, 605–617.

Zhao, J., Wang, T., Yatskar, M., Ordonez, V. and Chang, K. W. (2017). Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus-level Constraints. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2979–2989. arXiv.org e-Print arXiv:1707.09457. Available at: https://arxiv.org/abs/1707.09457.

Zhao, J., Wang, T., Yatskar, M., Ordonez, V. and Chang, K. W. (2018). Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2 (Short Papers). Association for Computational Linguistics, 15–20. arXiv.org e-Print arXiv:1804.06876. Available at: https://arxiv.org/abs/1804.06876.

Zimmer, B. and Carson, C. E. (2012). Among the New Words. American speech, 87(4), 491–510.

Downloads

Published

18.07.2023

How to Cite

Kostikova, A. (2023). Gender-neutral Language Use in the Context of Gender Bias in Machine Translation (A Review Literature). Journal of Computational and Applied Linguistics, 1, 94–109. https://doi.org/10.33919/JCAL.23.1.5