Skip to main content

Gender Bias in Machine Translation Systems

  • Chapter
  • First Online:

Part of the book series: Social and Cultural Studies of Robots and AI ((SOCUSRA))

Abstract

In recent years, headlines such as ‘Is Google Translate Sexist?’ (Mail Online, Is Google translate sexist? Users report biased results when translating gender-neutral languages into English in 2017) or ‘The Algorithm that Helped Google Translate Become Sexist’ (Olson, The Algorithm that Helped Google Translate Become Sexist in 2018) have appeared in the technology sections of the world’s news providers. The nature of our highly interconnected world has made online translators indispensable tools in our daily lives. However, their output has the potential to cause great social harm. Due to the continuous pursuit to create ever larger language models and, as a consequence thereof, the opaque nature of unsupervised training datasets, language-based AI systems, such as online translators, can easily produce biased content. If left unchecked, this will inevitable have detrimental consequences. This chapter addresses the nature, impact and risks of bias in training data by looking at the concrete example of gender bias in machine translation (MT). The first section will provide an introduction to recent proposals for ethical AI guidelines in different sectors and the field of natural language processing (NLP) will be presented. Next, I will explain different types of bias in machine learning and how they can manifest themselves in language models. This is followed by presenting the results of a corpus-linguistic analysis I performed of a sample dataset that was later used to train a MT system. I will explore the gender-related imbalances in the corpus that are likely to give rise to biased results. In the final section of this chapter, I will discuss different approaches to reduce gender bias in MT and present findings from a set of experiments my colleagues and I conducted ourselves to mitigate bias in MT. The research presented in this chapter takes a highly interdisciplinary approach, as it takes expertise from linguistics, philosophy, computer science and engineering in order to successfully dismantle and solve the complex problem of bias in NLP.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   129.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    See Engström’s chapter in this anthology for a critique of these international frameworks.

  2. 2.

    Note, however, that not all language communities and countries have equal access to NLP systems (Siavoshi 2020).

  3. 3.

    The most common equivalent of a female ‘nurse’ in German is Krankenschwester or its abbreviated form Schwester. However, the counts for other, less frequent terms like Arzthelferin, were also included here. This is simply due to the fact that the English ‘nurse’ has multiple equivalents in German. Also, the German Schwester may also be translated into the English ‘sister’. These, however, were kept separate in the word counts.

  4. 4.

    These approaches all address binary linguistic realisations of gender and more work is needed to ensure improved non-binary inclusivity in these systems (Darwin 2017; Ackerman 2019).

References

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Stefanie Ullmann .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Ullmann, S. (2022). Gender Bias in Machine Translation Systems. In: Hanemaayer, A. (eds) Artificial Intelligence and Its Discontents. Social and Cultural Studies of Robots and AI. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-030-88615-8_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-88615-8_7

  • Published:

  • Publisher Name: Palgrave Macmillan, Cham

  • Print ISBN: 978-3-030-88614-1

  • Online ISBN: 978-3-030-88615-8

  • eBook Packages: Social SciencesSocial Sciences (R0)

Publish with us

Policies and ethics