The functioning of Google Translate shows how the outputs of data-driven services are, in many cases, intrinsically biased as a result of our own human biases. Ultimately, after extensive criticism, Google made some changes to its translation service. Now, instead of giving a single output when using the personal pronoun hän, it gives two – one for both of the main gender pronouns. Nonetheless, gender bias can still be seen when translating several sentences at once. For example, “Hän maksaa laskut. Hän hoitaa lapsia.” is still translated into “He pays the bills. She takes care of the children.” This goes to show that the slight modification does not necessarily prevent a similar problem from emerging. After all, the foundations of data-driven services remain largely the same. The biases embedded into algorithms via machine-learning reflect human knowledge and data. Hence, issues pertaining to algorithmic bias should not be directly pinned on programmers. The role of sources from which data are accumulated should be taken into consideration as well, as many of these biases originate there.
For instance, the Oxford English Dictionary, has been previously criticized due to its sexist synonyms regarding women. Specifically, the word “women” has been associated with condescending terms, such as “bitch.” It should be noted, however, that such ascriptions reflect more on the lack of revision of outdated words and definitions rather than the wrongdoings of contemporary wordsmiths.
Biases are intrinsically human. However, this built-in partiality should not be an excuse for embedding them within the products and services available to citizens of this modern world. At the end of the day, the contents of digital landscapes are known to operate tools that influence people’s behavior and perceptions, either consciously or unconsciously. Therefore, the extent to which these biases are incorporated in services such as translation software should be actively scrutinized in order to diminish their potential ramifications. Biases should not be left lingering in algorithmic codes, as these may have wide-ranging societal repercussions, such as creating preconceptions regarding gender or perpetuating existing stereotypes.