A Chinese-language Google user types in “Buddhism” and launches a search. A French-language user searches the same term, as does an English-language user. Will all three get the same results? New Harvard research says not necessarily, finding online search responses can vary significantly and even conflict depending on the topic and language of the query.
The variation is a result of a hidden “language bias” embedded in the search algorithms of Google, ChatGPT, YouTube, and Wikipedia, says Queenie Luo, who studies artificial intelligence ethics and early Chinese history and is co-author of a paper with her Ph.D. adviser, Michael Puett, the Walter C. Klein Professor of Chinese History and Anthropology, and Michael D. Smith, former dean of the Faculty of Arts & Sciences who now teaches at the Harvard Paulson School of Engineering and Applied Sciences (SEAS).
That bias, they contend, distorts users’ understanding of search topics by limiting their exposure to a full range of information and viewpoints and raises questions about larger implications for relations between nations and peoples.
The Gazette spoke with Luo, who earned a master’s degree from SEAS in data science in 2023, about language bias in searches and the potential social and political harms that can arise from this hidden filter in major search platforms. The interview has been edited for clarity and length.
You tested a number of search terms, but the paper focuses on two that are extremely complex and abstract, Buddhism and liberalism. What did you find on Google?
Buddhism is a global religion that has developed distinct cultural traditions in different language communities worldwide. Chinese Buddhism is very different than Japanese Buddhism, Thai Buddhism, and Vietnamese Buddhism. Essentially, Western Buddhism has evolved into another branch of Buddhism over the past two centuries.
Our research found that when searching for Buddhism-related phrases on Google using different languages, the top-ranked websites tend to reflect the dominant Buddhist tradition of the query’s language community.
The general assumption among internet users is that they can access global information through search engines, with Google often perceived as providing objective, unfiltered results. However, our research found that as Google and most online platforms use search language as a significant filter, different language users end up reading highly distinct information online.
Such problems aren’t just limited to Buddhism-related queries, but extend to a wide range of topics, such as liberalism and international trade policy. For example, if you search liberalism using English on Google, you tend to get very positive views about liberalism and nearly no references to neoliberalism. The concepts of free market, human rights, and equality tend to be strongly emphasized in English-language search results.
However, if you switch your search language to Chinese, Google’s top-ranked search results on liberalism tend to benegative and frequently connect it to neoliberalism. Similarly, when you ask Google, “What constitutes good economic policy” using a European language like Italian or French, the top-ranked websites tend to emphasize aspects such as protective market economy, but the top-ranked websites would instead focus on “free market economy” or “limited government intervention” when you search using English.
These ideas aren’t mutually exclusive but can be contradictory based on context.
What about other platforms people use to search, like ChatGPT?
Things changed with ChatGPT. As ChatGPT is predominantly trained on English-language data, it always presents the Anglo-American perspectives by default. The version integrated with Bing behaves similarly to Google in that it searches into websites in the query’s language and summarizes the main content for you.
Wikipedia and YouTube are two major platforms featured prominently on Google. We found that language bias also exists on these two platforms. For example, if you search on Wikipedia using English about Buddhist meditation, the English article gives you an overview of world traditions of Buddhist meditation.
But if you switch to the French article, it includes a section of “neuroscience and Buddhism” that does not exist in articles in other languages. This difference could be partly due to the influence of the French monk Matthieu Ricard, who has participated in a series of neuroscience experiments to demonstrate the benefits of meditation and attracted a lot of attention in the French-speaking community.
Language bias becomes more dramatic on YouTube because YouTube videos tend to provide highly concentrated and focused information on one narrow aspect. For example, when searched using Japanese, the top-ranked videos include Buddhist music performed by a Japanese monk, whereas the English videos teach users about the wisdom of the Buddha. The impressions that different linguistic users get from watching these videos are very different.
So users are steered to different information and only shown the prevailing views of the search language, and not getting a global picture of the query topic?
Exactly. Such filtering effect can be neutral or useful for utility-based questions, such as visa requirements. However, it can pose a significant threat to our society on sensitive and complex topics like liberalism and international trade policy. Unlike math or computer science, which have definitive answers, complex topics demand diversity and mutual debate.
We use the fable of the blind men and the elephant to describe this phenomenon, that each language community is like a blind person touching a small portion of the elephant while believing they have seen the whole.
As Google’s ranking algorithm is designed to capture the “common case” and majority’s interests while also filtered by language, English-language users typically get positive views about liberalism while Chinese-language users get criticisms of it through Google.
Over time, such technology serves as a facilitator for social division. Mutual engagement is no longer possible because each language community sees different facts.
What factors are driving this?
There are many factors that contribute to the existing pattern. First, using the language filter is an algorithmic choice. Without a good translation system, users cannot read information written in other languages, so the language filter serves a practical function. However, now with machine translation, the language filter might not be necessary.
Second, language is intrinsically tied to culture, history, and group identity, so any concept that is expressed through a certain linguistic system is inseparable from its cultural roots. On concepts surrounding Buddhism and liberalism, different language corpora do exhibit very different opinions and perspectives.
And thirdly, the extent of discrepancy across different language searches varies depending on the topic you’re searching. For example, on topics like “Jacobean matrix,” we did not observe precipitable differences across languages. “Jacobean matrix” is a relatively new and very technical term, and has a well-defined mathematical definition, so when you’re searching this term across different languages, you don’t see much difference.
Scientific, mathematical, and technical terms, especially recent technical terms, as they have well-defined definitions, tend to have consistent interpretations. However, for terms that have a longer history, like Newton’s First Law, the top-ranked websites often include a lot of historical narratives surrounding these topics
Why is language bias showing up in AI-powered searches?
As mentioned earlier, current large language models are mainly trained on English-language data and always follow the Anglo-American perspectives by default.
There are many layers of technical issues that contribute to such problems. The first layer is with the imbalanced training data. The second layer has to do with debiasing techniques, “alignment,” and human review. Biases in existing training data are unavoidable — issues with gender bias and racial bias are very common in AI models.
However, as biases embedded in complex topics like liberalism and Buddhism haven’t caught much attention within the AI community, people haven’t started testing and working on such issues. So, most large language models currently follow the dominant and most popular perspectives in their training data, which happen to be the Anglo-American views. Fortunately, these issues can be mitigated in ways similar to how gender and racial biases are addressed.
In the paper, you warn that such language bias creates “a strong, invisible cultural barrier that has important sociopolitical implications for bridging divides.” Can you explain?
The general internet user tends to attribute authority to Google and believe Google’s search results are neutral and objective, especially compared with social media platforms. They are not aware of the skewed perspectives they get from Google.
As people continue being reinforced by the dominant views in their own language community while believing they have seen the whole, this technology is not serving us well as a communication mediator.
The danger lies in the long term. On sensitive and complex topics like what constitutes a good market economy, if one side is reinforced with the idea that a free-market economy is good while the other side is constantly fed with the advantages of a protective market economy, it can be quite challenging for both sides to reach consensus.
The language filter on the internet sets a strong barrier that prevents us from mutually understanding each other while reinforcing our existing beliefs without hearing the other side.
What can users or even the tech companies do to minimize the effects of language bias in online search?
From a user perspective, you can use Google Translate and translate your search phrase into different languages and then use the translated phrases to search and translate them back to your own language. However, these actions can be so costly to the user.
From the technical aspect, there are many ways to minimize language bias if the goal is to help users to access information from different language corpora. First, adopting a recommendation system — [that would work] like Amazon’s shopping recommendation — can help users get exposed to alternative opinions out there. Right now, the“related search” in Google search is not helping because those related search suggestions are suggesting the majority’s view within the same language.
Second, the recently rolled-out Google AI Overview has the potential to overcome language barriers. As it is searching and summarizing content for users, it can identify a spectrum of viewpoints from their entire repository, regardless of language, and then summarize and translate the main points back to users, helping users to break down the language barrier.
Source Link
Support Techcratic
If you find value in Techcratic’s insights and articles, consider supporting us with Bitcoin. Your support helps me, as a solo operator, continue delivering high-quality content while managing all the technical aspects, from server maintenance to blog writing, future updates, and improvements. Support innovation! Thank you.
Bitcoin Address:bc1qlszw7elx2qahjwvaryh0tkgg8y68enw30gpvge
Please verify this address before sending funds.Bitcoin QR Code
Simply scan the QR code below to support Techcratic.
Please read the Privacy and Security Disclaimer on how Techcratic handles your support.
Disclaimer: As an Amazon Associate, Techcratic may earn from qualifying purchases.