Google expands AI translation with Gemini, but language gaps persist

[Chinese characters engraved on the wall. Photo Credit to Unsplash]
On December 12, 2025, Google integrated its Gemini model into Google Translate, enabling real-time speech-to-speech translation in over 70 languages, including Mandarin Chinese, marking one of the most significant advancements in AI-assisted language access in the platform's history.
The update introduced live headphone translation in beta across Android devices in the United States, Mexico, and India, with Google confirming plans to extend the feature to iOS and additional countries throughout 2026.
Chinese was also added to Gemini's Google Workspace side panel in February 2025, allowing users to generate content, summarize documents, and complete workplace tasks in Chinese for the first time.
Collectively, these advancements mark a measurable increase in Chinese language accessibility through mainstream AI platforms.
The reach, however, remains uneven.
For Chinese users, the issue extends beyond access to underlying data imbalance. While tools now support Chinese input and output, the systems powering them are still largely trained on English-dominant datasets, which shapes both translation quality and informational depth.
Research from Stanford University published in May 2025 found that most major large language models are predominantly trained on English data, with non-English languages underrepresented in both quantity and quality of training material.
Researchers cautioned that the gap is likely to grow, noting that workers and communities fluent in English will continue to gain advantages as AI transforms global workplaces, while those relying on other languages face compounding technological barriers.
A November 2025 study published in Nature confirmed that Common Crawl—a large-scale open dataset of web pages widely used to train AI models—remains heavily skewed toward English-language content, reinforcing a structural bias that persists across most leading models.
Accessibility and accuracy are also separate problems that need to be addressed.
A study published in Humanities and Social Sciences Communications in November 2025 analyzed over 42,000 viewer comments on AI-translated Chinese-language television dramas and found that AI-generated subtitles frequently mishandled idiomatic expressions, culturally embedded metaphors, and emotionally charged dialogue.
Roughly 35 percent of the viewer feedback highlighted emotional resonance as a priority, a quality AI translation consistently failed to preserve without human editorial intervention.
Industry data indicates AI translation tools misinterpret culturally specific phrases approximately 40 percent of the time, compared to an error rate below 5 percent for professional human translators handling equivalent content.
The language a user queries in also shapes the substance of information returned, not only its fluency.
Johns Hopkins University researchers, presenting their findings at the 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics, found that the same border dispute query produced India-aligned responses in Hindi and China-aligned responses in Chinese, across multiple major AI models.
A study published in PNAS Nexus in February 2026 compared AI models developed in China against those developed outside, finding substantially higher refusal rates—the frequency with which a model declines to answer a question—and inaccurate responses across 145 political questions among Chinese-originating models. This gap, researchers attributed to government regulation rather than technological limitations alone.
A separate bilingual benchmark study from the same month tested 17 large language models on politically sensitive queries and found that 15 out of 17 produced measurably different responses depending on whether the question was asked in Chinese or English.
AI’s expanded accessibility to the Chinese language comes with challenges.
However, Google’s Gemini-powered expansion does not guarantee consistency in meaning, context, or perspective across languages.
As the platform continues its global expansion,, the challenge is no longer whether users can access information in their own language, but whether that information reflects the same level of accuracy and neutrality promised at launch.
- Haryn Lee / Grade 11
- Liberta-Scholars College Prep