Google Launches TranslateGemma: Open AI Translation Models for 55 Languages | Quick Digest
Google has launched TranslateGemma, a new suite of open translation models built on Gemma 3, supporting 55 languages including Hindi. These efficient models come in three sizes for various devices, enhancing high-quality translation for developers and users globally.
Google introduced TranslateGemma, a new suite of open translation models.
Models are built on Gemma 3 and are available in 4B, 12B, and 27B parameter sizes.
Supports 55 languages, including major ones like Hindi, Spanish, and French.
The 12B model demonstrates superior efficiency, outperforming larger baselines in quality.
TranslateGemma retains multimodal capabilities, allowing translation of text within images.
Models are accessible to developers on platforms such as Kaggle and Hugging Face.
Google has officially unveiled TranslateGemma, a new collection of open translation models built on its advanced Gemma 3 architecture, marking a significant stride in democratizing high-quality language translation. These models are available in three distinct parameter sizes – 4B, 12B, and 27B – to cater to diverse deployment environments, ranging from mobile and edge devices to consumer laptops and robust cloud infrastructure. A key highlight is the 12B model's impressive efficiency, as it reportedly outperforms the larger Gemma 3 27B baseline in translation quality while requiring less than half the computing power. This breakthrough enables developers to achieve high-fidelity translation with greater throughput and lower latency.
TranslateGemma rigorously supports 55 language pairs, encompassing widely spoken languages like Spanish, French, Chinese, and critically for India, Hindi. Beyond these core languages, Google has also trained the models on nearly 500 additional language pairs, laying a robust foundation for further research and adaptation. The models leverage a specialized two-stage fine-tuning process, distilling the 'intuition' from Google's proprietary Gemini models into an open architecture, combining supervised fine-tuning with reinforcement learning. Furthermore, TranslateGemma inherits the multimodal capabilities of Gemma 3, allowing it to translate text embedded within images even without specific training for this function. These open-weight models are readily available for developers on platforms like Kaggle and Hugging Face, fostering innovation and enabling the creation of custom, privacy-focused translation solutions.
Read the full story on Quick Digest