Microsoft boosts its AI translation capabilities
Machine translation just got a significant upgrade in Microsoft Translator and Azure AI services
When you purchase through links on our site, we may earn an affiliate commission.Here’s how it works.
Microsofthas given itstranslation softwareand services a major boost by adopting a newAItechnology that significantly improves the quality of production translation models.
The software giant eventually aims to combine AI models for text, vision, audio and language through its larger XYZ-code initiative. As a component of this initiative, Z-code supports the creation of AI systems that are capable of speaking, seeing, hearing and understanding.
Microsoft has updated itsMicrosoft Translatorsoftware as well as its other Azure AI services with its new Z-code models. In order to get these models into production, the software giant is usingNvidia GPUsand Triton Inference Server to efficiently scale and deploy them.
It’s also worth noting that Microsoft Translator is the first machine translation provider to introduce Z-code Mixture of Experts models live for customers.
Z-code Mixture of Experts
Unlike previousAI models, Z-code models utilize a new architecture called Mixture of Experts (MoE) where different parts of the models can learn different tasks. As such, the models learn to translate between multiple languages simultaneously.
At the same time, newly introduced Z-code MoE models take advantage of transfer learning which enables efficient knowledge sharing across similar languages such as English and French. The models also use both parallel and monolingual data during the training process which allows for high quality machine translation beyond high-resource languages.
This cool Microsoft Teams update will make sure nothing gets lost in translation
Microsoft, OpenAI may have solved a fundamental AI bottleneck
This intriguing new platform lets you translate your words into code
In October of last year, Microsoft announced in ablog postthat Microsoft Translator is now capable of translating over 100 languages. To do this, the company used 200bn parameters supporting 100 language pairs. However, as training large models with billions of parameters is challenging, the Translator team worked together with Microsoft DeepSpeed to develop a high-performance system that it used to help train its massive scale Z-code MoE models.
Are you a pro? Subscribe to our newsletter
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
Microsoft then partnered withNvidiato optimize faster engines that can be used at runtime to deploy its new Z-code/MoE models on GPUs. For its part, Nvidia developed customCUDAkernels that leveraged the CUTLASS and FasterTransformer libraries to implement MoE layers on a singleV100 GPU.
Microsoft’s new Z-code models are nowavailable by invitationto customers using its Document Translation feature that translates entire documents or even volumes of documents in a variety of different file formats while keeping their original formatting intact.
After working with the TechRadar Pro team for the last several years, Anthony is now the security and networking editor at Tom’s Guide where he covers everything from data breaches and ransomware gangs to the best way to cover your whole home or business with Wi-Fi. When not writing, you can find him tinkering with PCs and game consoles, managing cables and upgrading his smart home.
New fanless cooling technology enhances energy efficiency for AI workloads by achieving a 90% reduction in cooling power consumption
Samsung plans record-breaking 400-layer NAND chip that could be key to breaking 200TB barrier for ultra large capacity AI hyperscaler SSDs
NYT Strands today — hints, answers and spangram for Sunday, November 10 (game #252)