Microsoft has mostly relied on OpenAI’s large language models to power its own AI products like Copilot. But now, the tech giant’s AI division, led by Mustafa Suleyman, has announced two new AI models – MAI-1-preview and MAI-Voice-1.
In a blog post, Microsoft said that while MAI-1-preview “offers a glimpse of future offerings inside Copilot”, MAI-Voice-1 can generate a 60-second-long audio clip in just one second using a single GPU, making it one of the most efficient speech systems available to date.
Talking of MAI-Voice-1, the tech giant says that the “highly expressive and natural speech generation model” is already powering certain features like Copilot Daily and Podcasts. The company is also planning to bring it to Copilot Labs, where users will be able to try it out with storytelling demos.
As for MAI-1-preview, the new large language model is currently available for testing on LMArena and was trained using approximately 15,000 NVIDIA H100 GPUs. Microsoft says that the AI model specialises “in following instructions and providing helpful responses to everyday queries.” In the following weeks, the AI model will be available in certain text-based use cases. The number of GPUs used to train Microsoft’s newest model is far less than the likes of xAI’s Grok, which took more than 1,00,000 of these chips just for training.
In a statement to The Verge last year, Microsoft’s AI division head, Mustafa Suleyman, had said that the company’s in-house developed AI models won’t be focusing on enterprise use cases, but instead they will “create something that works extremely well for the consumer and really optimize for our use case.”
Microsoft may have invested billions of dollars in ChatGPT maker OpenAI, but in the last few months, the relationship between the two companies has been rocky. Earlier this year, in June, OpenAI executives allegedly considered accusing Microsoft of anticompetitive behaviour in the partnership.
© IE Online Media Services Pvt Ltd
Average Rating