Ant Group Releases Open-Source Multimodal AI Model Ming-Flash-Omni 2.0

Stock News
02/11

Ant Group has officially open-sourced its latest multimodal large model, Ming-Flash-Omni 2.0. The model demonstrates exceptional performance across multiple public benchmarks, with particularly strong capabilities in visual language understanding, controllable speech generation, and image generation and editing. Some metrics have surpassed those of Gemini 2.5 Pro. Ming-Flash-Omni 2.0 is also the industry's first model capable of unified audio generation across all scenarios, enabling the synchronized synthesis of speech, ambient sound effects, and background music within a single audio track. Users can finely adjust parameters such as timbre, speech rate, tone, volume, emotion, and even dialect through natural language instructions. In terms of inference efficiency, the model achieves an extremely low inference frame rate of 3.1Hz, allowing real-time generation of high-fidelity, minute-long audio while significantly optimizing computational costs and response speed. Ant Group has been investing in multimodal research for several years, with the Ming-Omni series now in its third iteration. The open-sourcing of Ming-Flash-Omni 2.0 releases its core capabilities as a reusable foundation, providing a unified entry point for end-to-end multimodal application development. Users can also experience and access the model online through Ant's official platform, Ling Studio.

免责声明:投资有风险,本文并非投资建议,以上内容不应被视为任何金融产品的购买或出售要约、建议或邀请,作者或其他用户的任何相关讨论、评论或帖子也不应被视为此类内容。本文仅供一般参考,不考虑您的个人投资目标、财务状况或需求。TTM对信息的准确性和完整性不承担任何责任或保证,投资者应自行研究并在投资前寻求专业建议。

热议股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10