On September 1st morning, according to the official "LongCat" account, Meituan officially released and open-sourced LongCat-Flash-Chat today. According to the introduction, LongCat-Flash adopts an innovative Mixture-of-Experts (MoE) architecture with total parameters of 560B and activated parameters ranging from 18.6B to 31.3B (average 27B), achieving dual optimization of computational efficiency and performance. Based on comprehensive evaluation across multiple benchmark tests, LongCat-Flash-Chat demonstrates outstanding advantages in agent tasks while activating only a small number of parameters, with inference speed exceeding 100 tokens per second.