DeepSeek Rolls Out New Model Capable of Processing Extremely Long Texts in One Go

Deep News
02/11

On February 11th, multiple users reported that DeepSeek has updated its version on both web and app platforms, now supporting context lengths of up to 1 million tokens. This marks a significant increase from the 128K context length available in DeepSeek V3.1, which was released last August.

In practical tests, DeepSeek confirmed during interactions that it supports 1 million tokens of context, enabling it to process extremely long texts in a single session. When provided with the entire novel "Jane Eyre," a document exceeding 240,000 tokens, DeepSeek successfully recognized and processed the content.

Previously, informed sources suggested that DeepSeek was more likely to release minor updates to its V3 series models during the Spring Festival period. However, the same sources indicated that the main event is still upcoming. DeepSeek's next-generation flagship model is expected to be a foundational model with trillions of parameters. Due to this substantial scale increase, training speeds have noticeably slowed, resulting in delays to the release timeline.

免责声明:投资有风险,本文并非投资建议,以上内容不应被视为任何金融产品的购买或出售要约、建议或邀请,作者或其他用户的任何相关讨论、评论或帖子也不应被视为此类内容。本文仅供一般参考,不考虑您的个人投资目标、财务状况或需求。TTM对信息的准确性和完整性不承担任何责任或保证,投资者应自行研究并在投资前寻求专业建议。

热议股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10