DeepSeek Rolls Out New Model Capable of Processing Extremely Long Texts in One Go

Deep News
02/11

On February 11th, multiple users reported that DeepSeek has updated its version on both web and app platforms, now supporting context lengths of up to 1 million tokens. This marks a significant increase from the 128K context length available in DeepSeek V3.1, which was released last August.

In practical tests, DeepSeek confirmed during interactions that it supports 1 million tokens of context, enabling it to process extremely long texts in a single session. When provided with the entire novel "Jane Eyre," a document exceeding 240,000 tokens, DeepSeek successfully recognized and processed the content.

Previously, informed sources suggested that DeepSeek was more likely to release minor updates to its V3 series models during the Spring Festival period. However, the same sources indicated that the main event is still upcoming. DeepSeek's next-generation flagship model is expected to be a foundational model with trillions of parameters. Due to this substantial scale increase, training speeds have noticeably slowed, resulting in delays to the release timeline.

免責聲明:投資有風險,本文並非投資建議,以上內容不應被視為任何金融產品的購買或出售要約、建議或邀請,作者或其他用戶的任何相關討論、評論或帖子也不應被視為此類內容。本文僅供一般參考,不考慮您的個人投資目標、財務狀況或需求。TTM對信息的準確性和完整性不承擔任何責任或保證,投資者應自行研究並在投資前尋求專業建議。

熱議股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10