Cryptopolitan
2026-01-10 10:00:07

DeepSeek V4 rumored to outperform ChatGPT and Claude in long-context coding

DeepSeek V4 is rumored to outperform ChatGPT and Claude in long-context coding, targeting elite-level coding tasks. Insiders claim that Silicon Valley’s AI landscape should be concerned if internal tests hint at its expected performance after the rollout in mid-February. Chinese-based AI start-up DeepSeek is reportedly planning to release DeepSeek V4, its latest large language model, on February 17. People familiar with the matter claim that the model is poised to cast a shadow over existing large language models, such as OpenAI’s ChatGPT and Anthropic’s Claude, when handling long-context code prompts and tasks. Developers express deep anticipation for the DeepSeek V4 release The Chinese company has not publicly disclosed any information about the imminent release or confirmed the rumors as of the time of writing. Developers across different social networks have expressed deep anticipation for the release. Yuchen Jin, an AI developer and co-founder of Hyperbolic Labs, wrote on X that “DeepSeek V4 is rumored to drop soon, with stronger coding than Claude and GPT.” Subreddit r/DeepSeek also heated up, with one user explaining that their obsession with DeepSeek’s imminent V4 model was not normal. The user said that they frequently “check news, possible rumors, and I even go to read the Docs on the DS website to look for any changes or signs that indicate an update.” DeepSeek’s previous releases have had a significant impact on global markets. The Chinese AI start-up released its R1 reasoning model in January 2025, leading to a trillion-dollar sell-off. The release matched OpenAI’s 01 model on math and reasoning benchmarks, despite costing significantly less than the US AI startup spent on its 01 model. The Chinese company reportedly spent only $6 million on the model release. Meanwhile, global competitors spend nearly 70 times more for the same output. Its V3 model also logged a 90.2% score on the MATH-500 benchmark, compared to Claude’s 78.3%. DeepSeek’s more recent V3 upgrade (V3.2 Speciale) further improved its productivity. Its V4 model’s selling point has evolved from the V3’s emphasis on pure reasoning, formal proofs, and logical math. The new release is expected to be a hybrid model that combines both reasoning and non-reasoning tasks. The model aims to capture the developer market by filling an existing gap that demands high accuracy and long-context code generation. Claude Opus 4.5 currently claims dominance in the SWE benchmark, achieving an accuracy of 80.9%. The V4 needs to beat this to overturn Claude Opus 4.5. Based on previous successes, the incoming model may surpass this threshold and claim dominance in the benchmark. DeepSeek pioneers mHC for training LLMs DeepSeek’s success has left many in profound professional disbelief. How could such a small company achieve such milestones? The secret could be deeply entrenched in its research paper published on January 1. The company identified a new training method that allows developers to easily scale large language models. Liang Wenfeng, founder and CEO of DeepSeek, wrote in the research that the company is using Manifold-Constrained Hyper-Connections (mHC) to train its AI models. The executive proposed using mHC to address the issues encountered when developers train large language models. According to Wenfeng, mHC is an upgrade of Hyper-Connections (HC), a framework that other AI developers use to train their large language models. He explained that HC and other traditional AI architectures force all data through a single, narrow channel. At the same time, mHC widens that pathway into multiple channels, facilitating the transfer of data and information without causing training collapse. Lian Jye Su, chief analyst at Omdia, commended CEO Wenfeng for publishing their research. Su emphasized that DeepSeek’s decision to publish its training methods dictates renewed confidence in the Chinese AI sector. DeepSeek has dominated the developing world. Microsoft published a report on Thursday, showing that DeepSeek commands 89% of China’s AI market and has been gaining momentum in developing countries. Join a premium crypto trading community free for 30 days - normally $100/mo.

Crypto 뉴스 레터 받기
면책 조항 읽기 : 본 웹 사이트, 하이퍼 링크 사이트, 관련 응용 프로그램, 포럼, 블로그, 소셜 미디어 계정 및 기타 플랫폼 (이하 "사이트")에 제공된 모든 콘텐츠는 제 3 자 출처에서 구입 한 일반적인 정보 용입니다. 우리는 정확성과 업데이트 성을 포함하여 우리의 콘텐츠와 관련하여 어떠한 종류의 보증도하지 않습니다. 우리가 제공하는 컨텐츠의 어떤 부분도 금융 조언, 법률 자문 또는 기타 용도에 대한 귀하의 특정 신뢰를위한 다른 형태의 조언을 구성하지 않습니다. 당사 콘텐츠의 사용 또는 의존은 전적으로 귀하의 책임과 재량에 달려 있습니다. 당신은 그들에게 의존하기 전에 우리 자신의 연구를 수행하고, 검토하고, 분석하고, 검증해야합니다. 거래는 큰 손실로 이어질 수있는 매우 위험한 활동이므로 결정을 내리기 전에 재무 고문에게 문의하십시오. 본 사이트의 어떠한 콘텐츠도 모집 또는 제공을 목적으로하지 않습니다.