DeepSeek V4 Preview Released: Lower Memory and Compute Requirements, Leading Agent Capabilities

DeepSeek V4's preview version launches with reduced memory and compute needs, claiming top agent capabilities and performance surpassing Sonnet 4.5.

DeepSeek V4 Preview Released

On April 24, Chinese AI company DeepSeek announced the release of the preview version of its new model series, DeepSeek-V4, which is now open source. The model claims to lead in three dimensions: agent capabilities, world knowledge, and reasoning performance within the domestic and open-source fields.

DeepSeek-V4 is available in two versions, Pro and Flash, both supporting a super long context of 1 million tokens and significantly reducing computational and memory requirements.

Image 11

API services are also launched, allowing developers to call the model by changing the parameters to deepseek-v4-pro or deepseek-v4-flash. The interface is compatible with OpenAI ChatCompletions and Anthropic standards.

DeepSeek also disclosed that due to limitations in high-end computing supply, the Pro version currently has very limited service throughput. It is expected that prices for the Pro version will significantly decrease later this year with the mass release of Huawei’s Ascend 950 super nodes.

Notably, Ascend CANN will live stream the debut of DeepSeek V4 on the Ascend platform at 4 PM.

Image 12

This release coincides almost exactly with OpenAI’s launch of GPT-5.5 the previous day, with starkly contrasting pricing strategies. Some netizens pointed out:

GPT-5.5 launched yesterday at a price of $30 per million output tokens, while DeepSeek V4 was released today under the MIT license, marking a significant drop in AI intelligence costs that forces every AI product company to reassess its profit structure.

Image 13

User Enrico commented that DeepSeek V4 is “really impressive, fast, and intelligent,” although he believes the output price of $3.48 per million tokens is “not cheap,” but noted that LocalAI will help promote the model to a broader user base.

Image 14

DeepSeek-V4-Pro: Performance Comparable to Top Closed-Source Models

DeepSeek-V4-Pro is the flagship version of this release, positioned to match the performance of top closed-source models.

Image 15

In terms of reasoning performance, V4-Pro claims to surpass all currently published open-source models in mathematical, STEM, and competitive coding assessments, achieving results comparable to the world’s top closed-source models.

Image 16

In world knowledge assessments, V4-Pro significantly outperforms other open-source models, only slightly trailing behind Google’s Gemini-Pro-3.1.

Agent capabilities have greatly improved. Compared to previous models, DeepSeek-V4-Pro’s agent capabilities have been notably enhanced. In Agentic Coding assessments, V4-Pro has reached the best level among current open-source models and performed excellently in other agent-related evaluations.

Currently, DeepSeek-V4 has become the Agentic Coding model used by the company’s internal employees, with feedback indicating that the user experience surpasses that of Sonnet 4.5, and the delivery quality is close to Opus 4.6 in non-thinking mode, though there remains a gap compared to Opus 4.6 in thinking mode.

DeepSeek-V4’s detailed technical report was also released alongside its launch.

Image 17

DeepSeek-V4-Flash: A Faster and More Efficient Economic Choice

V4-Flash is positioned as a faster and more economical lightweight option.

Compared to DeepSeek-V4-Pro, DeepSeek-V4-Flash is slightly inferior in world knowledge reserves but demonstrates nearly comparable reasoning abilities.

Due to smaller model parameters and activation scales, its API service offers significant advantages in speed and cost.

In agent assessments, V4-Flash performs comparably to V4-Pro on simple tasks, but still lags behind on more complex tasks.

This positioning makes V4-Flash more suitable for enterprise applications that are sensitive to latency and cost, with moderate task complexity.

Structural Innovation and High Context Efficiency

DeepSeek-V4 introduces a novel attention mechanism at the underlying architecture level.

By compressing on the token dimension and combining it with self-developed DSA sparse attention technology (DeepSeek Sparse Attention), the company claims to have achieved globally leading long-context capabilities while significantly reducing the demand for computational resources and memory compared to traditional methods.

A direct result of this architectural innovation is that a 1M context window will become the standard for all official DeepSeek services.

Image 18

For enterprise users needing to handle long documents, extended dialogues, or complex multi-step tasks, this capability’s availability is of substantial significance.

By reducing computational consumption while expanding the context window, it also helps further lower reasoning costs, reinforcing DeepSeek’s competitive advantage in terms of cost performance.

Agent Ecosystem Adaptation Progressing Simultaneously

DeepSeek stated that the V4 series has been specially adapted and optimized for mainstream agent products such as Claude Code, OpenClaw, OpenCode, and CodeBuddy, resulting in performance improvements in coding tasks and document generation tasks.

Image 19

At the API level, both models support a maximum context length of 1M and are compatible with both non-thinking and thinking modes.

Thinking mode allows setting reasoning intensity through the reasoning_effort parameter, with options for high or max levels. DeepSeek recommends enabling thinking mode and setting the intensity to max for complex agent scenarios.

Image 20

Was this helpful?

Likes and saves are stored in your browser on this device only (local storage) and are not uploaded to our servers.

Comments

Discussion is powered by Giscus (GitHub Discussions). Add repo, repoID, category, and categoryID under [params.comments.giscus] in hugo.toml using the values from the Giscus setup tool.