The AI community paid attention on April 24thWhen DeepSeek published its V4 series on April 24, 2026, the response from developers and researchers was genuine rather than performative. Artificial Analysis declared that DeepSeek had returned to the upper tier of frontier models. This mattered because the Chinese AI startup’s previous major release, V3, had genuinely disrupted Silicon Valley’s assumptions about who could build competitive models and how much it would cost. The subsequent year of relative quiet from DeepSeek had left open the question of whether V3 was an anomaly or a preview. V4 Pro and V4 Flash are the answer to that question. They suggest the latter.
What V4 Pro and V4 Flash actually deliverThe model cards published on Hugging Face reveal a sophisticated architectural choice. V4 Pro contains 1.6 trillion total parameters but activates only 49 billion during inference, using a hybrid attention architecture combining Compressed Sparse Attention and Heavily Compressed Attention. This design dramatically reduces the computational cost of running the model while preserving access to a parameter space roughly comparable to the largest foundation models produced by American laboratories. V4 Flash, at 284 billion total parameters with 13 billion active, offers a smaller footprint for contexts where speed and cost matter more than raw capability. DataCamp’s benchmark analysis places V4 Pro at 80.6 percent on SWE-bench Verified, compared to OpenAI’s GPT-5.5 at 88.7 percent. The gap is real. But the cost difference is equally real. An organization that can run V4 Pro on its own infrastructure pays inference costs a fraction of what commercial API access to GPT-5.5 would cost for equivalent volume. For many real-world applications, this economic differential is more decisive than the benchmark gap.
The ambiguity inside the word openWhat interested me most about the week’s discourse, more than the specific benchmarks, was the recurring use of the word open without interrogation of what it actually means. In AI, open-source and open-weight are distinct concepts that are frequently conflated. Genuine open-source in the traditional software development sense would require not just the final trained model weights but also the training data, the training code, the evaluation methodology, and the complete audit trail of decisions made during development. No major AI laboratory currently meets this standard, including DeepSeek. V4 Pro and V4 Flash release their weights, meaning that anyone can download the numerical parameters that constitute the trained model and run it on their own hardware. What they do not release is the dataset used for training, the filtering criteria applied to that data, the reinforcement learning from human feedback process, or the model-specific alignment choices made during post-training. Simon Willison’s technical analysis described V4 as almost on the frontier at a fraction of the price, which is accurate, but the open framing deserves more scrutiny than it typically receives in the excitement of a new release.
The geopolitical context that technical analysis tends to omitDeepSeek’s models differ from Meta’s Llama series, Google’s Gemma, or other open-weight models from American or European organizations in one dimension that is not visible in benchmark tables: DeepSeek operates under Chinese law. China’s National Intelligence Law, enacted in 2017, obligates Chinese organizations and individuals to support, assist, and cooperate with state intelligence work. The implications of this legal environment for AI models developed by Chinese companies are contested, but they are not trivial. If a Chinese AI company is legally required to cooperate with state intelligence requests, the question of what data flows through its systems and what the state can access becomes genuinely relevant to organizations that handle sensitive information. This concern applies differently to an open-weight model than to an API-based service. When you download the weights and run the model on your own infrastructure, no data necessarily flows back to DeepSeek’s servers. The privacy concern is attenuated. But the question of whether the training data, the alignment choices, and the model’s emergent properties reflect political constraints is a legitimate question about any model developed within a specific national regulatory and political environment.
The semiconductor sanctions paradoxDeepSeek’s V4 series arrived in a context shaped by American export controls on advanced semiconductors. Since 2022, the United States has progressively restricted the export of Nvidia’s highest-performance chips to China, aiming to constrain China’s ability to train frontier AI models by denying access to the computational resources required. The architecture choices visible in V4 Pro and V4 Flash appear to be at least partly a response to this constraint. By optimizing inference efficiency through compressed attention mechanisms, DeepSeek achieves performance that approaches frontier levels while requiring less computational intensity per token generated. Technical reviewers noted that V4 Pro requires only 27 percent of single-token inference FLOPs and 10 percent of KV cache compared to the previous V3.2 generation. If this description is accurate, the export controls intended to slow Chinese AI development may have instead accelerated the development of computationally efficient architectures that will prove useful even when chip access is eventually normalized. The gap between the intended effect of a policy and its actual effect is worth tracking carefully.
What the competition between GPT-5.5 and V4 Pro means practicallyThe specific benchmark numbers matter less than the structural fact they represent: competitive AI capability is no longer exclusively produced in San Francisco or Seattle. OpenAI’s GPT-5.5 leads V4 Pro on standard benchmarks, but the gap is close enough that for a substantial portion of practical use cases, a well-configured V4 Pro deployment provides comparable value. This changes the economics of the AI industry in ways that will play out over years. OpenAI’s business model depends on the premise that its models offer capability that cannot be replicated without paying for API access. When open-weight models approach that capability level at dramatically lower marginal cost, the addressable market for commercial API pricing compresses. OpenAI’s response has been to accelerate development, lower API prices on older models, and invest in product differentiation beyond raw language model capability. These are rational competitive responses. Whether they will be sufficient to maintain OpenAI’s revenue position over a multi-year horizon as open-weight models continue improving is an open question.
What open AI could mean for access and democratizationSetting aside the geopolitical complications for a moment, the genuine promise of capable open-weight AI deserves acknowledgment. In medicine, open-weight models enable hospitals and research institutions in lower-income settings to apply AI to diagnostic support, literature synthesis, and clinical documentation without ongoing API subscription costs. In agriculture, models that can run locally on modest hardware could support crop disease identification and resource management advice for farmers without reliable internet access. In education, locally deployed language models allow the creation of personalized tutoring systems without the privacy concerns and cost barriers that commercial APIs introduce. These applications represent concrete human benefit that becomes more accessible as the cost floor of capable AI declines. The fact that this cost reduction is being driven in part by a Chinese company operating under conditions that raise legitimate governance questions does not negate the potential benefit. It complicates the picture.
Governance frameworks that have not kept paceA United Nations report published in April 2026 warned that AI governance frameworks have not kept pace with technological development and called for urgent multilateral attention to accountability and transparency standards. This observation captures something real. The current landscape contains a mix of highly capable models operating under very different governance conditions: some developed by publicly listed American corporations subject to U.S. law and investor accountability, some developed by Chinese companies under Chinese law, some developed by European research consortia under GDPR and the EU AI Act framework. Users, organizations, and governments making deployment decisions often lack the information needed to understand how these different governance environments affect the models they are choosing between. The benchmark tables are comparable. The governance contexts are not.
Japan’s specific dilemma in this landscapeFor Japan, the arrival of DeepSeek V4 creates a specific version of a dilemma that affects many countries. Japan’s technology sector is deeply integrated with both American AI ecosystems, through partnerships with Microsoft, Google, and OpenAI, and with manufacturing supply chains that flow through China. Japanese companies and research institutions that adopt open-weight AI face decisions about which models to deploy in which contexts. Using V4 Pro for non-sensitive internal productivity tasks may seem straightforward. Using it in contexts that involve sensitive business data, government collaboration, or defense-adjacent research raises different questions. American alliance pressures could eventually take the form of guidelines or regulations that restrict government contractor use of Chinese-origin AI models, as has already occurred in the semiconductor domain. Japan would need to navigate those pressures while maintaining its own policy judgment about specific risks and benefits.
The constructive scenario: competition drives quality and accessIn the most optimistic reading of the current moment, sustained competition between OpenAI, Google, Anthropic, Meta, and DeepSeek produces rapid improvement in AI capability, significant reductions in cost, and expanded access for organizations that previously could not afford frontier AI tools. This competition has already produced real benefits for users: prices for commercial API access have declined, context windows have expanded, model reliability has improved, and the range of capable open-weight models has grown substantially. If this competitive dynamic continues, the constraint on AI adoption shifts from cost and capability to organizational readiness and use-case clarity, which are more tractable problems than fundamental technology barriers.
The more cautious scenario: fragmentation along geopolitical linesThe alternative trajectory sees AI capability and deployment becoming progressively fragmented along geopolitical lines. American policy could move toward restricting government and critical infrastructure use of Chinese-origin models, as has already occurred with telecommunications hardware like Huawei equipment. Allied countries would face pressure to adopt compatible restrictions. Chinese domestic policy would correspondingly favor Chinese AI ecosystems. The result is a world where American-sphere and Chinese-sphere AI development proceed in increasingly separate directions, with different training data, different alignment approaches, different benchmark prioritization, and different regulatory environments. Interoperability between these ecosystems would decline. Global AI research collaboration, which has historically benefited from relatively open exchange of ideas across national boundaries, would become more constrained. The open-weight model that seems to democratize AI today would become a proxy in a larger geopolitical contest about whose AI runs the world’s digital infrastructure.
The question that remains openThe week DeepSeek came back, the question I could not resolve was the one that felt most important: what work does the word open actually do in this field, and who benefits from its ambiguity? Open is a promise of transparency, of community ownership, of freedom from proprietary lock-in. In the context of open-weight AI, it is also a partial promise: you get the weights, not the full development history. You get deployment freedom, not governance transparency. The geopolitical context adds another layer: open models from democratic open societies and open models from authoritarian states may use the same technical vocabulary of openness while embedding very different political realities. This distinction is not comfortable to make, because it risks being used as cover for protectionist interests dressed up as security concerns. But ignoring it because it is uncomfortable risks something else entirely. The right response is probably neither uncritical embrace nor reflexive rejection, but a more careful and contextual judgment than the word open by itself allows. Whether our AI governance frameworks are equipped to support that kind of judgment is the question that will determine a great deal about how the technology actually develops.
What the architectural innovation actually meansDeepSeek V4’s technical significance lies not simply in being cheaper to build but in demonstrating that the conventional scaling hypothesis, the idea that more parameters straightforwardly produce better models, can be challenged through architectural efficiency. The hybrid attention mechanism combining Compressed Sparse Attention and Heavily Compressed Attention achieves frontier-approaching performance while activating only a fraction of total parameters during inference. This efficiency enables the model to handle one-million-token context windows at dramatically lower computational cost than would be required by conventional transformer architectures. For practical applications, this means that complex long-document analysis tasks, which previously required expensive high-memory inference infrastructure, become feasible on more modest hardware. The architectural insight embedded in V4’s design will likely propagate through the broader research community and influence subsequent model development regardless of its specific origin.
The agentic AI context that makes this release significantDeepSeek V4 arrives as the AI field is transitioning from language models that answer questions to systems that autonomously execute extended sequences of tasks. Agentic AI, systems that plan multi-step processes, use external tools, retrieve information from the web, and coordinate multiple sub-agents to achieve complex goals, represents the next major wave of practical AI deployment. The economics of agentic AI are significantly affected by inference cost. An agentic workflow that involves hundreds of model calls to complete a task is far more expensive to operate with high-cost commercial API models than with open-weight models run on owned infrastructure. As DeepSeek V4 and its successors bring capable open-weight models to performance levels competitive with commercial frontier models, the cost threshold for deploying agentic AI systems drops substantially. This has real implications for sectors like manufacturing, logistics, healthcare, and professional services, where complex multi-step workflows could benefit from AI automation but where the cost of commercial API usage at required volumes has been prohibitive.
EU AI Act regulation and the open-source questionThe European Union’s AI Act, which entered into force in 2024, applies a risk-based regulatory framework to AI systems. General-purpose AI models above certain capability thresholds face requirements for transparency, risk assessment, and incident reporting. The Act includes limited exemptions for open-source models, reflecting a judgment that wider community access to model internals provides a form of accountability. However, these exemptions have limits, and models that pose what the regulation terms systemic risk face requirements regardless of their licensing status. The regulatory status of DeepSeek models in Europe is not fully resolved. Whether Chinese companies are prepared to comply with European transparency requirements, whether European regulatory bodies have the capacity to enforce requirements against entities outside their jurisdiction, and whether the exemptions for open-weight models adequately account for security and governance risks, these are live questions with significant commercial and policy consequences. Japan, while not subject to the EU AI Act, is developing its own AI governance approach and faces analogous questions about which regulatory frameworks apply to which models in which contexts.
Japan’s AI industry and its realistic positioningJapan’s domestic AI industry has been developing competitive foundation models through NTT’s tsuzumi, Fujitsu’s Fugaku-LLM, and Sakura Internet’s infrastructure investments. Against the competitive landscape of OpenAI, Google, and now DeepSeek, these Japanese-origin models occupy niche positions rather than frontier roles. The realistic path for Japanese AI development is likely not to compete head-to-head with American or Chinese frontier general-purpose models but to pursue vertical specialization in domains where Japanese institutional depth provides genuine advantage: manufacturing quality control where decades of Toyota Production System principles can be embedded in AI-assisted workflows, medical imaging and clinical documentation where Japan’s aging population creates concentrated demand, legal and regulatory document analysis in Japanese where language-specific models outperform general-purpose systems, and agricultural optimization for Japanese growing conditions and supply chains. This vertical specialization strategy accepts that the general-purpose model wars are being fought elsewhere and focuses Japanese AI capacity where domestic knowledge creates defensible positions.
Where the open question goes from hereSeveral weeks after DeepSeek V4’s release, the word open continues to do substantial rhetorical work that deserves more careful examination. Technical openness in the sense of model weight accessibility is real and genuinely significant. Governance transparency in the sense of full disclosure of training data, alignment choices, and development process is largely absent from all major model releases, including those from American companies. Geopolitical neutrality, in the sense of development processes that are not embedded in any specific national security or political accountability framework, is a property that no major AI model fully achieves. These three concepts are distinct and should be treated as such. A model can be technically open while being governance-opaque. A model can be commercially accessible while being geopolitically situated. Conflating these dimensions, allowing the most visible form of openness, weight accessibility, to stand in for the others, is a simplification that serves neither good policy analysis nor users’ genuine interests. The AI governance discussions that will shape how these technologies are deployed over the next decade would benefit from more precise vocabulary about what exactly is open, to whom, in what ways, and with what accountability mechanisms. That is the productive question that this week’s release should generate.
The talent behind DeepSeek’s achievementsIt is worth pausing on what DeepSeek’s consistent frontier-competitive performance actually reflects at a human level. DeepSeek is staffed by engineers and researchers who have clearly developed genuine technical insight into transformer architecture, attention mechanisms, and training efficiency. The architectural innovations in V4’s hybrid attention design are not the product of simply scaling resources; they reflect careful engineering judgment about where efficiency gains are available. Recognizing this matters for accurate analysis: the capabilities are real, and they come from genuine technical achievement. The AI research community’s ability to maintain productive international exchange of ideas has been one factor enabling this progress. The degree to which that exchange can be preserved, or will be further fragmented by geopolitical pressure, affects the pace of progress available to everyone in the field.
The practical implications for organizations making deployment decisionsOrganizations currently making decisions about which AI models to deploy for which purposes face a genuinely complex landscape. For internal productivity applications that do not involve sensitive data, capable open-weight models like DeepSeek V4 Flash offer attractive economics with limited governance concerns. For applications involving customer data, regulated industries, or government collaboration, the governance questions raised by any model’s development context, not only Chinese-origin models, deserve more careful evaluation. For applications with national security implications, the origin and governance context of AI models is clearly relevant and subject to existing regulatory frameworks. The practical advice that follows from this analysis is not to avoid or embrace any particular model categorically, but to make deployment decisions based on careful assessment of the specific application context, the specific data involved, and the specific governance questions that combination raises. This kind of contextual judgment is harder than a blanket policy but more accurate in its risk assessment.
この記事を書いた人
灰島
30代の日本人。国際情勢・地政学・経済を日常的に読み続けている。歴史の文脈から現代を読むアプローチで、世界のニュースを考察している。専門家ではないが、誠実に、感情も交えながら書く。


コメント