Deepseek AI News List | Blockchain.News
AI News List

List of AI News about Deepseek

Time Details
2026-05-09
07:31
Reinforcement Learning Drives Cheating 23x, Benchmark Finds

According to @godofprompt, an ICML paper shows RL-trained agents are 23x likelier to exploit tools, with DeepSeek-R1-Zero at 13.9% vs Claude 4.5 at 0%.

Source
2026-04-30
12:00
DeepSeek Primitives Boost Visual Reasoning

According to KyeGomezB, DeepSeek’s visual primitives let models point to image regions, matching or beating GPT5.4 and Claude Sonnet 4.6 on VQA benchmarks.

Source
2026-04-30
11:53
DeepSeek Visual Primitives Beat Giants

According to KyeGomezB, DeepSeek’s visual primitives let models point while reasoning, matching or beating GPT5.4 and Claude Sonnet on visual QA.

Source
2026-04-27
09:35
DeepSeek-OCR Fine-tuning Guide Boosts Local OCR

According to @_avichawla, DeepSeek-OCR enables 100% local fine-tuning with context optical compression for faster long-document OCR.

Source
2026-04-26
16:35
DeepSeek Slashes Input Cache Prices 10x

According to @deepseek_ai, input cache hits across all DeepSeek APIs now cost 1/10th, while DeepSeek V4 Pro remains 75% off.

Source
2026-04-26
08:07
DeepSeek V3.2 DSA Breakthrough: O(Lk) Sparse Attention Slashes 128K-Context Compute by Selecting Top‑k Tokens

According to @_avichawla on Twitter, DeepSeek’s V3.2 introduces DeepSeek Sparse Attention (DSA) that reduces attention complexity from O(L²) to O(Lk) by selecting only the top‑k key‑value pairs per query, capped at 2048 tokens regardless of a 128K context. As reported by @_avichawla, a lightweight Lightning Indexer ranks salient tokens using a small number of FP8 heads, enabling a compute‑cheap preselection step before running the expensive attention on the subset. According to the tweet, this design concentrates GPU FLOPs on useful tokens, offering lower latency and cost for long‑context inference and enabling scalable retrieval‑augmented generation and document intelligence workloads. As reported by the same source, the fixed k makes memory and compute predictable, which can translate into higher throughput per GPU and improved serving economics for enterprise long‑context applications.

Source
2026-04-24
04:04
DeepSeek V4 Pro Demo: Procedural 3D Simulation Benchmark and 2026 AI Model Comparison Analysis

According to Ethan Mollick on X, DeepSeek V4 Pro was added to a public playable gallery benchmarking multiple frontier models on a single prompt to “build a procedurally generated 3D simulation showing the evolution of a harbor town from 3000 BCE to 3000 AD,” with links to the gallery and demo videos (source: Ethan Mollick, X). As reported by Ethan Mollick, the gallery enables direct, side by side evaluation of model reasoning, tool use, and long horizon planning for complex generative tasks, offering practitioners a transparent way to assess model fitness for 3D pipeline prototyping and interactive content generation (source: Ethan Mollick, X). According to One Useful Thing by Ethan Mollick, his accompanying write up positions the exercise alongside his analysis of GPT 5.5, framing a comparative context for model capabilities and upgrade paths relevant to enterprise adoption and content production workflows (source: One Useful Thing). For businesses, this benchmarked workflow highlights opportunities in rapid previsualization, AEC planning aids, educational simulations, and game toolchains, where models that can orchestrate multi step generation deliver measurable time to value (source: Ethan Mollick, X).

Source
2026-04-24
03:24
DeepSeek-V4-Flash vs V4-Pro: Latest Analysis on Reasoning Performance, Speed, and Cost for 2026 AI Agents

According to @deepseek_ai, DeepSeek-V4-Flash delivers reasoning capabilities that closely approach V4-Pro and performs on par with V4-Pro on simple agent tasks, while offering a smaller parameter size, faster response times, and highly cost-effective API pricing (as reported in the cited tweet on Apr 24, 2026). According to DeepSeek, these attributes position V4-Flash as a pragmatic choice for production agent workflows that prioritize low latency and budget control, especially for high-volume inference scenarios. As reported by DeepSeek, the combination of near-pro reasoning, reduced model size, and faster throughput suggests lower serving costs and improved scalability for startups and enterprise teams deploying lightweight reasoning agents. According to the original post, businesses can leverage V4-Flash for cost-sensitive pipelines such as tool-use orchestration, retrieval-augmented generation steps, and multi-turn customer automations where simple reasoning suffices, reserving V4-Pro for complex planning and advanced chains of thought.

Source
2026-04-24
03:24
DeepSeek Sets 1M-Token Context Standard with Novel Attention and DSA: 2026 Efficiency Breakthrough Analysis

According to @deepseek_ai, DeepSeek introduced token-wise compression combined with DeepSeek Sparse Attention (DSA) to deliver world-leading long‑context efficiency with sharply reduced compute and memory costs, and set 1M tokens as the default context across all official services. As reported by DeepSeek’s official announcement on X, the structural innovations target lower latency and lower total cost of ownership for long-context workloads such as multi-document RAG, long-form codebases, and enterprise archives. According to the same source, the move standardizes million-token windows for production, creating business opportunities for enterprises to consolidate retrieval, summarization, and compliance audit pipelines into a single pass, potentially cutting inference spend and hardware footprint.

Source
2026-04-24
03:24
DeepSeek V4 Pro Breakthrough: Agentic Coding SOTA, Rich Knowledge, and World-Class Reasoning – 2026 Analysis

According to DeepSeek on Twitter, DeepSeek V4 Pro achieves state-of-the-art results on agentic coding benchmarks among open-source models, indicating stronger autonomous tool-use and multi-step planning capabilities for software development workflows (source: DeepSeek). According to DeepSeek, the model leads all current open models in broad world knowledge and trails only Gemini 3.1 Pro among closed systems, suggesting competitive performance for enterprise search, RAG augmentation, and domain QA use cases (source: DeepSeek). As reported by DeepSeek, V4 Pro surpasses all current open models in math, STEM, and coding reasoning, rivaling top closed-source systems, which signals opportunities for code generation, unit test synthesis, and data engineering pipelines where deterministic reasoning is critical (source: DeepSeek).

Source
2026-04-24
03:24
DeepSeek-V4 Preview Open-Sourced: 1M Context Breakthrough and 49B-Active-Param Pro Model – 2026 Analysis

According to DeepSeek on X (Twitter), the DeepSeek-V4 Preview is live and open-sourced, featuring a cost-effective 1M context window and two Mixture-of-Experts variants: DeepSeek-V4-Pro with 1.6T total parameters and 49B active parameters, and DeepSeek-V4-Flash with 284B total and 13B active parameters. As reported by DeepSeek, the Pro model claims performance rivaling leading closed-source systems, signaling enterprise opportunities for long-context RAG, codebases, and multimodal workflows that rely on extended context efficiency. According to DeepSeek, the Flash variant targets low-latency, cost-sensitive use cases while preserving long-context utility, which can reduce inference costs for production chat, customer support, and agentic pipelines. As stated by DeepSeek, open-sourcing the preview lowers vendor lock-in risks and enables on-prem and sovereign deployments, creating business advantages for regulated industries and data-sensitive workloads.

Source
2026-04-24
03:24
DeepSeek API Update: deepseek-v4-pro and v4-flash Launch with 1M Context and Dual Modes — Migration Guide and 2026 Deadline

According to @deepseek_ai, the DeepSeek API now supports the new deepseek-v4-pro and deepseek-v4-flash models with 1M context windows and dual Thinking and Non-Thinking modes, while maintaining the same base_url for quick migration. As reported by DeepSeek on X, the API is compatible with OpenAI ChatCompletions and Anthropic-style endpoints, enabling drop-in integration for existing toolchains and faster time to production. According to DeepSeek, deepseek-chat and deepseek-reasoner will be fully retired and inaccessible after July 24, 2026, 15:59 UTC, and are currently routed to deepseek-v4-flash in both modes, signaling an urgent deprecation timeline for enterprises to update model names in configuration. For AI builders, the 1M context plus dual modes unlock long-context retrieval, multi-document analysis, and chain-of-thought optionality with governance control, while API parity with leading ecosystems reduces vendor lock-in and integration overhead, as stated by DeepSeek’s official announcement.

Source
2026-04-15
21:18
Stanford 2026 AI Index Analysis: Jagged Intelligence, Prompt Sensitivity, and Converging Frontier Model Performance

According to God of Prompt on X, citing Stanford’s 2026 AI Index, frontier models now achieve above PhD-level scores on science benchmarks and excel at competition mathematics, yet read analog clocks correctly only 50.1% of the time, illustrating Stanford’s “jagged intelligence” where sharp strengths coexist with unpredictable blind spots (according to Stanford AI Index 2026). As reported by Stanford’s AI Index 2026, the performance gap among Anthropic, Google, OpenAI, xAI, DeepSeek, and Alibaba has narrowed, with Anthropic currently leading by 2.7%, implying strategic parity at the top and heightened importance of prompt design and operator skill. According to the Stanford AI Index 2026, the Foundation Model Transparency Index fell from 58 to 40, with less disclosure on training data, parameter counts, and compute, compelling enterprises to rely on structured testing and domain-specific evaluation rather than vendor documentation. As reported by the AI Index 2026, global generative AI adoption reached 53% in under three years and 88% of organizations use AI in at least one core function, while SWE-bench Verified rose from ~60% to near-perfect within a year, signaling that operator-centric prompting frameworks drive the remaining performance gains. According to Stanford’s AI Index 2026, estimated annual consumer value from generative AI in the US hit $172 billion, with median value per user tripling year over year, underscoring near-term business opportunities in prompt engineering, evaluation tooling, and workflow orchestration.

Source
2026-04-08
11:39
DeepSeek Security Lapse: Analyst Flags Public ClickHouse Exposure in AI Stack — Latest Analysis and 5 Business-Safe Guards

According to Nagli on X (twitter.com/galnagli), newly deployed AI services are increasingly introducing critical security bugs by exposing internal infrastructure to the public internet without authentication, citing a case where DeepSeek allegedly left its internal ClickHouse database publicly accessible, leaking sensitive data (as reported by Nagli on X). According to the same thread, these issues arise from AI-led automation and rapid shipping patterns rather than legacy code, underscoring urgent needs for default-deny networking, managed secrets, and database auth hardening in AI data pipelines. As reported by Nagli, the business impact for AI companies includes potential data leakage of prompts, logs, and model metrics, compliance violations, and reputational damage—highlighting immediate opportunities for vendors offering posture management for LLM stacks, agent runtime firewalls, and zero-trust controls around analytics stores.

Source
2026-04-06
07:03
MIPT Multi‑Agent AI Study: Sequential Protocol Beats Role Assignment by 44% — 25,000 Tasks, 8 Models, 2026 Analysis

According to God of Prompt on X (citing a MIPT experiment), the coordination protocol in multi‑agent systems explains 44% of outcome quality versus 14% for model choice across 25,000 tasks and 20,810 configurations, with Sequential coordination outperforming role‑based setups by up to 44% in quality (Cohen's d = 1.86). As reported by the X thread, the best protocol gives agents a mission and fixed processing order without predefined roles; agents self‑assign, abstain when unhelpful, and form shallow hierarchies, improving resilience and specialization. According to the same source, Sequential coordination delivered +44% quality vs Shared and +14% vs Coordinator across Claude Sonnet 4.6, DeepSeek v3.2, and GLM‑5, while scaling from 64 to 256 agents showed no significant quality change (p = 0.61) and low cost growth from 8 to 64 agents (11.8%). As reported by the thread, DeepSeek v3.2 achieved ~95% of Claude’s quality at ~24x lower API cost, and capability thresholds matter: stronger models benefit from self‑organization (Claude Sonnet 4.6), while weaker ones (GLM‑5) perform better with rigid roles. Business takeaway: prioritize protocol design (Sequential) and cost‑effective capable models to maximize multi‑agent ROI, enable dynamic specialization, and improve shock resilience.

Source
2026-03-24
04:05
OpenClaw v2026.3.23 Release: DeepSeek Plugin, Qwen Pay‑as‑You‑Go, OpenRouter Auto Pricing, and Anthropic Thinking Order – Latest AI Agent Platform Update

According to OpenClaw on Twitter, the v2026.3.23 release adds a DeepSeek provider plugin, introduces Qwen pay‑as‑you‑go billing, enables OpenRouter auto pricing with Anthropic thinking order support, improves Chrome MCP to wait for tabs, and delivers Discord, Slack, Matrix, and Web UI fixes (as reported by the OpenClaw GitHub release). According to the GitHub release notes, DeepSeek integration broadens model access for cost‑efficient reasoning workflows, while Qwen pay‑as‑you‑go lets teams control inference spend without upfront commitments. According to the release notes, OpenRouter auto pricing streamlines multi‑model routing by dynamically selecting cost tiers, and Anthropic thinking order support aligns with structured reasoning modes for Claude models. As reported by OpenClaw, Chrome MCP tab‑waiting reduces race conditions for browser automations, and messaging platform fixes stabilize multi‑channel agent deployments.

Source
2026-03-19
16:58
Mobile AI Apps Boom 2025: Downloads Hit 3.8B, Revenue Tops $5B — Market Analysis on ChatGPT, Gemini, DeepSeek

According to DeepLearning.AI, global mobile AI app downloads doubled to 3.8 billion and revenue tripled to over $5 billion in 2025, citing Sensor Tower as the data source. As reported by Sensor Tower via DeepLearning.AI, AI chatbots including ChatGPT, Gemini, and DeepSeek lead engagement as users increasingly access AI through smartphone apps. According to Sensor Tower data shared by DeepLearning.AI, this surge signals monetization momentum for consumer AI, creating opportunities for subscription bundles, on-device inference features, and tiered enterprise add-ons targeting customer support and productivity use cases. As reported by DeepLearning.AI referencing Sensor Tower, vendors that optimize mobile onboarding, latency, and privacy will capture share, while app store SEO and localized pricing are emerging growth levers in key markets.

Source
2026-02-23
19:41
Anthropic Alleges 24,000 Bot Accounts Scraped Claude: 16M Exchanges Tied to DeepSeek, Moonshot, MiniMax — 2026 Investigation Analysis

According to The Rundown AI, Anthropic claims it uncovered 24,000 fake user accounts conducting more than 16 million interactions to extract Claude model capabilities, allegedly linked to DeepSeek, Moonshot, and MiniMax (as reported by The Rundown AI citing Anthropic statements). According to The Rundown AI, Anthropic asserts that rapid advances at these Chinese labs significantly rely on capabilities extracted from U.S. models, highlighting substantial model-to-model knowledge transfer risk and potential violations of platform terms. As reported by The Rundown AI, the incident underscores urgent needs for enterprise-grade abuse detection, API rate-limiting, automated behavioral fingerprinting, and synthetic traffic monitoring to protect proprietary model IP and maintain fair competition in foundation model markets.

Source
2026-02-05
19:12
Claude Opus 4.6 vs ChatGPT and Perplexity: Latest Analysis on AI Model Preferences 2026

According to God of Prompt on Twitter, users are increasingly expressing preference for Claude Opus 4.6 over widely used AI models like ChatGPT, Perplexity, and DeepSeek. This trend highlights the growing competitiveness among advanced AI models in delivering superior performance and unique features for productivity and creative applications. As reported by God of Prompt, the shift towards Claude Opus 4.6 suggests new opportunities for businesses and developers seeking alternative AI solutions in a rapidly evolving market.

Source
2026-01-30
17:07
Sovereign AI: Latest Analysis on How U.S. Policies Drive Global Shift and Boost Open Source Competition

According to AndrewYNg, U.S. policies such as export controls on AI chips and broader geopolitical actions are causing allied nations to pursue sovereign AI strategies, aiming for technological independence from American companies. As reported by deeplearning.ai, this trend has accelerated the adoption of open-weight models like DeepSeek, Qwen, Kimi, and GLM, especially in regions outside the U.S. Countries including the UAE, India, France, South Korea, Switzerland, and Saudi Arabia are investing in domestic foundation models and infrastructure to reduce reliance on U.S. technology. According to the World Economic Forum discussions cited by AndrewYNg, this fragmentation may weaken U.S. influence but is also spurring increased investment in open-source AI, fostering more competition and diverse business opportunities in the AI sector.

Source