List of AI News about MiniMax
| Time | Details |
|---|---|
|
2026-03-09 17:25 |
MiniMax Agent Platform Launch: Latest Analysis on agent.minimax.io and 2026 AI Agent Market Opportunities
According to @godofprompt on X, the link agent.minimax.io highlights MiniMax’s agent platform. As reported by MiniMax’s official site, the company offers conversational and multimodal large models and tool-use capabilities that enable autonomous AI agents for tasks like customer support and content operations. According to MiniMax product documentation, agent workflows integrate retrieval, function calling, and memory to support enterprise use cases such as lead qualification, knowledge base Q&A, and task automation. As reported by multiple MiniMax announcements, the platform targets developers with APIs and dashboards for building domain-specific agents, creating commercial opportunities in verticals including ecommerce chat, fintech onboarding, and marketing automation. |
|
2026-03-02 23:53 |
ARC-AGI-2 Results: Chinese Open-Weight Models Underperform Frontier LLMs — Data-Backed Analysis
According to ARC Prize on X, semi-private ARC-AGI-2 results show Kimi K2.5 scored 12% at $0.28, Minimax M2.5 5% at $0.17, GLM-5 5% at $0.27, and DeepSeek V3.2 4% at $0.12, all below July 2025 frontier lab models (as referenced by ARC Prize) (source: ARC Prize; post amplified by Ethan Mollick). According to ARC Prize, these outcomes indicate current Chinese open-weight models are strong in narrow tasks but weaker on generalization and out-of-distribution reasoning versus leading closed models, highlighting a performance gap with direct business impact on reliability-critical use cases like autonomous agents and complex tool-use pipelines. As reported by ARC Prize, the cost-performance figures suggest competitive token pricing but insufficient reasoning yield, guiding enterprises to consider hybrid stacks—using frontier closed models for hardest reasoning while deploying open-weight models for domain-specific, cost-sensitive workflows. |
|
2026-02-27 12:11 |
MiniMax M2.5 Agent Model: Latest Analysis on Code Generation, Edge-Case Handling, and Cost for Shipping AI Agents
According to @godofprompt on X, MiniMax’s M2.5 is positioned as an agent-first large model that plans architecture, writes modular code, addresses edge cases, and optimizes performance, aiming to function like a software engineer rather than a chat assistant. According to MiniMax’s platform site and docs, M2.5 is available via platform.minimax.io with text generation guides and a dedicated Coding Plan subscription, signaling a commercial focus on production-grade code agents. As reported by the MiniMax docs, the offering emphasizes multi-step planning and code reliability features that support autonomous agent workflows, creating opportunities for startups to reduce engineering cycle time and ship automation-heavy backends. According to MiniMax’s subscription page, pricing under the Coding Plan targets affordability for continuous agent runs, which can lower unit economics for code refactoring, test generation, and performance tuning use cases. |
|
2026-02-27 12:10 |
MiniMax M2.5 Beats Opus 4.6 on SWE-Bench Verified: 80.2% Score, 3x Faster, $1 Hour—AI Coding Benchmark Analysis
According to God of Prompt on X (Twitter), MiniMax M2.5 surpassed Opus 4.6 on the SWE-Bench Verified benchmark with an 80.2% score, delivers roughly 3x faster execution, and is offered at a flat $1 per hour, while using only 10B activated parameters, positioning it as the smallest Tier-1 model for coding tasks. As reported by the same source, these metrics imply lower latency and significantly reduced inference cost, enabling 24/7 autonomous coding agents and continuous integration bots at practical budgets. According to the post, the combination of high benchmark accuracy and small active parameter count suggests strong efficiency-per-dollar, which can improve ROI for software teams deploying code assistants, test repair bots, and maintenance agents in production pipelines. |
|
2026-02-23 19:41 |
Anthropic Alleges 24,000 Bot Accounts Scraped Claude: 16M Exchanges Tied to DeepSeek, Moonshot, MiniMax — 2026 Investigation Analysis
According to The Rundown AI, Anthropic claims it uncovered 24,000 fake user accounts conducting more than 16 million interactions to extract Claude model capabilities, allegedly linked to DeepSeek, Moonshot, and MiniMax (as reported by The Rundown AI citing Anthropic statements). According to The Rundown AI, Anthropic asserts that rapid advances at these Chinese labs significantly rely on capabilities extracted from U.S. models, highlighting substantial model-to-model knowledge transfer risk and potential violations of platform terms. As reported by The Rundown AI, the incident underscores urgent needs for enterprise-grade abuse detection, API rate-limiting, automated behavioral fingerprinting, and synthetic traffic monitoring to protect proprietary model IP and maintain fair competition in foundation model markets. |
