Close Menu
    Facebook X (Twitter) Instagram
    Articles Stock
    • Home
    • Technology
    • AI
    • Pages
      • About ArticlesStock — AI & Technology Journalist
      • Contact us
      • Disclaimer For Articles Stock
      • Privacy Policy
      • Terms and Conditions
    Facebook X (Twitter) Instagram
    Articles Stock
    AI

    Xiaomi Releases MiMo-V2.5-Professional and MiMo-V2.5: Matching Frontier Mannequin Benchmarks at Considerably Decrease Token Value

    Naveed AhmadBy Naveed Ahmad23/04/2026Updated:23/04/2026No Comments8 Mins Read
    blog 1 14


    Xiaomi MiMo staff publicly launched two new fashions: MiMo-V2.5-Professional and MiMo-V2.5. The benchmarks, mixed with some genuinely placing real-world job demos, make a compelling case that open agentic AI is catching as much as the frontier quicker than most anticipated. Each fashions can be found instantly through API, and priced competitively.

    What’s an Agentic Mannequin, and Why Does It Matter?

    Most LLM benchmarks check a mannequin’s capability to reply a single, self-contained query. Agentic benchmarks check one thing a lot tougher — whether or not a mannequin can full a multi-step aim autonomously, utilizing instruments (internet search, code execution, file I/O, API calls) over many turns, with out shedding monitor of the unique goal.

    Consider it because the distinction between a mannequin that may reply “how do I write a lexer?” versus one that may truly write an entire compiler, run checks in opposition to it, catch regressions, and repair them — all with out a human within the loop. The latter is strictly what Xiaomi MiMo staff is demonstrating right here.

    MiMo-V2.5-Professional: The Flagship

    MiMo-V2.5-Professional is Xiaomi’s most succesful mannequin to this point, delivering vital enhancements over its predecessor, MiMo-V2-Professional, on the whole agentic capabilities, advanced software program engineering, and long-horizon duties.

    The important thing benchmark numbers are aggressive with high closed-source fashions: SWE-bench Professional 57.2, Claw-Eval 63.8, and τ3-Bench 72.9 — putting it alongside Claude Opus 4.6 and GPT-5.4 throughout most evaluations. V2.5-Professional can maintain advanced, long-horizon duties spanning greater than a thousand instrument calls, demonstrating substantial enhancements in instruction following inside agentic situations, reliably adhering to refined necessities embedded in context and sustaining sturdy coherence throughout ultra-long contexts.

    One behavioral property that distinguishes V2.5-Professional from earlier fashions is what Xiaomi MiMo staff calls “harness consciousness”: it makes full use of the affordances of its harness surroundings, manages its reminiscence, and shapes how its personal context is populated towards the ultimate goal. This implies the mannequin doesn’t simply execute directions mechanically. It actively optimizes its personal working surroundings to remain on monitor throughout very lengthy duties.

    The three real-world job demos Xiaomi revealed illustrate precisely what “long-horizon agentic functionality” means in follow.

    Demo 1 — SysY Compiler in Rust: Referred from Peking College’s Compiler Principles course venture, this job asks the mannequin to implement an entire SysY compiler in Rust from scratch: lexer, parser, AST, Koopa IR codegen, RISC-V meeting backend, and efficiency optimization. The reference venture sometimes takes a PKU CS main pupil a number of weeks. MiMo-V2.5-Professional completed in 4.3 hours throughout 672 instrument calls, scoring an ideal 233/233 in opposition to the course’s hidden check suite.

    What’s notable isn’t simply the ultimate rating — it’s the structure of execution. Moderately than thrashing via trial and error, the mannequin constructed the compiler layer by layer: scaffold the complete pipeline first, good Koopa IR (110/110), then the RISC-V backend (103/103), then efficiency (20/20). The primary compile alone handed 137/233 checks, a 59% chilly begin that implies the structure was designed appropriately earlier than a single check was run. When a refactoring step later brought on regressions, the mannequin recognized the failures, recovered, and pushed on. That is structured, self-correcting engineering habits — not pattern-matched code technology.

    Demo 2 — Full-Featured Desktop Video Editor: With only a few easy prompts, MiMo-V2.5-Professional delivered a working desktop app: multi-track timeline, clip trimming, cross-fades, audio mixing, and export pipeline. The ultimate construct is 8,192 strains of code, produced over 1,868 instrument calls throughout 11.5 hours of autonomous work.

    Demo 3 — Analog EDA- FVF-LDO Design: That is essentially the most technically specialised demo: a graduate-level analog-circuit EDA job requiring the design and optimization of an entire FVF-LDO (Flipped-Voltage-Follower low-dropout regulator) from scratch within the TSMC 180nm CMOS course of. The mannequin needed to dimension the ability transistor, tune the compensation community, and decide bias voltages in order that six metrics land inside spec concurrently — section margin, line regulation, load regulation, quiescent present, PSRR, and transient response. Wired into an ngspice simulation loop, in about an hour of closed-loop iteration — calling the simulator, studying waveforms, tweaking parameters — the mannequin produced a design the place each goal metric is met, with 4 key metrics improved by an order of magnitude over its personal preliminary try.

    Token Effectivity: Intelligence at frontier degree is just helpful if it’s cost-effective. On ClawEval, V2.5-Professional lands at 64% Move^3 utilizing solely ~70K tokens per trajectory — roughly 40–60% fewer tokens than Claude Opus 4.6, Gemini 3.1 Professional, and GPT-5.4 at comparable functionality ranges. For engineers constructing manufacturing agent pipelines, this can be a materials value discount, not only a advertising and marketing stat.

    https://mimo.xiaomi.com/mimo-v2-5-pro/

    MiMo Coding Bench is Xiaomi’s in-house analysis suite designed to evaluate fashions on real-world developer duties inside agentic frameworks like Claude Code. It covers repo understanding, venture constructing, code assessment, structured artifact technology, planning, SWE, and extra. V2.5-Professional leads the sphere on this benchmark, and Xiaomi explicitly positions it as a drop-in backend for scaffolds together with Claude Code, OpenCode, and Kilo.

    MiMo-V2.5: Native Omnimodal at Half the Value

    Whereas V2.5-Professional targets the toughest long-horizon agentic duties, MiMo-V2.5 is a serious step ahead in agentic functionality and multimodal understanding. With native visible and audio understanding, MiMo-V2.5 causes seamlessly throughout modalities, surpasses MiMo-V2-Professional in agentic efficiency, and helps as much as 1 million tokens of context.

    The mannequin is designed with notion and motion unified from scratch. MiMo-V2.5 is educated from the begin to see, hear, and act on what it perceives, resulting in a single mannequin that understands every part and will get issues performed. That is architecturally vital — earlier multimodal fashions typically bolted imaginative and prescient on high of a textual content spine, creating functionality gaps on the perception-action boundary.

    On the coding facet, the worth proposition is obvious: in MiMo Coding Bench, MiMo-V2.5 delivers sturdy outcomes on on a regular basis coding duties, closing the hole with frontier fashions and matching MiMo-V2.5-Professional at half the fee. For groups that don’t want the acute long-horizon depth of V2.5-Professional, this can be a compelling working level.

    https://mimo.xiaomi.com/mimo-v2-5/

    On multimodal benchmarks: MiMo-V2.5 achieves a 62.3 on the Claw-Eval common subset, putting it on the Pareto frontier of efficiency and effectivity. On the multimodal agentic subset, MiMo-V2.5 reaches 23.8 on Claw-Eval Multimodal, matching Claude Sonnet 4.6, main MiMo-V2-Omni by eight factors, and trailing Claude Opus 4.6 by a single level.

    On video understanding, MiMo-V2.5 scores 87.7 on Video-MME, successfully tied with Gemini 3 Professional (88.4) and effectively forward of Gemini 3 Flash. Lengthy-horizon video comprehension — scene monitoring, temporal reasoning, visible grounding over minutes of footage — is now in frontier territory. On picture understanding, MiMo-V2.5 lands at 81.0 on CharXiv RQ and 77.9 on MMMU-Professional, closing in on Gemini 3 Professional.

    Pricing is easy: MiMo-V2.5 runs at 1x (1 token = 1 credit score), whereas MiMo-V2.5-Professional runs at 2x (1 token = 2 credit). Token Plans now not cost a multiplier for the 1M-token context window — beforehand a typical value friction for long-context agentic workloads.

    Key Takeaways

    • MiMo-V2.5-Professional matches frontier closed-source fashions on key agentic benchmarks (SWE-bench Professional 57.2, Claw-Eval 63.8, τ3-Bench 72.9), whereas utilizing 40–60% fewer tokens per trajectory than Claude Opus 4.6, Gemini 3.1 Professional, and GPT-5.4.
    • Lengthy-horizon autonomy is actual and measurable — V2.5-Professional autonomously constructed an entire SysY compiler in Rust (233/233 checks, 672 instrument calls, 4.3 hours) and a full-featured desktop video editor (8,192 strains of code, 1,868 instrument calls, 11.5 hours).
    • MiMo-V2.5 is natively omnimodal — educated from scratch to see, hear, and act throughout modalities with a local 1M-token context window, matching Claude Sonnet 4.6 on Claw-Eval Multimodal and almost tying Gemini 3 Professional on Video-MME (87.7 vs. 88.4).
    • Professional-level coding efficiency at half the fee — on MiMo Coding Bench, MiMo-V2.5 matches MiMo-V2.5-Professional on on a regular basis coding duties at 1x token pricing, making it the sensible selection for many manufacturing agent pipelines.
    • Each fashions are already suitable with well-liked agentic scaffolds like Claude Code, OpenCode, and Kilo — giving AI devs a drop-in, auditable, self-hostable path to frontier-level agentic AI.

    Take a look at the Technical details MiMo-V2.5, and Technical details MiMo-V2.5-Pro. Additionally, be at liberty to observe us on Twitter and don’t neglect to hitch our 130k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

    Have to companion with us for selling your GitHub Repo OR Hugging Face Web page OR Product Launch OR Webinar and so on.? Connect with us




    Source link

    Naveed Ahmad

    Naveed Ahmad is a technology journalist and AI writer at ArticlesStock, covering artificial intelligence, machine learning, and emerging tech policy. Read his latest articles.

    Related Posts

    Elon Musk admits tens of millions of Tesla homeowners want upgrades for true ‘Full Self-Driving’

    23/04/2026

    Shade lands $14M to let inventive groups search their video libraries in plain English

    23/04/2026

    India’s app market is booming — however international platforms are capturing many of the beneficial properties

    23/04/2026
    Leave A Reply Cancel Reply

    Categories
    • AI
    Recent Comments
      Facebook X (Twitter) Instagram Pinterest
      © 2026 ThemeSphere. Designed by ThemeSphere.

      Type above and press Enter to search. Press Esc to cancel.