Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing

    ‘Comment sections are not customers’: American Eagle brings back Sydney Sweeney amid celebrity push

    Media Briefing: Publishers explore selling AI visibility know-how to brands

    Facebook X (Twitter) Instagram
    • Artificial Intelligence
    • Business Technology
    • Cryptocurrency
    • Gadgets
    • Gaming
    • Health
    • Software and Apps
    • Technology
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Tech AI Verse
    • Home
    • Artificial Intelligence

      Read the extended transcript: President Donald Trump interviewed by ‘NBC Nightly News’ anchor Tom Llamas

      February 6, 2026

      Stocks and bitcoin sink as investors dump software company shares

      February 4, 2026

      AI, crypto and Trump super PACs stash millions to spend on the midterms

      February 2, 2026

      To avoid accusations of AI cheating, college students are turning to AI

      January 29, 2026

      ChatGPT can embrace authoritarian ideas after just one prompt, researchers say

      January 24, 2026
    • Business

      The HDD brand that brought you the 1.8-inch, 2.5-inch, and 3.5-inch hard drives is now back with a $19 pocket-sized personal cloud for your smartphones

      February 12, 2026

      New VoidLink malware framework targets Linux cloud servers

      January 14, 2026

      Nvidia Rubin’s rack-scale encryption signals a turning point for enterprise AI security

      January 13, 2026

      How KPMG is redefining the future of SAP consulting on a global scale

      January 10, 2026

      Top 10 cloud computing stories of 2025

      December 22, 2025
    • Crypto

      Is Bitcoin Price Entering a New Bear Market? Here’s Why Metrics Say Yes

      February 19, 2026

      Cardano’s Trading Activity Crashes to a 6-Month Low — Can ADA Still Attempt a Reversal?

      February 19, 2026

      Is Extreme Fear a Buy Signal? New Data Questions the Conventional Wisdom

      February 19, 2026

      Coinbase and Ledn Strengthen Crypto Lending Push Despite Market Slump

      February 19, 2026

      Bitcoin Caught Between Hawkish Fed and Dovish Warsh

      February 19, 2026
    • Technology

      Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing

      February 19, 2026

      ‘Comment sections are not customers’: American Eagle brings back Sydney Sweeney amid celebrity push

      February 19, 2026

      Media Briefing: Publishers explore selling AI visibility know-how to brands

      February 19, 2026

      How the MLS plans to convert World Cup interest into lasting soccer fandom

      February 19, 2026

      Philips Hue releases new upgraded Turaco outdoor lights

      February 19, 2026
    • Others
      • Gadgets
      • Gaming
      • Health
      • Software and Apps
    Check BMI
    Tech AI Verse
    You are at:Home»Technology»Xiaomi unveils open-source AI reasoning model MiMo
    Technology

    Xiaomi unveils open-source AI reasoning model MiMo

    TechAiVerseBy TechAiVerseApril 30, 2025No Comments5 Mins Read2 Views
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Xiaomi unveils open-source AI reasoning model MiMo
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp Email

    Xiaomi unveils open-source AI reasoning model MiMo


    ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

    Unlocking the Reasoning Potential of Language Model
    From Pretraining to Posttraining

    ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━


    This code repository is licensed under the Apache2.0 License.

    I. Introduction

    Currently, most successful RL works, including open-source research, rely on relatively large base models, e.g., 32B models, particularly for enhancing code reasoning capabilities. Moreover, it was widely considered that achieving uniform and simultaneous improvements in both mathematical and code capabilities within a small model is challenging. Nonetheless, we believe that the effectiveness of the RL trained reasoning model relies on the inherent reasoning potential of the base model. To fully unlock the reasoning potential of language models, efforts must focus not only on post-training but also on pre-training strategies tailored to reasoning.

    In this work, we present MiMo-7B, a series of models trained from scratch and born for reasoning tasks. Our RL experiments from MiMo-7B-Base show that our model possesses extraordinary reasoning potential, even surpassing much larger 32B models. Additionally, we perform RL training on a cold-started SFT model, resulting in MiMo-7B-RL, which demonstrates superior performance on both mathematics and code reasoning tasks, matching the performance of OpenAI o1-mini.

    We open-source MiMo-7B series, including checkpoints of the base model, SFT model, RL model trained from base model, and RL model trained from the SFT model.
    We believe this report along with the models will provides valuable insights to develop powerful reasoning LLM that benefit the larger community.

    🌟 Highlights

    • Pre-Training: Base Model Born for Reasoning

      • We optimize data preprocessing pipeline, enhancing text extraction toolkits and applying multi-dimensional data filtering to increase reasoning pattern density in pre-training data. We also employ multiple strategies to generate massive diverse synthetic reasoning data.
      • We adopt a three-stage data mixture strategy for pre-training. Overall, MiMo-7B-Base is pre-trained on approximately 25 trillion tokens.
      • We incorporate Multiple-Token Prediction as an additional training objective, which enhances model performance and accelerates inference.
    • Post-Training Recipe: Pioneering Reasoning Model

      • We curate 130K mathematics and code problems as RL training data, which can be verified by rule-based verifiers. Each problem undergoes careful cleaning and difficulty assessment to ensure quality. We employ only rule-based accuracy rewards to avoid potential reward hacking.
      • To mitigate the sparse reward issue for challenging code problems, we introduce a test difficulty driven code reward. By assigning fine-grained scores for test cases with varying difficulty levels, the policy can be more effectively optimized via dense reward signal.
      • We implement a data re-sampling strategy for easy problems to enhance rollout sampling efficiency and stabilize policy updates, particularly in the later phases of RL training.
    • RL Infrastructures

      • We develop a Seamless Rollout Engine to accelerate RL training and validation. Our design integrates continuous rollout, asynchronous reward computation, and early termination to minimize GPU idle time, achieving 2.29$times$ faster training and 1.96$times$ faster validation.
      • We support MTP in vLLM and enhance the robustness of the inference engine in RL system.

    II. Model Details

    Models are avaliable at https://huggingface.co/XiaomiMiMo

    Model Description Download
    MiMo-7B-Base Base model with extraordinary reasoning potential 🤗 XiaomiMiMo/MiMo-7B-Base
    MiMo-7B-RL-Zero RL model trained from base model 🤗 XiaomiMiMo/MiMo-7B-RL-Zero
    MiMo-7B-SFT SFT model trained from base model 🤗 XiaomiMiMo/MiMo-7B-SFT
    MiMo-7B-RL RL model trained from SFT model, superior performance matching OpenAI o1-mini 🤗 XiaomiMiMo/MiMo-7B-RL

    III. Evaluation Results

    Benchmark GPT-4o-0513 Claude-3.5-Sonnet-1022 OpenAI o1-mini QwQ-32B-Preview R1-Distill-Qwen-14B R1-Distill-Qwen-7B MiMo-7B-RL
    General
    GPQA Diamond
    (Pass@1)
    49.9 65.0 60.0 54.5 59.1 49.1 54.4
    SuperGPQA
    (Pass@1)
    42.4 48.2 45.2 43.6 40.6 28.9 40.5
    DROP
    (3-shot F1)
    83.7 88.3 83.9 71.2 85.5 77.0 78.7
    MMLU-Pro
    (EM)
    72.6 78.0 80.3 52.0 68.8 53.5 58.6
    IF-Eval
    (Prompt Strict)
    84.3 86.5 84.8 40.4 78.3 60.5 61.0
    Mathematics
    MATH-500
    (Pass@1)
    74.6 78.3 90.0 90.6 93.9 92.8 95.8
    AIME 2024
    (Pass@1)
    9.3 16.0 63.6 50.0 69.7 55.5 68.2
    AIME 2025
    (Pass@1)
    11.6 7.4 50.7 32.4 48.2 38.8 55.4
    Code
    LiveCodeBench v5
    (Pass@1)
    32.9 38.9 53.8 41.9 53.1 37.6 57.8
    LiveCodeBench v6
    (Pass@1)
    30.9 37.2 46.8 39.1 31.9 23.9 49.3

    MiMo-7B series

    Benchmark MiMo-7B-Base MiMo-7B-RL-Zero MiMo-7B-SFT MiMo-7B-RL
    Mathematics
    MATH500
    (Pass@1)
    37.4 93.6 93.0 95.8
    AIME 2024
    (Pass@1)
    32.9 56.4 58.7 68.2
    AIME 2025
    (Pass@1)
    24.3 46.3 44.3 55.4
    Code
    LiveCodeBench v5
    (Pass@1)
    32.9 49.1 52.3 57.8
    LiveCodeBench v6
    (Pass@1)
    29.1 42.9 45.5 49.3

    Important

    The evaluation are conducted with temperature=0.6.

    AIME24 and AIME25 are with averaged score of 32 repetitions. LiveCodeBench v5 (20240801-20250201), LiveCodeBench v6 (20250201-20250501), GPQA-Diamond and IF-Eval are with averaged score of 8 repetitions. MATH500 and SuperGPQA are with a single run.

    IV. Deployment

    vLLM inference

    1. [Recommended] We official support inference with MiMo-MTP using our fork of vLLM.

    Example script

    from vllm import LLM, SamplingParams
    
    model_path = "/path/to/MiMo"
    llm = LLM(
        model=model_path,
        trust_remote_code=True,
        num_speculative_tokens=1,
        disable_log_stats=False
    )
    sampling_params = SamplingParams(temperature=0.6)
    
    conversation = [
        {
            "role": "system",
            "content": ""
        },
        {
            "role": "user",
            "content": "Write an essay about the importance of higher education.",
        },
    ]
    
    outputs = llm.chat(conversation,
                       sampling_params=sampling_params,
                       use_tqdm=False)
    
    for output in outputs:
        prompt = output.prompt
        generated_text = output.outputs[0].text
        print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
    
    print("=" * 80)
    1. Or, you can register a vLLM loader for MiMo without loading MTP parameters.

    You can copy the registry/register_mimo_in_vllm.py to your directory and import it with

    import register_mimo_in_vllm
    
    from vllm import LLM, SamplingParams
    
    model_path = "/path/to/MiMo"
    llm = LLM(
        model=model_path,
        trust_remote_code=True,
        # num_speculative_tokens=1,
        disable_log_stats=False
    )
    sampling_params = SamplingParams(temperature=0.6)

    HuggingFace inference

    Example script

    from transformers import AutoModel, AutoModelForCausalLM, AutoTokenizer
    
    model_path = "/path/to/MiMo"
    model = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True)
    tokenizer = AutoTokenizer.from_pretrained(model_path)
    inputs = tokenizer(["Today is"], return_tensors='pt')
    output = model.generate(**inputs, max_new_tokens = 100)
    print(tokenizer.decode(output.tolist()[0]))

    Recommended environment and prompts

    • We recommend using our fork of vLLM which is developed based on vLLM 0.7.3.
    • We recommend using empty system prompt.

    We haven’t verified MiMo with other inference engines and welcome contributions based on the model definition in the Huggingface repo 💻.

    V. Citation

    @misc{xiaomi2025mimo,
          title={MiMo: Unlocking the Reasoning Potential of Language Model – From Pretraining to Posttraining}, 
          author={{Xiaomi LLM-Core Team}},
          year={2025},
          primaryClass={cs.CL},
          url={https://github.com/XiaomiMiMo/MiMo}, 
    }

    VI. Contact

    Please contact us at mimo@xiaomi.com or open an issue if you have any questions.

    Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
    Previous ArticleThe Leaderboard Illusion
    Next Article Finland Bans Smartphones in Schools
    TechAiVerse
    • Website

    Jonathan is a tech enthusiast and the mind behind Tech AI Verse. With a passion for artificial intelligence, consumer tech, and emerging innovations, he deliver clear, insightful content to keep readers informed. From cutting-edge gadgets to AI advancements and cryptocurrency trends, Jonathan breaks down complex topics to make technology accessible to all.

    Related Posts

    Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing

    February 19, 2026

    ‘Comment sections are not customers’: American Eagle brings back Sydney Sweeney amid celebrity push

    February 19, 2026

    Media Briefing: Publishers explore selling AI visibility know-how to brands

    February 19, 2026
    Leave A Reply Cancel Reply

    Top Posts

    Ping, You’ve Got Whale: AI detection system alerts ships of whales in their path

    April 22, 2025684 Views

    Lumo vs. Duck AI: Which AI is Better for Your Privacy?

    July 31, 2025273 Views

    6.7 Cummins Lifter Failure: What Years Are Affected (And Possible Fixes)

    April 14, 2025156 Views

    6 Best MagSafe Phone Grips (2025), Tested and Reviewed

    April 6, 2025118 Views
    Don't Miss
    Technology February 19, 2026

    Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing

    Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing…

    ‘Comment sections are not customers’: American Eagle brings back Sydney Sweeney amid celebrity push

    Media Briefing: Publishers explore selling AI visibility know-how to brands

    How the MLS plans to convert World Cup interest into lasting soccer fandom

    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us
    About Us

    Welcome to Tech AI Verse, your go-to destination for everything technology! We bring you the latest news, trends, and insights from the ever-evolving world of tech. Our coverage spans across global technology industry updates, artificial intelligence advancements, machine learning ethics, and automation innovations. Stay connected with us as we explore the limitless possibilities of technology!

    Facebook X (Twitter) Pinterest YouTube WhatsApp
    Our Picks

    Omnicom’s lack of surprises in its 2025 earnings is both a good and bad thing

    February 19, 20260 Views

    ‘Comment sections are not customers’: American Eagle brings back Sydney Sweeney amid celebrity push

    February 19, 20260 Views

    Media Briefing: Publishers explore selling AI visibility know-how to brands

    February 19, 20260 Views
    Most Popular

    7 Best Kids Bikes (2025): Mountain, Balance, Pedal, Coaster

    March 13, 20250 Views

    VTOMAN FlashSpeed 1500: Plenty Of Power For All Your Gear

    March 13, 20250 Views

    This new Roomba finally solves the big problem I have with robot vacuums

    March 13, 20250 Views
    © 2026 TechAiVerse. Designed by Divya Tech.
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions

    Type above and press Enter to search. Press Esc to cancel.