Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Microsoft sets 2026 deadline for Secure Boot certificate expiration

    Sony confirms new WH-1000XM6 release in official teaser

    Awkward debut: XPeng’s Iron robot falls on stage

    Facebook X (Twitter) Instagram
    • Artificial Intelligence
    • Business Technology
    • Cryptocurrency
    • Gadgets
    • Gaming
    • Health
    • Software and Apps
    • Technology
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Tech AI Verse
    • Home
    • Artificial Intelligence

      Read the extended transcript: President Donald Trump interviewed by ‘NBC Nightly News’ anchor Tom Llamas

      February 6, 2026

      Stocks and bitcoin sink as investors dump software company shares

      February 4, 2026

      AI, crypto and Trump super PACs stash millions to spend on the midterms

      February 2, 2026

      To avoid accusations of AI cheating, college students are turning to AI

      January 29, 2026

      ChatGPT can embrace authoritarian ideas after just one prompt, researchers say

      January 24, 2026
    • Business

      New VoidLink malware framework targets Linux cloud servers

      January 14, 2026

      Nvidia Rubin’s rack-scale encryption signals a turning point for enterprise AI security

      January 13, 2026

      How KPMG is redefining the future of SAP consulting on a global scale

      January 10, 2026

      Top 10 cloud computing stories of 2025

      December 22, 2025

      Saudia Arabia’s STC commits to five-year network upgrade programme with Ericsson

      December 18, 2025
    • Crypto

      HBAR Shorts Face $5 Million Risk if Price Breaks Key Level

      February 10, 2026

      Ethereum Holds $2,000 Support — Accumulation Keeps Recovery Hopes Alive

      February 10, 2026

      Miami Mansion Listed for 700 BTC as California Billionaire Tax Sparks Relocations

      February 10, 2026

      Solana Drops to 2-Year Lows — History Suggests a Bounce Toward $100 is Incoming

      February 10, 2026

      Bitget Cuts Stock Perps Fees to Zero for Makers Ahead of Earnings Season, Expanding Access Across Markets

      February 10, 2026
    • Technology

      Microsoft sets 2026 deadline for Secure Boot certificate expiration

      February 11, 2026

      Sony confirms new WH-1000XM6 release in official teaser

      February 11, 2026

      Awkward debut: XPeng’s Iron robot falls on stage

      February 11, 2026

      Limited edition Analogue 3D now available to buy

      February 11, 2026

      Unusual mid-range smartphone features dot matrix secondary display, camera shutter button and 6,500 mAh battery

      February 11, 2026
    • Others
      • Gadgets
      • Gaming
      • Health
      • Software and Apps
    Check BMI
    Tech AI Verse
    You are at:Home»Technology»Deep Cogito goes big, releasing 4 new open source hybrid reasoning models with self-improving ‘intuition’
    Technology

    Deep Cogito goes big, releasing 4 new open source hybrid reasoning models with self-improving ‘intuition’

    TechAiVerseBy TechAiVerseAugust 1, 2025No Comments8 Mins Read8 Views
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp Email

    Deep Cogito goes big, releasing 4 new open source hybrid reasoning models with self-improving ‘intuition’

    Deep Cogito, a lesser-known AI research startup based in San Francisco founded by ex-Googlers, has released four new open-ish large language models (LLMs) that attempt something few others do: Learning how to reason more effectively over time — and get better at it on their own.

    The models, released as part of Cogito’s v2 family, range from 70 billion to 671 billion parameters and are available for AI developers and enterprises to use under a mix of limited and fully open licensing terms. They include:

    Dense and MoE models are each suited to different needs. Dense 70B and 405B variant models activate all parameters on every forward pass, making them more predictable and easier to deploy across a wide range of hardware.

    They’re ideal for low-latency applications, fine-tuning and environments with limited GPU capacity. MoE models, such as the 109B and 671B versions, use a sparse routing mechanism to activate only a few specialized “expert” subnetworks at a time, allowing for much larger total model sizes without proportional increases in compute cost.


    The AI Impact Series Returns to San Francisco – August 5

    The next phase of AI is here – are you ready? Join leaders from Block, GSK, and SAP for an exclusive look at how autonomous agents are reshaping enterprise workflows – from real-time decision-making to end-to-end automation.

    Secure your spot now – space is limited: https://bit.ly/3GuuPLF


    This makes them well-suited for high-performance inference tasks, research into complex reasoning or serving frontier-level accuracy at lower runtime expense. In Cogito v2, the 671B MoE model serves as the flagship, leveraging its scale and routing efficiency to match or exceed leading open models on benchmarks — while using significantly shorter reasoning chains.

    The models are available now on Hugging Face for download and usage by enterprises and on Unsloth for local usage, or, for those who can’t host the model inferences on their own hardware, through application programming interfaces (APIs) from Together AI, Baseten and RunPod.

    There’s also a quantized “8-bit floating point (FP8)” version of the 671B model, which reduces the size of the numbers used to represent the model’s parameters from 16-bits to 8-bits, helping users run massive models faster, cheaper and on more accessible hardware — sometimes with only a negligible hit to performance (95 to 99%). However, this can slightly degrade model accuracy, especially for tasks requiring fine-grained precision (some math or reasoning problems).

    All four Cogito v2 models are designed as hybrid reasoning systems: They can respond immediately to a query, or, when needed, reflect internally before answering.

    Crucially, that reflection is not just runtime behavior — it’s baked into the training process itself.

    These models are trained to internalize their own reasoning. That means the very paths they take to arrive at answers — the mental steps, so to speak — are distilled back into the models’ weights.

    Over time, they learn which lines of thinking actually matter and which don’t.

    As Deep Cogito’s blog post notes, the researchers “disincentivize the model from ‘meandering more’ to be able to arrive at the answer, and instead develop a stronger intuition for the right search trajectory for the reasoning process.”

    The result, Deep Cogito claims, is faster, more efficient reasoning and a general improvement in performance, even in so-called “standard” mode.

    Self-improving AI

    While many in the AI community are just encountering the company, Deep Cogito has been quietly building for over a year.

    It emerged from stealth in April 2025 with a series of open-source models trained on Meta’s Llama 3.2. Those early releases showed promising results.

    As VentureBeat previously reported, the smallest Cogito v1 models (3B and 8B) outperformed Llama 3 counterparts across several benchmarks — sometimes by wide margins.

    Deep Cogito CEO and co-founder Drishan Arora — previously a lead LLM engineer at Google — described the company’s long-term goal as building models that can reason and improve with each iteration, much like how AlphaGo refined its strategy through self-play.

    Deep Cogito’s core method, iterated distillation and amplification (IDA), replaces hand-written prompts or static teachers with the model’s own evolving insights.

    What is ‘machine intuition’?

    With Cogito v2, the team took that loop to a much larger scale. The central idea is simple: Reasoning shouldn’t just be an inference-time tool; it should be part of the model’s core intelligence.

    So, the company implemented a system where the model runs reasoning chains during training, and then is trained on its intermediate thoughts.

    This process yields concrete improvements, according to internal benchmarks. The flagship 671B MoE model outperforms DeepSeek R1 in reasoning tasks, matching or beating its latest 0528 model while using 60% shorter reasoning chains.

    On MMLU, GSM8K and MGSM, Cogito 671B MoE’s performance was roughly on par with top open models like Qwen1.5-72B and DeepSeek v3, and approached the performance tier of closed models like Claude 4 Opus and o3.

    Specifically:

    • Cogito 671B MoE (reasoning mode) matched DeepSeek R1 0528 across multilingual QA and general knowledge tasks, and outperformed it on strategy and logical deduction.
    • In non-reasoning mode, it exceeded DeepSeek v3 0324, suggesting that the distilled intuition carried real performance weight even without an extended reasoning path.
    • The model’s ability to complete reasoning in fewer steps also had downstream effects: Lower inference costs and faster response times on complex prompts.

    Arora explains this as a difference between searching for a path versus already knowing roughly where the destination lies.

    “Since the Cogito models develop a better intuition of the trajectory to take while searching at inference time, they have 60% shorter reasoning chains than Deepseek R1,” he wrote in a thread on X.

    What kinds of tasks do Deep Cogito’s new models excel at when using their machine intuition?

    Some of the most compelling examples from Cogito v2’s internal testing highlight exactly how this manifests in use.

    In one math-heavy prompt, a user asks whether a train traveling at 80 mph can reach a city 240 miles away in under 2.5 hours.

    While many models simulate the calculation step-by-step and occasionally make unit conversion errors, Cogito 671B reflects internally, determines that 240 ÷ 80 = 3 hours, and correctly concludes that the train cannot arrive in time. It does so with only a short internal reasoning trace — under 100 tokens — compared to the 200-plus used by DeepSeek R1 to reach the same answer.

    In another example involving legal reasoning, a user asks whether a specific U.S. Supreme Court ruling would apply to a hypothetical case involving search and seizure. Cogito’s reasoning mode highlights a two-step logic: Dirst determining whether the hypothetical matches the precedent, then explaining why it does or doesn’t. The model reaches a nuanced answer with clear justification — a kind of interpretive reasoning that many LLMs still struggle with.

    Other tasks show improvements in handling ambiguity. On a classic multi-hop question — “If Alice is Bob’s mother, and Bob is Charlie’s father, what is Alice to Charlie?” — models often get tangled in pronouns. Cogito v2’s models correctly identify Alice as Charlie’s grandmother, even in slightly reworded variants where other open models falter.

    Efficiency at scale

    Despite the massive size of the new models, Deep Cogito claims to have trained all eight of its Cogito models — including smaller v1 checkpoints — for under $3.5 million in total, compared to the reported $100 million plus for some of OpenAI’s leading models.

    That includes data generation, synthetic reinforcement, infrastructure and more than 1,000 training experiments. Compared to the nine-figure budgets of other frontier models, it’s a fraction of the typical spend.

    Arora attributes this frugality to the company’s core thesis: Smarter models need better priors, not more tokens.

    By teaching the model to skip redundant or misleading reasoning paths, Cogito v2 delivers stronger performance without ballooning inference time.

    That’s a meaningful tradeoff for users running models on API infrastructure or edge devices where latency and cost matter.

    What’s next for Deep Cogito and v2?

    The release of Cogito v2 is not a final product, but an iterative step. Arora describes the company’s roadmap as “hill climbing” — running models, learning from their reasoning traces, distilling them and repeating the loop. Over time, each model becomes a stepping stone for the next.

    Every model Deep Cogito has released is open source, and the company says that will remain true for future iterations.

    Already, its work has attracted attention and support from backers like Benchmark’s Eric Vishria and South Park Commons’ Aditya Agarwal.

    Infrastructure partners include Hugging Face, Together AI, RunPod, Baseten, Meta’s Llama team and Unsloth.

    For developers, researchers, and enterprise teams, the models are available now. Developers can run them locally, compare modes or fine-tune for specific use cases.

    And, for the broader open-source AI community, Cogito v2 offers more than just a new benchmark winner — it proposes a different way to build intelligence. Not by thinking harder, but by learning how to think better.

    Daily insights on business use cases with VB Daily

    If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

    Read our Privacy Policy

    Thanks for subscribing. Check out more VB newsletters here.

    An error occured.

    Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
    Previous ArticleHard-won vibe coding insights: Mailchimp’s 40% speed gain came with governance price
    Next Article OpenAI removes ChatGPT feature after private conversations leak to Google search
    TechAiVerse
    • Website

    Jonathan is a tech enthusiast and the mind behind Tech AI Verse. With a passion for artificial intelligence, consumer tech, and emerging innovations, he deliver clear, insightful content to keep readers informed. From cutting-edge gadgets to AI advancements and cryptocurrency trends, Jonathan breaks down complex topics to make technology accessible to all.

    Related Posts

    Microsoft sets 2026 deadline for Secure Boot certificate expiration

    February 11, 2026

    Sony confirms new WH-1000XM6 release in official teaser

    February 11, 2026

    Awkward debut: XPeng’s Iron robot falls on stage

    February 11, 2026
    Leave A Reply Cancel Reply

    Top Posts

    Ping, You’ve Got Whale: AI detection system alerts ships of whales in their path

    April 22, 2025667 Views

    Lumo vs. Duck AI: Which AI is Better for Your Privacy?

    July 31, 2025251 Views

    6.7 Cummins Lifter Failure: What Years Are Affected (And Possible Fixes)

    April 14, 2025151 Views

    6 Best MagSafe Phone Grips (2025), Tested and Reviewed

    April 6, 2025111 Views
    Don't Miss
    Technology February 11, 2026

    Microsoft sets 2026 deadline for Secure Boot certificate expiration

    Microsoft sets 2026 deadline for Secure Boot certificate expiration – NotebookCheck.net News ⓘ news.microsoft.comMicrosoft signage…

    Sony confirms new WH-1000XM6 release in official teaser

    Awkward debut: XPeng’s Iron robot falls on stage

    Limited edition Analogue 3D now available to buy

    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us
    About Us

    Welcome to Tech AI Verse, your go-to destination for everything technology! We bring you the latest news, trends, and insights from the ever-evolving world of tech. Our coverage spans across global technology industry updates, artificial intelligence advancements, machine learning ethics, and automation innovations. Stay connected with us as we explore the limitless possibilities of technology!

    Facebook X (Twitter) Pinterest YouTube WhatsApp
    Our Picks

    Microsoft sets 2026 deadline for Secure Boot certificate expiration

    February 11, 20263 Views

    Sony confirms new WH-1000XM6 release in official teaser

    February 11, 20262 Views

    Awkward debut: XPeng’s Iron robot falls on stage

    February 11, 20263 Views
    Most Popular

    7 Best Kids Bikes (2025): Mountain, Balance, Pedal, Coaster

    March 13, 20250 Views

    VTOMAN FlashSpeed 1500: Plenty Of Power For All Your Gear

    March 13, 20250 Views

    This new Roomba finally solves the big problem I have with robot vacuums

    March 13, 20250 Views
    © 2026 TechAiVerse. Designed by Divya Tech.
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions

    Type above and press Enter to search. Press Esc to cancel.