Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media hires

    Why brands are shifting toward ‘less precise, more accurate’ gauges for paid social

    WTF is Markdown for AI agents? 

    Facebook X (Twitter) Instagram
    • Artificial Intelligence
    • Business Technology
    • Cryptocurrency
    • Gadgets
    • Gaming
    • Health
    • Software and Apps
    • Technology
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Tech AI Verse
    • Home
    • Artificial Intelligence

      Read the extended transcript: President Donald Trump interviewed by ‘NBC Nightly News’ anchor Tom Llamas

      February 6, 2026

      Stocks and bitcoin sink as investors dump software company shares

      February 4, 2026

      AI, crypto and Trump super PACs stash millions to spend on the midterms

      February 2, 2026

      To avoid accusations of AI cheating, college students are turning to AI

      January 29, 2026

      ChatGPT can embrace authoritarian ideas after just one prompt, researchers say

      January 24, 2026
    • Business

      The HDD brand that brought you the 1.8-inch, 2.5-inch, and 3.5-inch hard drives is now back with a $19 pocket-sized personal cloud for your smartphones

      February 12, 2026

      New VoidLink malware framework targets Linux cloud servers

      January 14, 2026

      Nvidia Rubin’s rack-scale encryption signals a turning point for enterprise AI security

      January 13, 2026

      How KPMG is redefining the future of SAP consulting on a global scale

      January 10, 2026

      Top 10 cloud computing stories of 2025

      December 22, 2025
    • Crypto

      Binance Denies Sanctions Breach Claims After $1 Billion Iran-Linked USDT Transactions Reported

      February 16, 2026

      Ray Dalio Says the World Order Has Broken Down: What Does It Mean for Crypto?

      February 16, 2026

      Cardano Whales are Trying to Rescue ADA Price

      February 16, 2026

      MYX Finance Lost 70% In a Week: What Triggered the Sharp Sell-Off?

      February 16, 2026

      What Really Happened Between Binance and FTX? CZ Finally Tells His Side

      February 16, 2026
    • Technology

      Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media hires

      February 16, 2026

      Why brands are shifting toward ‘less precise, more accurate’ gauges for paid social

      February 16, 2026

      WTF is Markdown for AI agents? 

      February 16, 2026

      ‘Being very careful’: Weeks after unveiling ad plans, OpenAI works to control the message

      February 16, 2026

      Hideki Sato, known as the father of Sega hardware, has reportedly died

      February 16, 2026
    • Others
      • Gadgets
      • Gaming
      • Health
      • Software and Apps
    Check BMI
    Tech AI Verse
    You are at:Home»Technology»Forget training, find your killer apps during AI inference
    Technology

    Forget training, find your killer apps during AI inference

    TechAiVerseBy TechAiVerseOctober 10, 2025No Comments6 Mins Read2 Views
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Forget training, find your killer apps during AI inference
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp Email

    Forget training, find your killer apps during AI inference

    Most organisations will never train their own AI models. Instead, most customer’s key challenge in AI lies in applying it to production applications and inference, with fine tuning and curation of data the core tasks.

    Key here are use of retrieval augmented generation (RAG) and vector databases, the ability to reuse AI prompts, and co-pilot capabilities that allow users to question corporate information in natural language.

    Those are the views of Pure Storage execs who spoke to Computerweekly.com this week at the company’s Accelerate event in London.

    Naturally, the key tasks identified fit well with areas of functionality added recently to Pure’s storage hardware offer – including its recently launched Key Value Accelerator – and also with its ability to provide capacity on demand.

    But they also illustrate the key challenges for organisations tackling AI at this stage in its maturity, which has been called a “post-training phase”.

    In this article, we look at what customers need from storage in AI in production phases, and with ongoing ingestion of data and inference taking place.

    Don’t buy GPUs; they’re changing too quickly

    Most organisations won’t train their own AI models because it’s simply too expensive at the moment. That’s because GPU hardware is incredibly costly to buy and also because it is evolving at such a rapid pace that obsolescence comes very soon.

    So, most organisations now tend to buy GPU capacity in the cloud for training phases.

    It’s pointless trying to build in-house AI training farms when GPU hardware can become obsolete within a generation or two.

    That’s the view of Pure Storage founder and chief visionary officer John “Coz” Colgrove.

    “Most organisations say, ‘Oh, I want to buy this equipment, I’ll get five years of use out of it, and I’ll depreciate it over five or seven years,” he said. “But you can’t do that with the GPUs right now.

    “I think when things improve at a fantastic rate, you’re better off leasing instead of buying. It’s just like buying a car,” said Colgrove. “If you’re going to keep it for six, seven, eight years or more, you buy it, but if you’re going to keep it for two years and change to a newer one, you lease it.”

    Find your AI killer app

    For most organisations practical exploitation of AI won’t happen in the modelling phase. Instead, it’s going to come where they can use it to build a killer app for their business.

    Colgrove gives the example of a bank. “With a bank we know the killer app is going to be something customer facing,” he said. “But how does AI work right now? I take all my data out of whatever databases I have for interacting with the customer. I suck it into some other system. I transform it like an old ETL batch process, spend weeks training on it and then I get a result.

    “That is never going to be the killer app,” said Colgrove. “The killer app will involve some kind of inferencing I can do. But that inferencing is going to have to be applied in the regular systems if it’s customer facing.

    “That means when you actually apply the AI to get value out of it, you’ll want to apply it to the data you already have, the things you’re already doing with your customers.”

    In other words, for most customers the challenges of AI lie in the production phase and more precisely the ability to (rapidly) curate and add data, and run inference on it to fine tune existing AI models. And then to be able to do that all again when you have the next idea how to further improve things.

    Pure Storage EMEA filed chief technology officer Fred Lherault, summed it up thus: “So it’s really about how do I connect models to my data? Which first of all means, have I done the right level of finding what my data is, curating my data, making it AI ready, and putting it into an architecture where it can be accessed by a model?”

    Key tech underpinnings of agile AI

    So, the inference phase has emerged as the key focus for most AI customers. Here, the challenge is to be able to curate and manage the data to build and re-iterate upon AI models during their production lifetime. That means customers connecting with their own data in an agile fashion.

    This means the use of technologies that include vector databases, RAG pipelines, co-pilot capability, and prompt caching and reuse.

    Key challenges for storage as it relates to these are twofold. It means being able to connect to, eg, RAG data sources, and to vector databases. It also means being able to handle big jumps in storage capacity, and to reduce the need to do so. The two are often connected.

    “An interesting thing happens when you put your data into vector databases,” said Lherault. “There’s some computation required, but then the data gets augmented with vectors that can then be searched. That’s the whole goal of the vector database, and that augmentation can sometimes result in a 10x amplification of data.

    “If you’ve got a terabyte of source data you want to use with an AI model, it means you’ll need a 10TB database to run it,” he said. “There’s all of that process that is new for many organisations when they want to use their data with AI models.”

    Deal with demands on storage capacity

    Such capacity jumps can also occur during tasks such as checkpointing, which can see huge volumes of data created as snapshot-like points to roll back to in AI processing.

    Pure aims to tackle these with its Evergreen as-a-service model, which allows customers to rapidly add to capacity.

    The company also suggests ways to keep storage volumes from rising too rapidly, as well as speeding performance.

    Its recently introduced Key Value Accelerator allows customers to store AI prompts so they can be reused. Ordinarily, an LLM would access cached tokens representing previous responses, but GPU cache is limited, so answers often need to be re-calculated anew. Pure’s KV Accelerator allows tokens to be held in its storage in file or object format.

    That can speed responses by up to 20x, said Lherault. “The more you start having users asking different questions, the faster you run out of cache,” he added. “If you’ve got two users asking the same question at the same time and do that on two GPUs, they both have to do the same computation. It’s not very efficient.

    “We’re allowing it to actually store those pre-computed key values on our storage so the next time someone asks a question that’s already been asked or requires the same token, if we’ve got it on our side, the GPU doesn’t need to do the computation,” said Lherault.

    “It helps to reduce the number of GPUs you need, but also on some complex questions that generate thousands of tokens, we’ve seen sometimes the answer coming 20 times faster.”

    Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
    Previous ArticleAI and autonomous vehicles drive UAE’s first commercial electric truck fleet
    Next Article Warlock ransomware may be linked to Chinese state
    TechAiVerse
    • Website

    Jonathan is a tech enthusiast and the mind behind Tech AI Verse. With a passion for artificial intelligence, consumer tech, and emerging innovations, he deliver clear, insightful content to keep readers informed. From cutting-edge gadgets to AI advancements and cryptocurrency trends, Jonathan breaks down complex topics to make technology accessible to all.

    Related Posts

    Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media hires

    February 16, 2026

    Why brands are shifting toward ‘less precise, more accurate’ gauges for paid social

    February 16, 2026

    WTF is Markdown for AI agents? 

    February 16, 2026
    Leave A Reply Cancel Reply

    Top Posts

    Ping, You’ve Got Whale: AI detection system alerts ships of whales in their path

    April 22, 2025680 Views

    Lumo vs. Duck AI: Which AI is Better for Your Privacy?

    July 31, 2025260 Views

    6.7 Cummins Lifter Failure: What Years Are Affected (And Possible Fixes)

    April 14, 2025154 Views

    6 Best MagSafe Phone Grips (2025), Tested and Reviewed

    April 6, 2025112 Views
    Don't Miss
    Technology February 16, 2026

    Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media hires

    Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media…

    Why brands are shifting toward ‘less precise, more accurate’ gauges for paid social

    WTF is Markdown for AI agents? 

    ‘Being very careful’: Weeks after unveiling ad plans, OpenAI works to control the message

    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us
    About Us

    Welcome to Tech AI Verse, your go-to destination for everything technology! We bring you the latest news, trends, and insights from the ever-evolving world of tech. Our coverage spans across global technology industry updates, artificial intelligence advancements, machine learning ethics, and automation innovations. Stay connected with us as we explore the limitless possibilities of technology!

    Facebook X (Twitter) Pinterest YouTube WhatsApp
    Our Picks

    Media Buying Briefing: Attivo breathes new life into Hill Holliday and DNY with senior media hires

    February 16, 20263 Views

    Why brands are shifting toward ‘less precise, more accurate’ gauges for paid social

    February 16, 20263 Views

    WTF is Markdown for AI agents? 

    February 16, 20263 Views
    Most Popular

    7 Best Kids Bikes (2025): Mountain, Balance, Pedal, Coaster

    March 13, 20250 Views

    VTOMAN FlashSpeed 1500: Plenty Of Power For All Your Gear

    March 13, 20250 Views

    This new Roomba finally solves the big problem I have with robot vacuums

    March 13, 20250 Views
    © 2026 TechAiVerse. Designed by Divya Tech.
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions

    Type above and press Enter to search. Press Esc to cancel.