Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week

    Galaxy S26: Samsung officially discusses performance and efficiency of first-ever 2nm chips

    New Seiko Pink Panther watch appears in leaked image

    Facebook X (Twitter) Instagram
    • Artificial Intelligence
    • Business Technology
    • Cryptocurrency
    • Gadgets
    • Gaming
    • Health
    • Software and Apps
    • Technology
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Tech AI Verse
    • Home
    • Artificial Intelligence

      Insurance companies are trying to avoid big payouts by making AI safer

      November 19, 2025

      State and local opposition to new data centers is gaining steam, study shows

      November 15, 2025

      Amazon to lay off 14,000 corporate employees

      October 29, 2025

      Elon Musk launches Grokipedia as an alternative to ‘woke’ Wikipedia

      October 29, 2025

      Fears of an AI bubble are growing, but some on Wall Street aren’t worried just yet

      October 18, 2025
    • Business

      Windows 11 gets new Cloud Rebuild, Point-in-Time Restore tools

      November 18, 2025

      Government faces questions about why US AWS outage disrupted UK tax office and banking firms

      October 23, 2025

      Amazon’s AWS outage knocked services like Alexa, Snapchat, Fortnite, Venmo and more offline

      October 21, 2025

      SAP ECC customers bet on composable ERP to avoid upgrading

      October 18, 2025

      Revenue generated by neoclouds expected to exceed $23bn in 2025, predicts Synergy

      October 15, 2025
    • Crypto

      Nvidia Posts $57B Record Revenue with Bitcoin Rebounding Above $91K

      November 20, 2025

      3 Reasons Why A Cardano Price Rebound Looks Likely

      November 20, 2025

      BitMine (BMNR) Stock Bounces As Q4 Results Near — Is the Price Preparing Another Early Move?

      November 20, 2025

      Fed Minutes Reveal December Rate Cut on a Knife’s Edge, Bitcoin Slips Below $89,000

      November 20, 2025

      TRUMP Price Holds Above $7, Even As Epstein Files Release Approved

      November 20, 2025
    • Technology

      Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week

      November 20, 2025

      Galaxy S26: Samsung officially discusses performance and efficiency of first-ever 2nm chips

      November 20, 2025

      New Seiko Pink Panther watch appears in leaked image

      November 20, 2025

      NieR creator Yoko Taro says many projects were axed behind the scenes: ‘I haven’t been doing nothing’

      November 20, 2025

      Samsung Galaxy A57 to surpass downgraded Galaxy S26 in crucial battery area

      November 20, 2025
    • Others
      • Gadgets
      • Gaming
      • Health
      • Software and Apps
    Check BMI
    Tech AI Verse
    You are at:Home»Technology»Nvidia and Microsoft accelerate AI processing on PCs
    Technology

    Nvidia and Microsoft accelerate AI processing on PCs

    TechAiVerseBy TechAiVerseMay 19, 2025No Comments9 Mins Read2 Views
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    Nvidia and Microsoft accelerate AI processing on PCs
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp Email

    Nvidia and Microsoft accelerate AI processing on PCs

    May 19, 2025 9:00 AM

    AI PCs are getting Tensor RT for RTX.

    Image Credit: Nvidia

    Nvidia and Microsoft announced work to accelerate the performance of AI processing on Nvidia RTX-based AI PCs.

    Generative AI is transforming PC software into breakthrough experiences — from digital humans to writing assistants, intelligent agents and creative tools.

    Nvidia RTX AI PCs are powering this transformation with technology that makes it simpler to get started experimenting with generative AI, and unlocking greater performance on Windows 11.

    TensorRT for RTX AI PCs

    TensorRT has been reimagined for RTX AI PCs, combining industry leading TensorRT performance with just-in-time on-device engine building and an 8x smaller package size for fast AI deployment to the more than 100 million RTX AI PCs.

    Announced at Microsoft Build, TensorRT for RTX is natively supported by Windows ML — a new inference stack that provides app developers with both broad hardware compatibility and state of the art performance.

    Gerardo Delgado, director of product for AI PC at Nvidia, said in a press briefing that the AI PCs start with Nvidia’s RTX hardware, CUDA programming and an array of AI models. He noted that at a high level, an AI model is basically a set of mathematical operations along with a way to run them. And the combination of operations and how to run them is what is normally known as a graph in machine learning.

    He added, “Our GPUs are going to execute these operations with Tensor cores. But Tensor cores change from generation to generatio. We have been implementing them from time to time, and then within a generation of GPUs, you also have different Tensor code counts depending on the schema. Being able to match what’s the right Tensor code for each mathematical operation is the key to achieving performance. So a TensorRT does this in a two step approach.”

    First, Nvidia has to optimize the AI model. It has to quantize the model so it reduces the precision of parts of the model or some of the layers. Once Nvidia has optimized model, TensorRT consumes that optimized model, and then Nvidia basically prepares a plan with a pre-selection of kernels.”

    If you compare this to a standard way of running AI on Windows, Nvidia can achieve about a 1.6 times performance on average.

    Now there will be a new version of TensorRT for RTX to improve this experience. It’s designed specifically for RTX AI PCs and it provides the same TensorRT performance, but instead of having to pre-generate the TensorRT engines per GPU, it will focus on optimizing the model, and it will ship a generic TensorRT engine.

    “Then once the application is installed, TensorRT for RTX will generate the right TensorRT engine for your specific GPU in just seconds. This greatly simplifies the developer workflow,” he said.

    Among the results are a reduction in size of of libraries, better performance for video generation, and better quality livestreams, Delgado said.

    Nvidia SDKs make it easier for app developers to integrate AI features and accelerate their apps on GeForce RTX GPUs. This month top software applications from Autodesk, Bilibili, Chaos, LM Studio and Topaz are releasing updates to unlock RTX AI features and acceleration.

    AI enthusiasts and developers can easily get started with AI using Nvidia NIM, pre-packaged, optimized AI models that run in popular apps like AnythingLLM, Microsoft VS Code and ComfyUI. The FLUX.1-schnell image generation model is now available as a NIM, and the popular FLUX.1-dev NIM has been updated to support more RTX GPUs.

    For a no-code option to dive into AI development, Project G-Assist — the RTX PC AI assistant in the Nvidia app — has enabled a simple way to build plug-ins to create assistant workflows. New community plug-ins are now available including Google Gemini web search, Spotify, Twitch, IFTTT and SignalRGB.

    Accelerated AI inference with TensorRT for RTX

    Today’s AI PC software stack requires developers to choose between frameworks that have broad hardware support but lower performance, or optimized paths that only cover certain hardware or model types and require the developer to maintain multiple paths.

    The new Windows ML inference framework was built to solve these challenges. Windows ML is built on top of ONNX Runtime and seamlessly connects to an optimized AI execution layer provided and maintained by each hardware manufacturer. For GeForce RTX GPUs, Windows ML automatically uses TensorRT for RTX — an inference library optimized for high performance and rapid deployment. Compared to DirectML, TensorRT delivers over 50% faster performance for AI workloads on PCs.

    Windows ML also delivers quality of life benefits for the developer. It can automatically select the right hardware to run each AI feature, and download the execution provider for that hardware, removing the need to package those files into their app. This allows Nvidia to provide the latest TensorRT performance optimizations to users as soon as they are ready. And because it’s built on ONNX Runtime, Windows ML works with any ONNX model.

    To further enhance the experience for developers, TensorRT has been reimagined for RTX. Instead of having to pre-generate TensorRT engines and package them with the app, TensorRT for RTX uses just-in-time, on-device engine building to optimize how the AI model is run for the user’s specific RTX GPU in mere seconds. And the library has been streamlined, reducing its file size by a massive eight times. TensorRT for RTX is available to developers through the Windows ML preview today, and will be available directly as a standalone SDK at Nvidia Developer, targeting a June release.

    Developers can learn more in Nvidia’s Microsoft Build Developer Blog, the TensorRT for RTX launch blog, and Microsoft’s Windows ML blog.

    Expanding the AI ecosystem on Windows PCs

    Developers looking to add AI features or boost app performance can tap into a broad range of Nvidia SDKs. These include CUDA and TensortRT for GPU acceleration; DLSS and Optix for 3D graphics; RTX Video and Maxine for multimedia; and Riva, Nemotron or ACE for generative AI.

    Top applications are releasing updates this month to enable Nvidia unique features using these SDKs. Topaz is releasing a generative AI video model to enhance video quality accelerated by CUDA. Chaos Enscape and Autodesk VRED are adding DLSS 4 for faster performance and better image quality. BiliBili is integrating Nvidia Broadcast features, enabling streamers to activate Nvidia Virtual Background directly within Bilibili Livehime to enhance the quality of livestreams.

    Local AI made easy with NIM Microservices and AI blueprints

    Getting started with developing AI on PCs can be daunting. AI developers and enthusiasts have to select from over 1.2 million AI models on Hugging Face, quantize it into a format that runs well on PC, find and install all the dependencies to run it, and more. Nvidia NIM makes it easy to get started by providing a curated list of AI models, pre-packaged with all the files needed to run them, and optimized to achieve full performance on RTX GPUs. And as containerized microservices, the same NIM can be run seamlessly across PC or cloud.

    A NIM is a package — a generative AI model that’s been prepackaged with everything you need to run it.

    It’s already optimized with TensorRT for RTX GPUs, and it comes with an easy to use API that’s open-API compatible, which makes it compatible with all of the top AI applications that users are using today.

    At Computex, Nvidia is releasing the FLUX.1-schnell NIM — an image generation model from Black Forest Labs for fast image generation — and updating the FLUX.1-dev NIM to add compatibility for a wide range of GeForce RTX 50 and 40 Series GPUs. These NIMs enable faster performance with TensorRT, plus additional performance thanks to quantized models. On Blackwell GPUs, these run over twice as fast as running them natively, thanks to FP4 and RTX optimizations.

    AI developers can also jumpstart their work with Nvidia AI Blueprints — sample workflows and projects using NIM.

    Last month Nvidia released the 3D Guided Generative AI Blueprint, a powerful way to control composition and camera angles of generated images by using a 3D scene as a reference. Developers can modify the open source blueprint for their needs or extend it with additional functionality.

    New Project G-Assist plug-ins and sample projects now available

    Nvidia recently released Project G-Assist as an experimental AI assistant integrated into the Nvidia app. G-Assist enables users to control their GeForce RTX system using simple voice and text commands, offering a more convenient interface compared to manual controls spread across multiple legacy control panels.

    Developers can also use Project G-Assist to easily build plug-ins, test assistant use cases and publish them through Nvidia’s Discord and GitHub.

    To make it easier to get started creating plug-ins, Nvidia has made available the easy-to use Plug-in Builder — a ChatGPT-based app that allows no-code/low-code development with natural language commands. These lightweight, community-driven add-ons leverage straightforward JSON definitions and Python logic.

    New open-source samples are available now on GitHub, showcasing diverse ways how on device AI can enhance your PC and gaming workflows.

    ● Gemini: The existing Gemini plug-in that uses Google’s cloud-based free-to-use LLM has been updated to include real-time web search capabilities.

    ● IFTTT: Enable automations from the hundreds of end points that work with IFTTT, such as IoT and home automation systems, enabling routines spanning digital setups and physical surroundings.

    ● Discord: Easily share game highlights, or messages directly to Discord servers without disrupting gameplay.

    Explore the GitHub repository for additional examples — including hands-free music control via Spotify, livestream status checks with Twitch, and more.

    Project G-Assist — AI Assistant For Your RTX PC

    Companies are also adopting AI as the new PC interface. For example, SignalRGB is developing a G-Assist plugin that enables unified lighting control across multiple manufacturers. SignalRGB users will soon be able to install this plug-in directly from the SignalRGB app.

    Enthusiasts interested in developing and experimenting with Project G-Assist plug-ins are invited to join the Nvidia Developer Discord channel to collaborate, share creations and receive support during development.

    Each week, the RTX AI Garage blog series features community-driven AI innovations and content for those looking to learn more about NIM microservices and AI Blueprints, as well as building AI agents, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.

    Daily insights on business use cases with VB Daily

    If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

    Read our Privacy Policy

    Thanks for subscribing. Check out more VB newsletters here.

    An error occured.

    Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
    Previous ArticleChinese cyber spooks lure laid-off US government workers
    Next Article GitHub Copilot evolves into autonomous agent with asynchronous code testing
    TechAiVerse
    • Website

    Jonathan is a tech enthusiast and the mind behind Tech AI Verse. With a passion for artificial intelligence, consumer tech, and emerging innovations, he deliver clear, insightful content to keep readers informed. From cutting-edge gadgets to AI advancements and cryptocurrency trends, Jonathan breaks down complex topics to make technology accessible to all.

    Related Posts

    Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week

    November 20, 2025

    Galaxy S26: Samsung officially discusses performance and efficiency of first-ever 2nm chips

    November 20, 2025

    New Seiko Pink Panther watch appears in leaked image

    November 20, 2025
    Leave A Reply Cancel Reply

    Top Posts

    Ping, You’ve Got Whale: AI detection system alerts ships of whales in their path

    April 22, 2025409 Views

    Lumo vs. Duck AI: Which AI is Better for Your Privacy?

    July 31, 2025109 Views

    6.7 Cummins Lifter Failure: What Years Are Affected (And Possible Fixes)

    April 14, 202575 Views

    Is Libby Compatible With Kobo E-Readers?

    March 31, 202555 Views
    Don't Miss
    Technology November 20, 2025

    Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week

    Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week – NotebookCheck.net News…

    Galaxy S26: Samsung officially discusses performance and efficiency of first-ever 2nm chips

    New Seiko Pink Panther watch appears in leaked image

    NieR creator Yoko Taro says many projects were axed behind the scenes: ‘I haven’t been doing nothing’

    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us
    About Us

    Welcome to Tech AI Verse, your go-to destination for everything technology! We bring you the latest news, trends, and insights from the ever-evolving world of tech. Our coverage spans across global technology industry updates, artificial intelligence advancements, machine learning ethics, and automation innovations. Stay connected with us as we explore the limitless possibilities of technology!

    Facebook X (Twitter) Pinterest YouTube WhatsApp
    Our Picks

    Hubble image of the ‘Lost Galaxy’ emerges as Picture of the Week

    November 20, 20250 Views

    Galaxy S26: Samsung officially discusses performance and efficiency of first-ever 2nm chips

    November 20, 20250 Views

    New Seiko Pink Panther watch appears in leaked image

    November 20, 20250 Views
    Most Popular

    Xiaomi 15 Ultra Officially Launched in China, Malaysia launch to follow after global event

    March 12, 20250 Views

    Apple thinks people won’t use MagSafe on iPhone 16e

    March 12, 20250 Views

    French Apex Legends voice cast refuses contracts over “unacceptable” AI clause

    March 12, 20250 Views
    © 2025 TechAiVerse. Designed by Divya Tech.
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions

    Type above and press Enter to search. Press Esc to cancel.