-> Back to overview

Lunch & Learns

NVIDIA's Playbook: From Raw GPUs to a Functioning AI Factory

-> Back to overview

Key takeaways

1

NVIDIA’s NIMs are designed to simplify the complex process of deploying AI models.

2

AI factories are in-house systems that turn a company's raw data into valuable intelligence.

3

Running AI models locally is a key strategy for securing sensitive enterprise data.

Building with NIMs: The 'Lego Kit' Approach to Local AI

From Gaming Hardware to AI Infrastructure

Artificial intelligence has moved from a speculative topic to a core part of modern business operations. The central question for companies is no longer if they should use AI, but how to build with it effectively. During a recent session, NVIDIA’s Jean Pierre Van Gastel and Rebatch's Jorge de Corte broke down the industry’s next logical step: the shift from consuming AI services to building dedicated, in-house "AI Factories."

To explain NVIDIA’s current strategy, Van Gastel started with the company's origins. Founded in 1993, NVIDIA initially focused on 3D graphics for the gaming industry, solving the complex challenge of real-time rendering. He highlighted the introduction of CUDA (Compute Unified Device Architecture) as a critical turning point. CUDA allowed developers to access the parallel processing capabilities of GPUs for general-purpose computing, essentially turning graphics cards into powerful, versatile processors.

This foundation, built to serve gamers, positioned NVIDIA perfectly to power the AI boom. The launch of ChatGPT, which Van Gastel called the "iPhone moment" for AI, was trained on NVIDIA hardware. This widespread adoption created a new challenge for businesses: how to transition from being a consumer of AI to a producer.

Building with NIMs: The 'Lego Kit' Approach to Local AI

Van Gastel framed the journey with a simple analogy: building with Lego. A company can take a do-it-yourself approach by sorting through the more than two million open-source models available on platforms like Hugging Face. He compared this to building something from a giant, unsorted bin of bricks: it offers maximum flexibility but is incredibly complex to assemble and maintain.

The alternative is a complete Lego kit with an instruction manual. This is the idea behind NVIDIA Inference Microservices (NIMs). Van Gastel described NIMs as pre-packaged, optimized, and secure AI models that can be deployed in about five minutes. They are essentially ready-to-use building blocks for an AI factory.

To show what this looks like in practice, Jorge de Corte stepped in. He addressed a common business problem: clients who want to use AI for tasks like code assistance but are prohibited from sending sensitive data to external cloud APIs. Using a compact NVIDIA Spark desktop machine, de Corte benchmarked a locally-run open-source model against leading cloud services.

The results were impressive. The local model performed nearly as well as a top-tier cloud model from six months prior, and the speed was more than sufficient for a single user's productivity. While the affordable Spark system is geared toward individuals, de Corte noted that larger NVIDIA servers can easily scale to support entire teams. He concluded that NIMs make this local-first approach practical by removing the significant software setup and configuration headaches.

 

The Next Step: Digital Twins and Physical AI

With text and image generation becoming more accessible, the conversation shifted to what’s next. Van Gastel identified the next frontier as Physical AI, which is focused on teaching AI to understand and interact with the three-dimensional world. He introduced NVIDIA Omniverse as the platform for building this future. Omniverse is a development environment for creating and operating 'digital twins'—photorealistic, physics-accurate virtual replicas of real-world environments like factories, warehouses, or even entire cities.

Van Gastel explained the practical value of this technology. Instead of training a warehouse robot in a live facility, where mistakes can be costly or dangerous, it can be trained for thousands of hours in a high-fidelity virtual simulation. This allows the AI to learn from countless scenarios safely and efficiently before being deployed in the real world. As many industries face labor shortages, the ability to automate physical tasks is becoming increasingly important. Omniverse, which is built on the open standard OpenUSD, provides the foundational technology to make concepts like automated logistics and humanoid robots a commercial reality.

The session with Van Gastel and de Corte clarified the path from simply using AI to building custom solutions with it. NVIDIA’s strategy isn't just about selling powerful GPUs; it’s about offering a full-stack roadmap for development. This stack ranges from the CUDA platform and easy-to-deploy NIMs to the advanced simulation capabilities of Omniverse. This layered approach makes building secure, proprietary AI systems a practical goal for companies of all sizes. The journey to creating an in-house AI Factory is underway, and the initial steps are more accessible than ever before.

Ready to start building your own AI factory and explore what's possible? Contact us for more information and practical insights

Looking for a sparring partner for your AI journey?

Contact us to discover how Cronos.AI can help your business.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.