NVIDIA NeMo

Build, customize, and deploy multimodal generative and agentic AI applications.

Overview

What Is NVIDIA NeMo?

NVIDIA NeMo™ is an end-to-end platform for developing custom generative AI—including large language models (LLMs), vision language models (VLMs), retrieval models, video models,  and speech AI—anywhere. 

With NeMo, you can easily build data flywheels to continuously optimize AI agents with the latest information. NeMo accelerates the data flywheel by curating AI and human feedback, refining and evaluating models, and deploying with guardrails and retrieval-augmented generation (RAG) to keep agents delivering peak performance. 

Build and maintain your agents with NVIDIA NeMo—secure, scalable, and supported enterprise-grade software, part of NVIDIA AI Foundry.

Maximizing AI Agent Performance with NVIDIA NeMo Microservices

Get an introduction to the NVIDIA NeMo microservices, offering developers insights into its key capabilities for building adaptive, and secure enterprise agentic AI systems.

Build Scalable Data Flywheels for Continuously Improving Generative AI Applications

Learn how NeMo enables enterprises to quickly and easily collect and process data, customize the generative AI models, evaluate the model's performance, and implement guardrails to ensure responsible and ethical use of the model.

Benefits

Explore the Benefits of NVIDIA NeMo for Generative AI

Continuously Improve AI Agent Performance

Easily build data flywheels that use enterprise data to improve AI agents, powering the entire flywheel with a simple Helm chart deployment or API calls for various parts of the workflow.

Production Ready

Deploy into production with a secure, optimized, full-stack solution that offers support, security, and API stability as part of NVIDIA AI Enterprise.

Increased ROI

Quickly train, customize, and deploy large language models (LLMs), VLMs, video, and speech AI at scale, reducing time to solution and increasing ROI.

Accelerated Performance

Maximize throughput and minimize training time with multi-node, multi-GPU training and inference.

Run Anywhere

Train and deploy generative AI anywhere, across clouds, data centers, and the edge.

Superior Visual Generation

State-of-the-art reconstruction quality using Cosmos tokenizer across a wide spectrum of image and video categories.

Features

The Complete Solution for Building Enterprise-Ready Generative AI Models

NeMo Framework
Accelerated Training and Customization

NVIDIA NeMo framework provides extensive configurability and advanced training and RL techniques with NeMo-Aligner for building and customizing reasoning and generative AI models.

NeMo Curator
Accelerate Data Curation

NVIDIA NeMo Curator improves generative AI model accuracy by processing text, image, and video data at scale. It also provides pre-built pipelines for generating synthetic data to customize and evaluate generative AI systems.

With NeMo Curator, you can accelerate video processing from years to days, compared to alternatives.

Video and Image Tokenizer
Generate High-Quality Visuals

NVIDIA Cosmos tokenizers are open models designed to simplify the development and customization of VLMs and video AI models. They offer high-quality compression and fast, excellent visual reconstruction, lowering TCO during model development and deployments.

NeMo Customizer
Simplify Fine-Tuning

NVIDIA NeMo Customizer is a high-performance, scalable microservice that simplifies fine-tuning and alignment of LLMs for domain-specific use cases, making it easier to adopt generative AI across industries.

NeMo Evaluator
Evaluate Models

NVIDIA NeMo Evaluator provides a microservice to assess generative AI models and pipelines across academic and custom benchmarks on any platform.

NeMo Retriever
Seamless Data Retrieval

NVIDIA NeMo™ Retriever is a collection of microservices that provide world-class information retrieval with high accuracy and maximum data privacy.

NeMo Guardrails
Generative AI Guardrails

NVIDIA NeMo Guardrails is a scalable rail orchestration platform for ensuring the security, safety, accuracy, and topical relevance of LLM interactions.

NVIDIA NIM™
Generative AI Inference

NVIDIA NIM, part of NVIDIA AI Enterprise, is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI model inferencing across clouds, data centers, and workstations.

Use Cases

How NeMo Is Being Used

See how NVIDIA NeMo supports industry use cases and jump-starts your AI development.

AI Agents

AI agents are transforming customer service across sectors, helping companies enhance customer conversations, achieve high resolution rates, and improve human representative productivity. AI agents can handle predictive tasks, reason and problem-solve, be trained to understand industry-specific terms, and pull relevant information from an organization’s knowledge bases, wherever that data resides.

AI Assistant

Businesses are deploying AI assistants to efficiently address the queries of millions of customers and employees around the clock. Powered by customized NVIDIA NIM microservices for LLMs, RAG, and speech and translation AI, these AI teammates deliver immediate and accurate spoken responses, even in the presence of background noise, poor sound quality, and diverse dialects and accents.

Information Retrieval

Trillions of PDF files are generated every year, each file likely consisting of multiple pages filled with various content types, including text, images, charts, and tables. This goldmine of data can only be used as quickly as humans can read and understand it. But with generative AI and RAG, this untapped data can be used to uncover business insights that can help employees work more efficiently and result in lower costs.

Content Generation

Generative AI makes it possible to generate highly relevant, bespoke, and accurate content grounded in the domain expertise and proprietary IP of your enterprise.

Humanoid Robot

Humanoid robots are built to adapt quickly to existing human-centric urban and industrial work spaces, tackling tedious, repetitive, or physically demanding tasks. Their versatility has them in such varied locations as factory floors to healthcare facilities, where these robots are assisting humans and helping alleviate labor shortages with automation.

Apptronik

Starting Options

Ways to Get Started With NVIDIA NeMo

Use the right tools and technologies to take generative AI models from development to production.

Try

Start prototyping with leading NVIDIA-built and open-source generative AI models that can be deployed using NVIDIA NIM™ microservices and customized with NeMo.

Build

Jump-start building your generative AI solutions with NVIDIA Blueprints, customizable reference applications, available for free on the NVIDIA API catalog.

Develop

For those looking to use NeMo for development, the software is available to download for free or apply for early access.

Deploy

Get a free license to try NVIDIA AI Enterprise in production for 90 days using your existing infrastructure. 

Customer Stories

How Industry Leaders Are Driving Innovation With NeMo

Dropbox
Bringing Personalized Generative AI to Customers

Dropbox plans to leverage NVIDIA’s AI foundry to build custom models and improve AI-powered knowledge work with the Dropbox Dash universal search tool and Dropbox AI.

Perplexity
Enhance Model Performance for AI-Powered Search Engines

Using NVIDIA NeMo, Perplexity aims to quickly customize frontier models to improve the accuracy and quality of search results and optimize them for lower latency and high throughput for a better user experience.

Amdocs
Bringing Custom Generative AI to the Global Telco Industry

Amdocs plans to build custom LLMs for the $1.7 trillion global telecommunications industry using NVIDIA’s AI foundry on Microsoft Azure.

Adopters

Leading Adopters Across All Industries

Resources

The Latest in NVIDIA NeMo Resources

Get Started With LLM Customization

In this course, you’ll go beyond prompt-engineering LLMs and learn techniques to efficiently customize pretrained LLMs for your specific use cases. Using NVIDIA NIM microservices, NeMo Curator, and NeMo Framework, you’ll learn various parameter-efficient fine-tuning methods to customize LLM behavior for your organization.

Elevate Your LLM Skills

Take advantage of our comprehensive LLM learning path, covering fundamental to advanced topics featuring hands-on training developed and delivered by NVIDIA experts. You can opt for the flexibility of self-paced courses or enroll in instructor-led workshops to earn a certificate of competency.

Get Certified by NVIDIA

Showcase your Generative AI skills and advance your career by getting certified by NVIDIA. Our new professional certification program offers two developer exams focusing on proficiency in large language models (LLMs) and multimodal workflow skills.

Building and Deploying Generative AI Models

Enterprises are turning to generative AI to revolutionize the way they innovate, optimize operations, and build a competitive advantage. NeMo is an end-to-end platform for curating data; training, customizing, and evaluating multimodal models; and running inference at scale. It supports text, image, video, and speech generation.

Unlocking Synthetic Data Generation with Llama 3.1

Learn how to use the Meta Llama 3.1 405B model to generate tailored synthetic data for your specific domain and explore how to evaluate this data using the Nemotron-4 340B Reward model and ensure alignment with human preferences through NVIDIA NeMo.

Build World-Class AI Virtual Assistants for Customer Service with RAG

Learn how companies can use the AI virtual assistant for customer service NVIDIA AI Blueprint to improve the operational efficiency of existing contact center solutions or build new customer service-centric systems.

Next Steps

Ready to Get Started?

Use the right tools and technologies to take generative AI models from development to production.

For Developers

Explore everything you need to start developing with NVIDIA NeMo, including the latest documentation, tutorials, technical blogs, and more.

Get in Touch

Talk to an NVIDIA product specialist about moving from pilot to production with the assurance of security, API stability, and support that comes with NVIDIA AI Enterprise.

AI Sweden

Accelerate Industry Applications With LLMs

AI Sweden facilitated regional language model applications by providing easy access to a powerful 100 billion-parameter model. They digitized historical records to develop language models for commercial use.

Amazon

How Amazon and NVIDIA Help Sellers Create Better Product Listings With AI

Amazon doubles inference speeds for new AI capabilities using NVIDIA TensorRT-LLM and GPUs to help sellers optimize product listings faster.

Amdocs

NVIDIA and Amdocs Bring Custom Generative AI to Global Telco Industry

Amdocs plans to build custom LLMs for $1.7 trillion global telecommunications industry using NVIDIA AI foundry service on Microsoft Azure.

AWS

NVIDIA Powers Training for Some of the Largest Amazon Titan Foundation Models

Amazon leveraged the NVIDIA NeMo framework, GPUs, and AWS EFAs to train its next-generation LLM, giving some of the largest Amazon Titan foundation models customers a faster, more accessible solution for generative AI.

Accenture

Accelerate Generative AI Adoption for Enterprises

ServiceNow, NVIDIA, and Accenture announced the launch of AI Lighthouse, a first-of-its-kind program designed to fast-track the development and adoption of enterprise generative AI capabilities.

Azure

Harnessing the Power of NVIDIA AI Enterprise on Azure Machine Learning

Get access to a complete ecosystem of tools, libraries, frameworks, and support services tailored for enterprise environments on Microsoft Azure.

Bria

Bria Builds Responsible Generative AI for Enterprises Using NVIDIA NeMo, Picasso

Bria, a startup based in Tel Aviv, is helping businesses who are seeking responsible ways to integrate visual generative AI technology into their enterprise products with a generative AI service that emphasizes model transparency alongside fair attribution and copyright protections.

Cohesity

Unlock Your Data Superpower: NVIDIA Microservices Unleash Enterprise-Grade Secure Generative AI for Cohesity

With NVIDIA NIM and optimized models, Cohesity DataProtect customers can add generative AI intelligence to data backups and archives. This allows Cohesity and NVIDIA to bring the power of generative AI to all Cohesity DataProtect customers. Leveraging the power of NIM and NVIDIA optimized models, Cohesity DataProtect customers obtain the power of data-driven insights from their data backups and archives, unleashing new levels of efficiency, innovation, and growth.

CrowdStrike

Shaping the Future of AI in the Cybersecurity Domain

CrowdStrike and NVIDIA are leveraging accelerated computing and generative AI to provide customers with an innovative range of AI-powered solutions tailored to efficiently address security threats.

Dell

Dell Validated Design for Generative AI With NVIDIA

Dell Technologies and NVIDIA announced an initiative to make it easier for businesses to build and use generative AI models on premises quickly and securely.

Deloitte

Unlock the Value of Generative AI Across Enterprise Software Platforms

Deloitte will use NVIDIA AI technology and expertise to build high-performing generative AI solutions for enterprise software platforms to help unlock significant business value.

Domino Data Lab

Domino Offers Production-Ready Generative AI Powered by NVIDIA

With NVIDIA NeMo, data scientists can fine-tune LLMs in Domino’s platform for domain-specific use cases based on proprietary data and IP—without needing to start from scratch. 

Dropbox

Dropbox and NVIDIA to Bring Personalized Generative AI to Millions of Customers

Dropbox plans to leverage NVIDIA’s AI foundry to build custom models and improve AI-powered knowledge work with Dropbox Dash universal search tool and Dropbox AI.

Google Cloud

AI Titans Collaborate to Create Generative AI Magic

At its Next conference, Google Cloud announced the availability of its A3 instances powered by NVIDIA H100 Tensor Core GPUs. Engineering teams from both companies have collaborated to bring NVIDIA NeMo to the A3 instances for faster training and inference.

HuggingFace

Leading AI Community to Accelerate Data Curation Pipeline

Hugging Face, the leading open platform for AI builders, is collaborating with NVIDIA to integrate NeMo Curator and accelerate DataTrove, their data filtering and deduplication library. “We are excited about the GPU acceleration capabilities of NeMo Curator and can’t wait to see them contributed to DataTrove!” says Jeff Boudier, Product Director at Hugging Face.

KT

Creating New Customer Experiences With LLMs

South Korea’s leading mobile operator builds billion-parameter LLMs trained with the NVIDIA DGX SuperPOD platform and NeMo framework to power smart speakers and customer call centers.

Lenovo

New Reference Architecture for Generative AI Based on LLMs

Solution to expedite innovation by empowering global partners and customers to develop, train, and deploy AI at scale across industry verticals with utmost safety and efficiency.

Quantiphi

Enabling Enterprises to Fast-Track Their AI-Driven Journeys

Quantiphi specializes in training and fine-tuning foundation models using the NVIDIA NeMo framework, as well as optimizing deployments at scale with the NVIDIA AI Enterprise software platform, while adhering to responsible AI principles.

SAP

SAP and NVIDIA Accelerate Generative AI Adoption Across Enterprise Applications Powering Global Industries

Customers can harness their business data in cloud solutions from SAP using customized LLMs deployed with NVIDIA AI foundry services and NVIDIA NIM Microservices.

ServiceNow

Building Generative AI Across Enterprise IT

ServiceNow develops custom LLMs on its ServiceNow platform to enable intelligent workflow automation and boost productivity across enterprise IT processes.

Perplexity

Enhance Model Performance for AI-Powered Search Engines

Using NVIDIA NeMo, Perplexity aims to quickly customize frontier models to improve the accuracy and quality of search results and optimize them for lower latency and high throughput for a better user experience.

VMware

VMware and NVIDIA Unlock Generative AI for Enterprises

VMware Private AI Foundation with NVIDIA will enable enterprises to customize models and run generative AI applications, including intelligent chatbots, assistants, search, and summarization.

Weight & Biases

Debug, Optimize, and Monitor LLM Pipelines 

Weights & Biases helps teams working on generative AI use cases or with LLMs track and visualize all prompt-engineering experiments—helping users debug and optimize LLM pipelines—as well as provides monitoring and observability capabilities for LLMs. 

Writer

Startup Pens Generative AI Success Story With NVIDIA NeMo

Using NVIDIA NeMo, Writer is building LLMs that are helping hundreds of companies create custom content for enterprise use cases across marketing, training, support, and more.