NVIDIA AI Foundry is a platform and service for building custom generative AI models with enterprise data and domain-specific knowledge. Just as TSMC manufactures chips designed by other companies, NVIDIA AI Foundry enables organizations to develop their own AI models.
A chip foundry provides state-of-the-art transistor technology, manufacturing process, large chip fabs, expertise, and a rich ecosystem of third-party tools and library providers. Similarly, NVIDIA AI Foundry includes NVIDIA-created AI models like Nemotron and Edify, popular open foundation models, NVIDIA NeMo™ software for customizing models, and dedicated capacity on NVIDIA DGX™ Cloud—built and backed by NVIDIA AI experts. The output is NVIDIA NIM™—an inference microservice that includes the custom model, optimized engines, and a standard API—which can be deployed anywhere.
NVIDIA AI Foundry and its libraries are integrated into the world’s leading AI ecosystem of startups, enterprise software providers, and global service providers.
Customize Generative AI Models for Enterprise Applications With Llama 3.1
With NVIDIA AI Foundry, enterprises can build generative AI solutions tailored for their specific organizational, industrial, or cultural needs.
The NVIDIA AI Foundry uses enterprise data, along with synthetically generated data, to augment and alter the general knowledge contained in a pretrained foundation model. Once the model is customized, evaluated, and has guardrails, it’s output as an NVIDIA NIM inference microservice. Developers use the NIM’s standard API to build generative AI-powered applications. Knowledge gained from applications in deployment can be fed back into the foundry to further improve custom models.
Benefits
Explore the Benefits of NVIDIA AI Foundry
Models Built to Your Needs
Start with state-of-the-art foundation models and then create custom models built specifically for your applications’ needs, using your enterprise data and domain expertise.
Faster Time to Solution
Focus on harnessing the knowledge of your organization and leveraging the game-changing insights of AI, instead of maintaining and tuning your AI development platform.
Production Ready
Protect data privacy and intellectual property by creating your own models. Run confidently in production environments with NVIDIA Enterprise Support, API stability, and reliable security updates.
Unprecedented Performance
Build models efficiently using the latest NVIDIA architecture on DGX Cloud with your preferred cloud service providers (CSPs). Deploy custom models with optimized engines, packaged as an easy-to-use NVIDIA NIM inference microservice.
Components
Build Custom Models for Generative AI Enterprise Applications
Start With Optimized Foundation Models
Generative AI Models
Enterprises can start with community and NVIDIA-built models, optimized to deliver the best performance on NVIDIA accelerated infrastructure. These models can be customized with NVIDIA NeMo and deployed in production at scale anywhere with NVIDIA NIM.
NVIDIA NeMo is an end-to-end platform for developing custom generative AI—including large language models (LLMs), multimodal, vision, and speech AI—anywhere. Deliver enterprise-ready models with precise data curation, cutting-edge customization, and optimized performance.
It includes NeMo Curator for accelerated data curation, NeMo Customizer for simplified fine-tuning, NeMo Evaluator for evaluating AI models, and NeMo Guardrails for safeguarding your LLMs.
To customize an AI model requires access to significant computing resources, AI expertise, and software to streamline AI model development.
NVIDIA DGX Cloud is an end-to-end AI platform for developers, offering dedicated capacity built on the latest NVIDIA AI architecture and co-engineered with the world’s leading CSPs. Build your custom AI models on DGX Cloud.
The output of the NVIDIA AI Foundry is an NVIDIA NIM to quickly deploy and scale the custom model anywhere.
NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI model inferencing across clouds, data centers, and workstations.
Hundreds of NVIDIA technology partners are integrating NVIDIA NIM, part of NVIDIA AI Enterprise, into their platforms to speed generative AI deployments for domain-specific applications.
Partners include the major public clouds, system builders, enterprise infrastructure providers, MLOps and AIOps leaders, and many others.
NVIDIA AI Foundry is built and backed by NVIDIA AI experts.
NVIDIA and the NVIDIA Partner Network (NPN)—including leading global systems integrators, service delivery partners, and solution providers—are ready to help you build custom models for your enterprise applications.
See how NVIDIA AI Foundry supports industry use cases and jump-starts your AI development.
AI Chatbot
Content Generation
Network Operations Centers
Intelligent Document Processing
Security Vulnerability Analysis
Hyperpersonalized Shopping
AI Chatbots
Organizations are looking to build smarter AI chatbots using custom LLMs and retrieval-augmented generation (RAG). With RAG, chatbots can accurately answer domain-specific questions by retrieving current information from an organization’s knowledge base and providing real-time responses in natural language. These chatbots can be used to enhance customer support, personalize AI avatars, manage enterprise knowledge, streamline employee onboarding, provide intelligent IT support, create content, and more.
Generative AI makes it possible to generate highly relevant, bespoke, and accurate content grounded in the domain expertise and proprietary IP of your enterprise.
Global telecommunications companies are exploring how to cost-effectively deliver new AI applications to the edge over 5G and upcoming 6G networks. With NVIDIA accelerated computing and AI, telcos, CSPs, and enterprises can build high-performance cloud-native networks—both fixed and wireless—with improved energy efficiency and security.
In financial services, processing documents involves complex data, such as loan records, external regulatory filings, transaction records, public market filings, and more. Financial institutions can use custom generative AI for IDP, such as building chatbots with RAG to automate loan processes or developing market insights for portfolio construction and trade execution.
Patching software security issues is becoming progressively more challenging as the number of reported security flaws in the common vulnerabilities and exposures (CVE) database hit a record high in 2022. Generative AI can improve vulnerability defense while decreasing the load on security teams.
A retailer’s inventory is complex, with thousands if not millions of products that change seasonally. In the immensely competitive marketplace that’s emerged over the past decade, leading retailers have harnessed the power of AI and data science to offer real-time, hyperpersonalized customer experiences that increase cart size, build brand affinity, and increase conversion.
Partners include the major public clouds, system builders, enterprise infrastructure providers, MLOps and AIOps leaders, and many others.
Model Builders
AIOps and MLOps Partners
DGX Cloud Providers
DGX Cloud Partners
Production Deployment Partners
Consulting Services Partners
Next Steps
Ready to Get Started?
Talk to an NVIDIA expert about taking custom generative AI applications from pilot to production with the security, API stability, and support of NVIDIA AI Enterprise.