NVIDIA Unveils AI Foundry And NeMo Retriever For Custom Generative AI Models Using Llama 3.1

NVIDIA Unveils AI Foundry And NeMo Retriever For Custom Generative AI Models Using Llama 3.1

Categories :

By Pallavi Singal
  • NVIDIA introduces AI Foundry to enable enterprises and nations to build custom AI ‘supermodels’ using their open data with Llama 3.1 and NVIDIA Nemotron Models. 
  • NVIDIA AI Foundry offers comprehensive generative AI Model services to deploy custom Llama 3.1 NVIDIA NIM microservices with new NVIDIA NeMo Retriever microservices for accurate responses in curation, synthetic data generation, fine-tuning, retrieval, guardrails, and evaluation. 
  • Accenture is the first to use this new service to build custom Llama 3.1 models for its clients. Aramco, AT&T, Uber and other industry leaders are among the first to access new Llama NVIDIA NIM microservices.

 

NVIDIA Unveils AI Foundry And NeMo Retriever For Custom Generative AI Models Using Llama 3.1

NVIDIA has launched the NVIDIA AI Foundry with NVIDIA NIM™ inference microservices, designed to empower enterprises and nations to develop bespoke ‘supermodels’ designed to specific industry needs using the latest Llama 3.1 collection of openly available models.

Meta’s openly available Llama 3.1 models mark a pivotal moment for the adoption of generative AI within the world’s enterprises,” said Jensen Huang, founder and CEO of NVIDIA. “Llama 3.1 opens the floodgates for every enterprise and industry to build state-of-the-art generative AI applications. NVIDIA AI Foundry has integrated Llama 3.1 throughout and is ready to help enterprises build and deploy custom Llama ‘supermodels’.”

The new Llama 3.1 models are a super-important step for open-source AI,” said Mark Zuckerberg, founder and CEO of Meta. “With NVIDIA AI Foundry, companies can easily create and customise the state-of-the-art AI services people want and deploy them with NVIDIA NIM. I’m excited to get this in people’s hands.”

NVIDIA AI Foundry: A comprehensive suite of custom generative AI models

The NVIDIA AI foundry service integrates three essential components to tailor a model for a specific dataset or company: a suite of NVIDIA AI Foundation Models, the NVIDIA NeMo framework and tools, and NVIDIA DGX Cloud AI supercomputing services. This comprehensive approach provides enterprises with an end-to-end solution for developing custom generative AI models.

NVIDIA AI Foundry facilitates enterprises and nations to create ‘supermodels’, customised for their domain-specific industry use cases using Llama 3.1 and NVIDIA software, computing and expertise. These ‘supermodels’ can be trained with proprietary data as well as synthetic data generated from Llama 3.1 405B and the NVIDIA Nemotron™ Reward model.

Image credits NVIDIA

NVIDIA DGX™ Cloud AI platform powers the NVIDIA AI Foundry. This Cloud AI platform is co-engineered with the world’s leading public clouds. This overcomes the scaling issues by giving enterprises significant computing resources.

Once custom models are built, enterprises can deploy them in production with NVIDIA NIM inference microservices. These microservices enable up to 2.5x higher throughput compared to running inference without NIM. The new NVIDIA NeMo Retriever RAG microservices further enhance response accuracy for AI applications by integrating with custom Llama 3.1 models and NIM microservices.

Global corporations supercharge AI with NVIDIA and Llama

Hundreds of NVIDIA NIM partners providing enterprise, data and infrastructure platforms can now integrate the new microservices in their AI solutions to advance generative AI capabilities for the NVIDIA community of more than 5 million developers and 19,000 startups.

Accenture is the first to utilise NVIDIA AI Foundry to develop custom Llama 3.1 models using the Accenture AI Refinery framework. This initiative aims to deploy generative AI applications that align with clients' cultural, linguistic, and industry-specific needs.

The world’s leading enterprises see how generative AI is transforming every industry and are eager to deploy applications powered by custom models,” said Julie Sweet, chair and CEO of Accenture. “Accenture has been working with NVIDIA NIM inference microservices for our internal AI applications, and now, using NVIDIA AI Foundry, we can help clients quickly create and deploy custom Llama 3.1 models to power transformative AI applications for their own business priorities.”

Image Credits NVIDIA

Companies in healthcare, energy, financial services, retail, transportation, and telecommunications are already leveraging NVIDIA NIM microservices for Llama. Among the first to access the new NIM microservices for Llama 3.1 are Aramco, AT&T and Uber.

Production support is available through NVIDIA AI Enterprise, with members of the NVIDIA Developer Program soon able to access NIM microservices for free for research, development, and testing.

NVIDIA: Leading the way to global AI adoption

NVIDIA, founded by Jensen Huang, is renowned for revolutionising visual computing and AI technologies. Specialising in GPU-accelerated computing, NVIDIA has played a pivotal role in advancing AI research and development across industries. The company's innovations span across various domains, including graphics processing units (GPUs), AI platforms, and advanced computing infrastructure.

NVIDIA’s portfolio includes the latest technologies such as the NVIDIA DGX Cloud, NVIDIA NeMo platform, and NVIDIA AI Foundry, which support enterprises in developing customised AI models and applications. The company’s GPUs, including NVIDIA Tensor Core, RTX, and GeForce RTX models, are integral to their accelerated computing solutions.

The Llama 3.1 collection of multilingual LLMs is a collection of generative AI models in 8B-, 70B- and 405B-parameter sizes. The model is trained on over 16,000 NVIDIA H100 Tensor Core GPUs and optimised for NVIDIA accelerated computing and software — in the data centre, in the cloud and locally on workstations with NVIDIA RTX™ GPUs or PCs with GeForce RTX GPUs.

Combined with NVIDIA NIM inference microservices for Llama 3.1 405B, NeMo Retriever NIM microservices deliver the highest open and commercial text Q&A retrieval accuracy for RAG pipelines.

Tags

How Cities are Embracing Sustainability for a Better Tomorrow

How Cities are Embracing Sustainability for a Better Tomorrow

Nov 21, 2024
Why You Should Hire a Professional to Install Your Roof

Why You Should Hire a Professional to Install Your Roof

Nov 20, 2024
Is a PAMM Account Right for You? 7 Factors to Consider Before Making a Decision

Is a PAMM Account Right for You? 7 Factors to Consider Before Making a Decision

Nov 19, 2024
The Ins and Outs of Indoor Positioning Systems: How They Work

The Ins and Outs of Indoor Positioning Systems: How They Work

Nov 19, 2024
4 Ways Technology is Streamlining Merchant Account Provider Operations

4 Ways Technology is Streamlining Merchant Account Provider Operations

Nov 18, 2024
How to Gather Documentation for a Seamless SR&ED Claim Process

How to Gather Documentation for a Seamless SR&ED Claim Process

Nov 17, 2024
What Every Business Needs to Know About Fuel and Lubricant Providers

What Every Business Needs to Know About Fuel and Lubricant Providers

Nov 17, 2024
What Features Make a Protective Enclosure Suitable for Heavy-Duty Applications

What Features Make a Protective Enclosure Suitable for Heavy-Duty Applications

Nov 17, 2024
Why Investing in Exterior Renovations Can Boost Your Business’s Curb Appeal

Why Investing in Exterior Renovations Can Boost Your Business’s Curb Appeal

Nov 17, 2024
A Quick Look at Heavy Equipment Essentials for Foundation and Site Work

A Quick Look at Heavy Equipment Essentials for Foundation and Site Work

Nov 17, 2024