Skip to content

Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

License

Apache-2.0, Unknown licenses found

Licenses found

Apache-2.0
LICENSE.md
Unknown
LICENSE.DATA
Notifications You must be signed in to change notification settings

NVIDIA/GenerativeAIExamples

Repository files navigation

NVIDIA Generative AI Examples

This repository is a starting point for developers looking to integrate with the NVIDIA software ecosystem to speed up their generative AI systems. Whether you are building RAG pipelines, agentic workflows, or fine-tuning models, this repository will help you integrate NVIDIA, seamlessly and natively, with your development stack.

Table of Contents

What's New?

Knowledge Graph RAG

This example implements a GPU-accelerated pipeline for creating and querying knowledge graphs using RAG by leveraging NIM microservices and the RAPIDS ecosystem to process large-scale datasets efficiently.

Agentic Workflows with Llama 3.1

RAG with Local NIM Deployment and LangChain

  • Tips for Building a RAG Pipeline with NVIDIA AI LangChain AI Endpoints by Amit Bleiweiss. [Blog, Notebook]

For more information, refer to the Generative AI Example releases.

Vision NIM Workflows

A collection of Jupyter notebooks, sample code and reference applications built with Vision NIMs.

To pull the vision NIM workflows, clone this repository recursively:

git clone https://github.com/nvidia/GenerativeAIExamples --recurse-submodules

The workflows will then be located at GenerativeAIExamples/vision_workflows

Follow the links below to learn more:

Try it Now!

Experience NVIDIA RAG Pipelines with just a few steps!

  1. Get your NVIDIA API key.

    1. Go to the NVIDIA API Catalog.
    2. Select any model.
    3. Click Get API Key.
    4. Run:
      export NVIDIA_API_KEY=nvapi-...
  2. Clone the repository.

    git clone https://github.com/nvidia/GenerativeAIExamples.git
  3. Build and run the basic RAG pipeline.

    cd GenerativeAIExamples/RAG/examples/basic_rag/langchain/
    docker compose up -d --build
  4. Go to https://localhost:8090/ and submit queries to the sample RAG Playground.

  5. Stop containers when done.

    docker compose down

RAG

RAG Notebooks

NVIDIA has first-class support for popular generative AI developer frameworks like LangChain, LlamaIndex, and Haystack. These end-to-end notebooks show how to integrate NIM microservices using your preferred generative AI development framework.

Use these notebooks to learn about the LangChain and LlamaIndex connectors.

LangChain Notebooks

LlamaIndex Notebooks

RAG Examples

By default, these end-to-end examples use preview NIM endpoints on NVIDIA API Catalog. Alternatively, you can run any of the examples on premises.

Basic RAG Examples

Advanced RAG Examples

RAG Tools

Example tools and tutorials to enhance LLM development and productivity when using NVIDIA RAG pipelines.

RAG Projects

Documentation

Getting Started

How To's

Reference

Community

We're posting these examples on GitHub to support the NVIDIA LLM community and facilitate feedback. We invite contributions! Open a GitHub issue or pull request! See contributing Check out the community examples and notebooks.