- Notifications
You must be signed in to change notification settings - Fork0
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
License
pimlock/GenerativeAIExamples
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
This repository is a starting point for developers looking to integrate with the NVIDIA software ecosystem to speed up their generative AI systems. Whether you are building RAG pipelines, agentic workflows, or fine-tuning models, this repository will help you integrate NVIDIA, seamlessly and natively, with your development stack.
This tutorial demonstrates an end-to-end Data Flywheel implementation that uses NVIDIA NeMo Microservices. It features a tool-calling workflow with the NVIDIA NeMo Datastore, NeMo Entity Store, NeMo Customizer, NeMo Evaluator, NeMo Guardrails microservices, and NVIDIA NIMs.
This example implements a GPU-accelerated pipeline for creating and querying knowledge graphs using RAG by leveraging NIM microservices and the RAPIDS ecosystem to process large-scale datasets efficiently.
- Build an Agentic RAG Pipeline with Llama 3.1 and NVIDIA NeMo Retriever NIM microservices [Blog,Notebook]
- NVIDIA Morpheus, NIM microservices, and RAG pipelines integrated to create LLM-based agent pipelines
- Tips for Building a RAG Pipeline with NVIDIA AI LangChain AI Endpoints by Amit Bleiweiss. [Blog,Notebook]
For more information, refer to theGenerative AI Example releases.
A collection of Jupyter notebooks, sample code and reference applications built with Vision NIMs.
To pull the vision NIM workflows, clone this repository recursively:
git clone https://github.com/nvidia/GenerativeAIExamples --recurse-submodulesThe workflows will then be located atGenerativeAIExamples/vision_workflows
Follow the links below to learn more:
- Learn how to use VLMs to automatically monitor a video stream for custom events.
- Learn how to search images with natural language using NV-CLIP.
- Learn how to combine VLMs, LLMs and CV models to build a robust text extraction pipeline.
- Learn how to use embeddings with NVDINOv2 and a Milvus VectorDB to build a few shot classification model.
Experience NVIDIA RAG Pipelines with just a few steps!
Get your NVIDIA API key.
- Go to theNVIDIA API Catalog.
- Select any model.
- ClickGet API Key.
- Run:
export NVIDIA_API_KEY=nvapi-...
Clone the repository.
git clone https://github.com/nvidia/GenerativeAIExamples.gitBuild and run the basic RAG pipeline.
cd GenerativeAIExamples/RAG/examples/basic_rag/langchain/docker compose up -d --build
Go tohttps://localhost:8090/ and submit queries to the sample RAG Playground.
Stop containers when done.
docker compose down
AData Flywheel is a self-reinforcing cycle where user interactions generate data that improves AI models or products, leading to better outcomes that attract more users and further enhance data quality. This feedback loop relies on continuous data processing, model refinement, and guardrails to ensure accuracy and compliance while compounding value over time. Real-world applications range from personalized customer experiences to operational systems like inventory management, where improved predictions drive efficiency and growth.
Tool calling empowers Large Language Models (LLMs) to integrate with external APIs, execute dynamic workflows, and retrieve real-time data beyond their training scope. The NVIDIA NeMo microservices platform offers a modular infrastructure for deploying AI pipelines that includes fine-tuning, evaluation, inference, and guardrail enforcement—across Kubernetes clusters in cloud or on-premises environments.
This end-to-endtutorial demonstrates how to leverage NeMo Microservices to customizeLlama-3.2-1B-Instruct by using thexLAM function-calling dataset, assess its accuracy, and implement safety constraints to govern its behavior.
NVIDIA has first-class support for popular generative AI developer frameworks likeLangChain,LlamaIndex, andHaystack. These end-to-end notebooks show how to integrate NIM microservices using your preferred generative AI development framework.
Use thesenotebooks to learn about the LangChain and LlamaIndex connectors.
- RAG
- Agents
By default, these end-to-endexamples use preview NIM endpoints onNVIDIA API Catalog. Alternatively, you can run any of the exampleson premises.
Example tools and tutorials to enhance LLM development and productivity when using NVIDIA RAG pipelines.
- NVIDIA Tokkio LLM-RAG: Use Tokkio to add avatar animation for RAG responses.
- Hybrid RAG Project on AI Workbench: Run an NVIDIA AI Workbench example project for RAG.
- Changing the Inference or Embedded Model
- Customizing the Vector Database
- Customizing the Chain Server:
- Configuring LLM Parameters at Runtime
- Supporting Multi-Turn Conversations
- Speaking Queries and Listening to Responses with NVIDIA Riva
- Support Matrix
- Architecture
- Using the Sample Chat Web Application
- RAG Playground Web Application
- Software Component Configuration
We're posting these examples on GitHub to support the NVIDIA LLM community and facilitate feedback.We invite contributions! Open a GitHub issue or pull request! Seecontributing Check out thecommunity examples and notebooks.
About
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Languages
- Jupyter Notebook74.4%
- Python22.9%
- JavaScript1.0%
- HTML0.4%
- Shell0.4%
- CSS0.3%
- Other0.6%
