Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.

License

NotificationsYou must be signed in to change notification settings

opea-project/GenAIExamples

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Introduction

GenAIExamples are designed to give developers an easy entry into generative AI, featuring microservice-based samples that simplify the processes of deploying, testing, and scaling GenAI applications. All examples are fully compatible with both Docker and Kubernetes, supporting a wide range of hardware platforms such as Gaudi, Xeon, AMD EPYC CPUs, AMD Instinct GPUs, and other hardwares including NVIDIA GPUs, ensuring flexibility and efficiency for your GenAI adoption.

Architecture

GenAIComps is a service-based tool that includes microservice components such as llm, embedding, reranking, and so on. Using these components, various examples in GenAIExample can be constructed including ChatQnA, DocSum, etc.

GenAIInfra is part of the OPEA containerization and cloud-native suite and enables quick and efficient deployment of GenAIExamples in the cloud.

GenAIEval measures service performance metrics such as throughput, latency, and accuracy for GenAIExamples. This feature helps users compare performance across various hardware configurations easily.

Use Cases

Below are some highlighted GenAI use cases across various application scenarios:

ScenarioUse Case
Question AnsweringChatQnA ✨: Chatbot with Retrieval Augmented Generation (RAG).
VisualQnA ✨: Visual Question-answering.
Image GenerationText2Image ✨: Text-to-image generation.
Content SummarizationDocSum: Document Summarization Application.
Code GenerationCodeGen: Gen-AI Powered Code Generator.
Information RetrievalDocIndexRetriever: Document Retrieval with Retrieval Augmented Generation (RAG).
Fine-tuningInstructionTuning: Application of Instruction Tuning.

For the full list of the available use cases and their supported deployment type, please referhere.

Documentation

The GenAIExamplesdocumentation contains a comprehensive guide on all available examples including architecture, deployment guides, and more. Information on GenAIComps, GenAIInfra, and GenAIEval can also be found there.

Getting Started

GenAIExamples offers flexible deployment options that cater to different user needs, enabling efficient use and deployment in various environments. Three primary methods are presently used to do this: Python startup, Docker Compose, and Kubernetes.

Users can choose the most suitable approach based on ease of setup, scalability needs, and the environment in which they are operating.

Deployment Guide

Deployment is based on released docker images by default - checkdocker image list for detailed information. You can also build your own images following instructions.

Prerequisite

  • For Docker Compose-based deployment, you should have docker compose installed. Refer todocker compose install for more information.

  • For Kubernetes-based deployment, you can useHelm orGMC-based deployment.

    • You should have a kubernetes cluster ready for use. If not, you can refer tok8s install to deploy one.
    • (Optional) You should have Helm (version >= 3.15) installed if you want to deploy with Helm Charts. Refer to theHelm Installation Guide for more information.
    • (Optional) You should have GMC installed to your kubernetes cluster if you want to try with GMC. Refer toGMC install for more information.
  • Recommended Hardware Reference

    Based on different deployment model sizes and performance requirements, you may choose different hardware platforms or cloud instances. Here are some of the reference platforms:

Use CaseDeployment modelReference ConfigurationHardware access/instances
XeonIntel/neural-chat-7b-v3-364 vCPUs, 365 GB disk, 100 GB RAM, and Ubuntu 24.04Intel Tiber Developer Cloud
GaudiIntel/neural-chat-7b-v3-31 or 2 Gaudi Cards, 16 vCPUs, 365 GB disk, 100 GB RAM, and Ubuntu 24.04Intel Tiber Developer Cloud
Xeon (AWS)Intel/neural-chat-7b-v3-364 vCPUs, 100 GB disk, 64 GB RAM, and Ubuntu 24.04AWS Cloud (e.g.,c7i.16xlarge)
AMD EPYCmeta-llama/Meta-Llama-3-8B-Instruct64 vCPUs, 100 GB disk, 256 GB RAM, and Ubuntu 24.04Google Cloud Platform
Microsoft Azure
AWS
AMD Instinctmeta-llama/Llama-3.1-405BGPU: 8× MI300X, 1536 GB vRAM, and Ubuntu 24.04AMD Developer Cloud
Oracle Cloud Infrastructure
Microsoft Azure

Deploy Examples

Note: Check forsample guides first for your use case. If it is not available, then refer to the table below:

Use CaseDocker Compose Deployment on XeonDocker Compose Deployment on GaudiDocker Compose Deployment on AMD EPYCDocker Compose Deployment on ROCmKubernetes with Helm ChartsKubernetes with GMC
ChatQnAXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsChatQnA with Helm ChartsChatQnA with GMC
CodeGenXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsCodeGen with Helm ChartsCodeGen with GMC
CodeTransXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsCodeTrans with Helm ChartsCodeTrans with GMC
DocSumXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsDocSum with Helm ChartsDocSum with GMC
SearchQnAXeon InstructionsGaudi InstructionsEPYC InstructionsNot SupportedSearchQnA with Helm ChartsSearchQnA with GMC
TranslationXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsNot SupportedTranslation with GMC
AudioQnAXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsAudioQnA with Helm ChartsAudioQnA with GMC
VisualQnAXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsVisualQnA with Helm ChartsVisualQnA with GMC
MultimodalQnAXeon InstructionsGaudi InstructionsEPYC InstructionsROCm InstructionsNot SupportedNot Supported
ProductivitySuiteXeon InstructionsNot SupportedEPYC InstructionsNot SupportedNot SupportedNot Supported
Text2ImageXeon InstructionsGaudi InstructionsEPYC InstructionsNot SupportedText2Image with Helm ChartsNot Supported

Supported Examples

Checkhere for detailed information of supported examples, models, hardwares, etc.

Validated Configurations

Checkhere for the validated configurations of GenAIExamples, including hardware and software versions that have been tested for each release.

Contributing to OPEA

Welcome to the OPEA open-source community! We are thrilled to have you here and excited about the potential contributions you can bring to the OPEA platform. Whether you are fixing bugs, adding new GenAI components, improving documentation, or sharing your unique use cases, your contributions are invaluable.

Together, we can make OPEA the go-to platform for enterprise AI solutions. Let's work together to push the boundaries of what's possible and create a future where AI is accessible, efficient, and impactful for everyone.

Please check theContributing guidelines for a detailed guide on how to contribute a GenAI component and all the ways you can contribute!

Thank you for being a part of this journey. We can't wait to see what we can achieve together!

Additional Content

About

Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors128


[8]ページ先頭

©2009-2025 Movatter.jp