From 7c9ed04132c9f98d4cde4d25f53c60bf9fad8591 Mon Sep 17 00:00:00 2001 From: Lucas Melo Date: Wed, 13 Nov 2024 13:42:12 -0600 Subject: [PATCH] ChatQnA - Add Terraform and Ansible Modules information (#970) Signed-off-by: chensuyue Signed-off-by: lucasmelogithub Co-authored-by: chen, suyue Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: Malini Bhandaru --- ChatQnA/README.md | 21 ++++++++++++++++++++- 1 file changed, 20 insertions(+), 1 deletion(-) diff --git a/ChatQnA/README.md b/ChatQnA/README.md index 24569cc74..e3daf6850 100644 --- a/ChatQnA/README.md +++ b/ChatQnA/README.md @@ -4,7 +4,26 @@ Chatbots are the most widely adopted use case for leveraging the powerful chat a RAG bridges the knowledge gap by dynamically fetching relevant information from external sources, ensuring that responses generated remain factual and current. The core of this architecture are vector databases, which are instrumental in enabling efficient and semantic retrieval of information. These databases store data as vectors, allowing RAG to swiftly access the most pertinent documents or data points based on semantic similarity. -## Deploy ChatQnA Service +## šŸ¤– Automated Terraform Deployment using IntelĀ® Optimized Cloud Modules for **Terraform** + +| Cloud Provider | Intel Architecture | Intel Optimized Cloud Module for Terraform | Comments | +| -------------------- | --------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------- | -------------------------------------------------------------------- | +| AWS | 4th Gen Intel Xeon with Intel AMX | [AWS Module](https://github.com/intel/terraform-intel-aws-vm/tree/main/examples/gen-ai-xeon-opea-chatqna) | Uses Intel/neural-chat-7b-v3-3 by default | +| AWS Falcon2-11B | 4th Gen Intel Xeon with Intel AMX | [AWS Module with Falcon11B](https://github.com/intel/terraform-intel-aws-vm/tree/main/examples/gen-ai-xeon-opea-chatqna-falcon11B) | Uses TII Falcon2-11B LLM Model | +| GCP | 5th Gen Intel Xeon with Intel AMX | [GCP Module](https://github.com/intel/terraform-intel-gcp-vm/tree/main/examples/gen-ai-xeon-opea-chatqna) | Also supports Confidential AI by using IntelĀ® TDX with 4th Gen Xeon | +| Azure | 5th Gen Intel Xeon with Intel AMX | Work-in-progress | Work-in-progress | +| Intel Tiber AI Cloud | 5th Gen Intel Xeon with Intel AMX | Work-in-progress | Work-in-progress | + +## Automated Deployment to Ubuntu based system(if not using Terraform) using IntelĀ® Optimized Cloud Modules for **Ansible** + +To deploy to existing Xeon Ubuntu based system, use our Intel Optimized Cloud Modules for Ansible. This is the same Ansible playbook used by Terraform. +Use this if you are not using Terraform and have provisioned your system with another tool or manually including bare metal. +| Operating System | Intel Optimized Cloud Module for Ansible | +|------------------|------------------------------------------| +| Ubuntu 20.04 | [ChatQnA Ansible Module](https://github.com/intel/optimized-cloud-recipes/tree/main/recipes/ai-opea-chatqna-xeon) | +| Ubuntu 22.04 | Work-in-progress | + +## Manually Deploy ChatQnA Service The ChatQnA service can be effortlessly deployed on Intel Gaudi2, Intel Xeon Scalable Processors and Nvidia GPU.