2025-01-07

Quickly build and deploy LLMs with Retrieval Augmented Generation, featuring Intel Accelerators and Azure

Join our tutorial on January, 28

Register now

About the webinar

Retrieval-augmented generation (RAG) offers a new way to maximize the capabilities of large language models (LLMs) to produce more accurate, context-aware, and informative responses. Join Akash Shankaran (Intel), Ron Abellera (Microsoft), and Juan Pablo Norena (Canonical) for a tutorial on how RAG can enhance your LLMs.

  • The session will explore how to optimize LLMs with RAG using Charmed OpenSearch, which can serve multiple services like data ingestion, model ingestion, vector database, retrieval and ranking, and LLM connector.

  • We will also show the architecture of our RAG deployment in Microsoft® Azure Cloud. Notably, the vector search capabilities of RAG are enhanced by Intel AVX® Acceleration, delivering faster processing and high-throughput performance for the RAG workflow.

Learn more: