vLLM Server Using OpenAI API on Gaudi 3 | AI with Guy
In this episode of AI with Guy, learn how to build a real-world retrieval-augmented generation (RAG) system using vLLM, the OpenAI API, and Intel Gaudi 3 across multiple AWS instances. This demo shows how to deploy a scalable, production-ready Gen-AI application using the OPEA framework.
Whether you're a developer, ML engineer, or AI architect, this walkthrough covers:
Setting up a vLLM server
Connecting to the OpenAI API for inference
Deploying across AWS EC2 using Intel Gaudi 3
Coordinating workloads using OPEA for high performance and cost efficiency
Resources:
OPEA Documentation: https://opea.dev/
ChatQnA Gen-AI Example: https://opea-project.github.io/latest/GenAIExamples/ChatQnA/README.html
Intel Cloud Dev Tools: https://cloud.intel.com/
Tech Stack:
-vLLM
-OpenAI API
-Intel Gaudi 3 (AWS DL1/DL2 instances)
OPEA
AWS EC2
About Intel Software:
Intel® Developer Zone is committed to empowering and assisting software developers in creating applications for Intel hardware and software products. The Intel Software YouTube channel is an excellent resource for those seeking to enhance their knowledge. Our channel provides the latest news, helpful tips, and engaging product demos from Intel and our numerous industry partners. Our videos cover various topics; you can explore them further by following the links.
Connect with Intel Software:
INTEL SOFTWARE WEBSITE:https://intel.ly/2KeP1hDD
INTEL SOFTWARE on FACEBOOK:http://bit.ly/2z8MPFFF
INTEL SOFTWARE on TWITTER:http://bit.ly/2zahGSnn
INTEL SOFTWARE GITHUB:http://bit.ly/2zaih6zz
INTEL DEVELOPER ZONE LINKEDIN:http://bit.ly/2z979qss
INTEL DEVELOPER ZONE INSTAGRAM:http://bit.ly/2z9Xsbyy
INTEL GAME DEV TWITCH:http://bit.ly/2BkNshuu
#intelsoftware
vLLM Server Using OpenAI API on Gaudi 3 | AI with Guy | Intel Software