Learn to build LLM-powered apps with Python and learn to scale and deploy them with Ray using open-source models.
In this workshop, you'll learn the why as well as the how, as you gain hands-on experience implementing common designs for LLM systems including the retrieval-augmented generation ("RAG") patterns that power the most successful AI apps.
We'll cover all of the fundamentals as well as demystify the more advanced operational elements like dealing with large datasets (think: an archive of customer service interactions or insurance claims), autoscaling on GPUs, and more.