This immersive 3-day workshop is designed to equip students with the skills to deploy and manage large language models (LLMs) in production environments. From understanding cloud-based infrastructure to leveraging distributed training and inference, students will gain hands-on experience in optimizing and scaling LLMs for real-world applications.
Introduce the foundational AI infrastructure components required to manage LLMs in production.
Explore how to leverage GPUs and other hardware accelerators for LLM training and inference in production environments.
Dive into distributed training techniques for scaling LLM training across multiple nodes.
Learn how to efficiently run LLMs at scale using distributed inference techniques to handle high-volume requests.
Online
9:00am - 5:00pm ET
Online
9:00am - 5:00pm ET
Online
9:00am - 5:00pm ET
WeCloudData is the leading data science and AI academy. Our blended learning courses have helped thousands of learners and many enterprises make successful leaps in their data journeys.
"*" indicates required fields
Canada:
180 Bloor St W #1003
Toronto, ON, Canada M5S 2V6
US:
16192 Coastal Hwy
Lewes, DE 19958, USA