Course Overview
This course is designed for cloud solutions architects who want to gain a deeper understanding of Artificial Intelligence (AI) in the cloud. The course covers the basics of AI in the cloud, including the different classes of training and inference, and the distinct price/performance tradeoffs. Students will learn about the four main types of AI inference, including server application inference, client application inference, batch or streaming inference, and edge inference. The course also explores how Intel hardware can make a difference in AI workloads, including the use of Intel Xeon scalable processors and Intel accelerators. Through a combination of video lessons, demos, and hands-on labs, students will gain practical experience with AI pipeline, benchmarking, instance selection, and federated learning. By the end of the course, students will be able to identify which parts of the AI pipeline run best in the cloud, benchmark instances for best AI performance, and run AI workloads in the cloud using Intel tools and optimizations.