Course curriculum

Today, AI applications are becoming pervasive across all sectors of our industry. Driven by a few fundamental trends, there is no indication of slowing down. In fact, the trend continues rapidly, making distributed computing at scale a norm and necessity. But distributed computing is not easy. It has its challenges. Building distributed applications today requires tons of expertise. For many developers, it is out of reach. Current solutions to these challenges have their shortcomings and tradeoff. Ray aims to address these shortcomings. As a general-purpose distributed computing framework, it makes programming a cluster of machines as easy as programming a laptop, thereby enabling many more developers and practitioners to take advantage of the advances in cloud computing and scale their machine learning workloads to solve harder problems, without needing to be experts in distributed systems. Besides a core general-purpose distributed-compute system, Ray encompasses a collection of state-of-the-art native libraries targeting scalable machine learning. These include libraries for hyper parameter tuning, distributed training, reinforcement learning, model serving, and last-mile ML data pre-processing and ingestion for model training. This talk will introduce Ray’s overview survey its ecosystem of both native and integrated ML libraries and discuss key applications and developments in the Ray ecosystem, drawing upon lessons from discussions with practitioners over the years of developing Ray with the community—and at Any Scale. In particular, we will demonstrate how you can easily scale three common ML workloads, from your laptop to the cluster, with Ray’s native libraries training, hyper parameter tuning and optimisation (HPO), and large-scale batch inference. Using the popular XGBoost for classification, we will show how you can scale model training, hyper parameter tuning, and inference—from a laptop or single node to a Ray cluster, with tangible performance difference when using Ray. The takeaways from this talk are - Why distributed computing has become the norm and necessity, not an exception Learn Ray’s architecture, core concepts, and programming primitives.Understand Ray’s ecosystem of scalable ML libraries Easily extend or transition your laptop to a Ray cluster Scale three ML workloads using Ray’s native libraries - Training on a single node vs. Ray cluster, using XGBoost with/without Ray Tuning HPO using XGBoost with Ray and Ray Tune Inferencing at scale, using XGBoost with/without Ray

  • 1

    Scaling AI Workloads with the Ray Ecosystem

    • Scaling AI Workloads with the Ray Ecosystem

Instructor

CEO and CoFounder Anyscale

Robert Nishihara

Robert Nishihara is one of the creators of Ray, a distributed system for scaling Python and machine learning applications. He is one of the co-founders and CEO of Anyscale, which is the company behind Ray. He did his PhD in machine learning and distributed systems in the computer science department at UC Berkeley. Before that, he majored in math at Harvard.

ODSC APAC Virtual

Don't miss a chance to be among FIRST 100 to register for APAC 2022