Tutorial
Welcome to the "Big Model" Era: Techniques and Systems to Train and Serve Bigger Models
Hao Zhang · Lianmin Zheng · Zhuohan Li · Ion Stoica
Moderator : Hongyi Wang
Hall F
In recent years, researchers in ML and systems have been working together to bring big models -- such as GPT-3 with 175B parameters -- into research and production. It has been revealed that increasing model sizes can significantly boost ML performance, and even lead to fundamentally new capabilities.
However, experimenting and adopting big models call for new techniques and systems to support their training and inference on big data and large clusters. This tutorial identifies research and practical pain points in model-parallel training and serving. In particular, this tutorial introduces new algorithmic techniques and system architectures for addressing the training and serving of popular big models, such as GPT-3, PaLM, and vision transformers. The tutorial also consists of a session on how to use the latest open-source system toolsets to support the training and serving of big models. Through this tutorial, we hope to lower the technical barrier of using big models in ML research and bring the big models to the masses.
Schedule
Mon 12:30 p.m. - 12:35 p.m.
|
Opening Remarks
(
Talk
)
>
SlidesLive Video |
Hao Zhang 🔗 |
Mon 12:35 p.m. - 12:50 p.m.
|
Trends Driving Big Models ( Talk ) > link | Ion Stoica 🔗 |
Mon 12:50 p.m. - 1:15 p.m.
|
New Views of ML parallelism: Intra- and Inter-Operator Parallelism
(
Talk
)
>
SlidesLive Video |
Hao Zhang 🔗 |
Mon 1:15 p.m. - 1:45 p.m.
|
Inter-Operator Parallelism
(
Talk
)
>
SlidesLive Video |
Zhuohan Li 🔗 |
Mon 1:45 p.m. - 1:55 p.m.
|
Break and Q&A
SlidesLive Video |
🔗 |
Mon 1:55 p.m. - 2:25 p.m.
|
Intra-Operator Parallelism
(
Talk
)
>
SlidesLive Video |
Lianmin Zheng 🔗 |
Mon 2:25 p.m. - 2:45 p.m.
|
Auto Parallelization of ML Computation
(
Talk
)
>
SlidesLive Video |
Hao Zhang 🔗 |
Mon 2:45 p.m. - 2:50 p.m.
|
Tools for Big Model, Key Takeaways, and Q&A
(
Talk
)
>
SlidesLive Video |
Lianmin Zheng 🔗 |