Timezone: »

ES-FoMo: Efficient Systems for Foundation Models
Julien Launay · Daniel Y Fu · Tri Dao · Daniel Hesslow · Beidi Chen · Azalia Mirhoseini · Percy Liang

Sat Jul 29 11:55 AM -- 12:00 AM (PDT) @ Ballroom A
Event URL: http://gpusgobrrr.com/ »

As models increase in size and training budget, they not only systematically improve in upstream quality, but also exhibit novel emergent capabilities. This increase in scale raises proportionate difficulties for practitioners: foundation model training and inference lie at a unique interdisciplinary crossroad, combining open problems in algorithms, system design, and software engineering.

Machine learning practitioners are key stakeholders here: on the one hand, researchers may contribute algorithmic insights and novel methods to improving training and inference of large models; on the other hand, novel research findings may be best demonstrated at scale—which may require training models as efficiently as possible to make the best use of available resources.

The goal of this workshop is to bring together interdisciplinary experts working on the emerging research questions and challenges associated with foundation model training and inference. We welcome submissions around training and inference systems/algorithms for foundation models, focusing on scaling-up or on reducing compute, time, memory, bandwidth, and energy requirements. Notably, we encourage submissions concerning the entire spectrum of foundation models: from BERT-sized Transformers, to large models with 100B+ parameters. Topics include but are not limited to:

* Training and inference systems, either distributed at large scale or in resource-constrained scenarios;
* Algorithms for improved training and inference efficiency;
* Systems for foundation models, such as novel programming languages or compilers.

Author Information

Julien Launay (HuggingFace)
Daniel Y Fu (Stanford University)
Tri Dao (Stanford)
Daniel Hesslow (Huggingface)
Beidi Chen (CMU / FAIR)
Azalia Mirhoseini
Percy Liang (Stanford University)

More from the Same Authors