Timezone: »
Until recently, many industrial Machine Learning applications have been the remit of consulting academics, data scientists within larger companies, and a number of dedicated Machine Learning research labs within a few of the world’s most innovative tech companies. Over the last few years we have seen the dramatic rise of companies dedicated to providing Machine Learning software-as-a-service tools, with the aim of democratizing access to the benefits of Machine Learning. All these efforts have revealed major hurdles to ensuring the continual delivery of good performance from deployed Machine Learning systems. These hurdles range from challenges in MLOps, to fundamental problems with deploying certain algorithms, to solving the legal issues surrounding the ethics involved in letting algorithms make decisions for your business.
This workshop will invite papers related to the challenges in deploying and monitoring ML systems. It will encourage submission on subjects related to: MLOps for deployed ML systems; the ethics around deploying ML systems; useful tools and programming languages for deploying ML systems; specific challenges relating to deploying reinforcement learning in ML systems and performing continual learning and providing continual delivery in ML systems;
and finally data challenges for deployed ML systems.
We will also invite the submission of open problems and encourage the discussion (through two live panels) on topics related to the areas of: "Deploying machine learning applications in the legal system" and "Deploying machine learning on devices or constrained hardware".
These subjects represent a wealth of topical and high-impact issues for the community to work on.
Fri 2:00 a.m. - 2:10 a.m.
|
Opening remarks
(
Introduction
)
SlidesLive Video » |
Alessandra Tosi · Nathan Korda · Fariba Yousefi · Andrei Paleyes · Stephen Roberts 🔗 |
Fri 2:10 a.m. - 2:50 a.m.
|
Deploying end-to-end machine learning systems for social impact
(
Invited Talk
)
SlidesLive Video » Speaker: Engineer Bainomugisha Bio: I am an Associate Professor of Computer Science and the Chair of the Department of Computer Science at Makerere University. My research focuses on Computer Science-driven solutions to the prevailing world challenges. I am also passionate about contributing to quality Computer Science education that is of sufficient breath and depth, practical and fast enough. Currently, I lead several innovative and research initiatives that aim to create and apply computational methods and tools that can improve the quality of life especially in the developing world setting. https://ibaino.net/ |
Engineer Bainomugisha 🔗 |
Fri 2:50 a.m. - 3:30 a.m.
|
Machine Learning and Legal Decisions
(
Invited Talk
)
SlidesLive Video » Can machine learning help to reduce costs and facilitate access to justice within the legal system? We first consider constitutional constraints to which deployment of ML in the legal system are subject; in particular, the protection of fundamental rights and the need to give reasons in legal decisions. We then turn to the technical state of the art with ML analysis of caselaw decisions. It is possible to predict outcomes of cases, given a set of facts, with a high degree of accuracy, but the explainability of these methods is limited. The research frontier therefore explores ways to provide legal reasons for case predictions. Speaker: John Armour is Professor of Law and Finance at Oxford University and a Fellow of the British Academy and the European Corporate Governance Institute. https://www.law.ox.ac.uk/people/john-armour |
John Armour 🔗 |
Fri 3:30 a.m. - 4:30 a.m.
|
Applications in the legal system
(
Panel Discussion
)
SlidesLive Video » Chair: JOHN ARMOUR, University of Oxford Panelists - Jessica Montgomery, University of Cambridge. "ML for policy: deploying machine learning to tackle public policy challenges " - Teresa Scantamburlo, European Centre for Living Technology. "The laborious exercise of human oversight in workforce surveillance " - Charles Brecque, Legislate.tech. "Can machine learning ever take the lawyer out of the loop completely? Can machine learning make the legal system fairer, make contracts less confusing and give legal advice?" |
Jessica Montgomery · Charles Brecque · Teresa Scantamburlo 🔗 |
Fri 4:30 a.m. - 4:40 a.m.
|
Short Break
|
🔗 |
Fri 4:40 a.m. - 5:30 a.m.
|
Machine Learning for Chip Design
(
Invited Talk
)
SlidesLive Video » The automated design of chips is facing growing challenges due to a high volume of smartphones, the increasing functionality, and the corresponding heterogeneity of the chips. In this talk, I will survey how machine learning has recently emerged as a core technique that promises to rescue the reducing gains in power performance and area in this field. In particular, I will focus on the challenges in deploying learning algorithms in electronic design automation and outline the solution that we take at Qualcomm which combines machine learning with combinatorial optimization solvers. Speaker: Roberto Bondesan, Qualcomm https://scholar.google.com/citations?user=l2z7p3oAAAAJ&hl=en |
Roberto Bondesan 🔗 |
Fri 5:30 a.m. - 6:15 a.m.
|
Machine Learning and Legal Decisions
(
Invited Talk
)
SlidesLive Video » Speaker's bio: Professor Richard Susskind OBE is an author, speaker, and independent adviser to major professional firms and to national governments. His main area of expertise is the future of professional service and, in particular, the way in which the IT and the Internet are changing the work of lawyers. He has worked on legal technology for over 30 years. He lectures internationally, has written many books, and advised on numerous government inquiries. https://www.susskind.com/ |
Richard Susskind 🔗 |
Fri 6:15 a.m. - 6:30 a.m.
|
Short Break
|
🔗 |
Fri 6:30 a.m. - 6:39 a.m.
|
Who is Responsible for Adversarial Defense?
(
Contributed Talk
)
SlidesLive Video » We have seen a surge in research aims toward adversarial attacks and defenses in AI/ML systems. While it is crucial to formulate new attack methods and devise novel defense strategies for robustness, it is also imperative to recognize who is responsible for implementing, validating and justifying the necessity of these defenses. In particular, which components of the system are vulnerable to what type of adversarial attacks, and the expertise needed to realize the severity of adversarial attacks. Also how to evaluate and address the adversarial challenges in order to recommend defense strategies for different applications. This paper opened a discussion on who should examine and implement the adversarial defenses and the reason behind such efforts. Authors: Kishor Datta Gupta ( University of Memphis ) Dipankar Dasgupta ( University of Memphis ) |
Kishor Datta Gupta 🔗 |
Fri 6:39 a.m. - 6:50 a.m.
|
Towards Efficient Machine Unlearning via Incremental View Maintenance
(
Contributed Talk
)
SlidesLive Video » Recent laws such as the GDPR require machine learning applications to "unlearn" parts of their training data if a user withdraws consent for their data. Current unlearning approaches accelerate the retraining of models, but come with hidden costs due to the need to reaccess training data and redeploy the resulting models. We propose to look at machine unlearning as an "incremental view maintenance" problem, leveraging existing research from the data management community on efficiently maintaining the results of a query in response to changes in its inputs. Our core idea is to consider ML models as views over their training data, and express the training procedure as a differential dataflow computation, whose outputs can be automatically updated. As a consequence, the resulting models can be continuously trained over streams of updates and deletions. We discuss important limitations of this approach, and provide preliminary experimental results for maintaining a state-of-the-art sequential recommendation model. Authors: Sebastian Schelter ( University of Amsterdam ) |
Sebastian Schelter 🔗 |
Fri 6:50 a.m. - 7:00 a.m.
|
DuckDQ: Data Quality Assertions for Machine Learning Pipelines
(
Contributed Talk
)
SlidesLive Video » Data quality validation plays an important role in ensuring the proper behaviour of productive machine learning (ML) applications and services. Observing a lack of existing solutions for quality control in medium-sized production systems, we developed DuckDQ: A lightweight and efficient Python library for data quality validation, that seamlessly integrates with existing scikit-learn ML pipelines and does not require a distributed computing environment or ML platform infrastructure, while outperforming existing solutions by a factor 3 to 40 in terms of runtime. We introduce the notion of data quality assertions, which can stop a pipeline when quality constraints of the input data or the model's output are not met. Furthermore, we employ stateful metric computations, which greatly enhance the possibilities for post-hoc failure analysis and drift detection, even when the serving data is not around anymore. Authors Till Doehmen ( Fraunhofer FIT ) Mark Raasveldt ( CWI ) Hannes Mühleisen ( Centrum Wiskunde & Informatica ) Sebastian Schelter ( University of Amsterdam ) |
Till Döhmen 🔗 |
Fri 7:00 a.m. - 7:10 a.m.
|
MLDemon: Deployment Monitoring for Machine Learning Systems
(
Contributed Talk
)
SlidesLive Video » Post-deployment monitoring of the performance ML systems is critical for ensuring reliability, especially as new user inputs can differ from the training distribution. Here we propose a novel approach, MLDemon, for ML DEployment MONitoring. MLDemon integrates both unlabeled features and a small amount of labeled examples which arrive over time to produce a real-time estimate of the ML model's current performance. Subject to budget constraints, MLDemon decides when to acquire additional, potentially costly, labels to verify the model. On temporal datasets with diverse distribution drifts and models, MLDemon substantially outperforms existing monitoring approaches. Moreover, we provide theoretical analysis to show that MLDemon is minimax rate optimal up to logarithmic factors and is provably robust against broad distribution drifts whereas prior approaches are not. Authors Tony Ginart ( Stanford University ) Martin Zhang ( Harvard School of Public Health ) James Zou ( Stanford University ) |
Tony Ginart 🔗 |
Fri 7:10 a.m. - 7:20 a.m.
|
Have the Cake and Eat It Too? Higher Accuracy and Less Expense when Using Multi-label ML APIs Online
(
Contributed Talk
)
SlidesLive Video » The fast-growing machine learning as a service industry has incubated many APIs for multi-label classification tasks such as OCR and multi-object recognition. The heterogeneity in those APIs' price and performance, however, often forces users to choose between accuracy and expense. In this work, we propose FrugalMCT, a principled framework that jointly maximizes the accuracy while minimizes the expense by adaptively selecting the APIs to use for different data. FrugalMCT combines different APIs' predictions to improve accuracy and selects which combination to use to respect expense constraints. Preliminary experiments using ML APIs from Google, Microsoft, and other providers for multi-label image classification show that FrugalMCT often achieves more than 50% cost reduction while matching the accuracy of the best single API. Authors: Lingjiao Chen ( Stanford University ) James Zou ( Stanford University ) Matei Zaharia ( Stanford and Databricks ) |
Lingjiao Chen 🔗 |
Fri 7:20 a.m. - 7:30 a.m.
|
Competition over data: when does data purchase benefit users?
(
Contributed Talk
)
SlidesLive Video » As machine learning (ML) is deployed by many competing service providers, the underlying ML predictors also compete against each other, and it is increasingly important to understand the impacts and biases from such competition. In this paper, we study what happens when the competing predictors can acquire additional labeled data to improve their prediction quality. We introduce a new environment that allows ML predictors to use active learning algorithms to purchase labeled data within their budgets while competing against each other to attract users. Our environment models a critical aspect of data acquisition in competing systems which has not been well-studied before. When predictors can purchase additional labeled data, their overall performance improves. Surprisingly, however, the quality that users experience---i.e. the accuracy of the predictor selected by each user---can decrease even as the individual predictors get better. We show that this phenomenon naturally arises due to a trade-off whereby competition pushes each predictor to specialize in a subset of the population while data purchase has the effect of making predictors more uniform. Authors: Yongchan Kwon ( Stanford University ) Tony Ginart ( Stanford University ) James Zou ( Stanford University ) |
Yongchan Kwon 🔗 |
Fri 7:30 a.m. - 8:00 a.m.
|
Poster Session
link »
The links to live Q&A: [ protected link dropped ] [ protected link dropped ] 3- Towards Efficient Machine Unlearning via Incremental View Maintenance [ protected link dropped ] [ protected link dropped ] [ protected link dropped ] |
Kishor Datta Gupta · Sebastian Schelter · Till Döhmen · Tony Ginart · Lingjiao Chen · Yongchan Kwon 🔗 |
Fri 8:00 a.m. - 9:00 a.m.
|
Deployment and monitoring on constrained hardware and devices
(
Panel Discussion
)
SlidesLive Video » Chair: Stephen Roberts, University of Oxford Panelists: - Partha Maji, ARM. "How can we extract true model uncertainty with little to no additional computational costs?", "Are popular Bayesian techniques equally expressible after they have gone through several stages of model compression?" - Cecilia Mascolo, University of Cambridge. "On-device Uncertainty Estimation" - Ivan Kiskin, University of Oxford. "Data Drift", "Updating ML lifecycle" - Yunpeng Li, University of Surrey. "Server-side vs client-side implementation" - Maria Nyamukuru, Dartmouth College. "Pathways to unified embedded machine learning algorithms" |
Cecilia Mascolo · Maria Nyamukuru · Ivan Kiskin · Partha Maji · Yunpeng Li · Stephen Roberts 🔗 |
Fri 9:00 a.m. - 9:10 a.m.
|
Short Break
|
🔗 |
Fri 9:10 a.m. - 9:50 a.m.
|
Ethics of developing ML in healthcare
(
Invited Talk
)
SlidesLive Video » Speaker: Shalmali Joshi, Postdoctoral Fellow at the Center for Research on Computation on Society, Harvard University (SEAS) https://shalmalijoshi.github.io/ |
Shalmali Joshi 🔗 |
Fri 9:50 a.m. - 10:30 a.m.
|
Model-less Inference Serving for ease-to-use and cost-efficiency
(
Invited Talk
)
SlidesLive Video » The number of applications relying on inference from Machine Learning (ML) models is already large and expected to keep growing. Facebook, for instance, serves tens-of-trillions of inference queries per day. Distributed inference dominates ML production costs: on AWS, it accounts for over 90% of ML infrastructure cost. Despite existing work in machine learning inference serving, ease-of-use and cost efficiency remain challenges at large scales. Developers must manually search through thousands of model-variants—versions of already-trained models that differ in hardware, resource footprints, latencies, costs, and accuracies—to meet the diverse application requirements. Since requirements, query load, and applications themselves evolve over time, these decisions need to be made dynamically for each inference query to avoid excessive costs through naive autoscaling. To avoid navigating through the large and complex trade-off space of model-variants, developers often fix a variant across queries, and replicate it when load increases. However, given the diversity across variants and hardware platforms in the cloud, a lack of understanding of the trade-off space can incur significant costs to developers. In this talk, I will primarily focus on INFaaS, an automated model-less system for distributed inference serving, where developers simply specify the performance and accuracy requirements for their applications without needing to specify a specific model-variant for each query. INFaaS generates model-variants from already trained models, and efficiently navigates the large trade-off space of model-variants on behalf of developers to meet application-specific objectives: (a) for each query, it selects a model, hardware architecture, and model optimizations, (b) it combines VM-level horizontal autoscaling with model-level autoscaling, where multiple, different model-variants are used to serve queries within each machine. By leveraging diverse variants and sharing hardware resources across models, INFaaS achieves significant improvement in performance (throughput and latency of model serving) while saving costs compared to existing inference serving systems. I will conclude the talk with a brief discussion on future directions. |
Neeraja J Yadwadkar 🔗 |
Fri 10:30 a.m. - 11:30 a.m.
|
Invited Speakers' Panel
(
Panel Discussion
)
SlidesLive Video » Chair: Stephen Roberts, University of Oxford |
Neeraja J Yadwadkar · Shalmali Joshi · Roberto Bondesan · Engineer Bainomugisha · Stephen Roberts 🔗 |
Author Information
Alessandra Tosi (Mind Foundry)
Nathan Korda (Mind Foundry)
Michael A Osborne (U Oxford)
Stephen Roberts (University of Oxford)
Andrei Paleyes (University of Cambridge)
Fariba Yousefi (University of Sheffield)
More from the Same Authors
-
2021 : Attacking Graph Classification via Bayesian Optimisation »
Xingchen Wan · Henry Kenlay · Binxin Ru · Arno Blaas · Michael A Osborne · Xiaowen Dong -
2021 : Revisiting Design Choices in Offline Model Based Reinforcement Learning »
Cong Lu · Philip Ball · Jack Parker-Holder · Michael A Osborne · Stephen Roberts -
2022 : Challenges and Opportunities in Offline Reinforcement Learning from Visual Observations »
Cong Lu · Philip Ball · Tim G. J Rudner · Jack Parker-Holder · Michael A Osborne · Yee-Whye Teh -
2022 Poster: Robust Multi-Objective Bayesian Optimization Under Input Noise »
Samuel Daulton · Sait Cakmak · Maximilian Balandat · Michael A Osborne · Enlu Zhou · Eytan Bakshy -
2022 Spotlight: Robust Multi-Objective Bayesian Optimization Under Input Noise »
Samuel Daulton · Sait Cakmak · Maximilian Balandat · Michael A Osborne · Enlu Zhou · Eytan Bakshy -
2022 Poster: Stabilizing Off-Policy Deep Reinforcement Learning from Pixels »
Edoardo Cetin · Philip Ball · Stephen Roberts · Oya Celiktutan -
2022 Spotlight: Stabilizing Off-Policy Deep Reinforcement Learning from Pixels »
Edoardo Cetin · Philip Ball · Stephen Roberts · Oya Celiktutan -
2021 : Invited Speakers' Panel »
Neeraja J Yadwadkar · Shalmali Joshi · Roberto Bondesan · Engineer Bainomugisha · Stephen Roberts -
2021 : Deployment and monitoring on constrained hardware and devices »
Cecilia Mascolo · Maria Nyamukuru · Ivan Kiskin · Partha Maji · Yunpeng Li · Stephen Roberts -
2021 : Opening remarks »
Alessandra Tosi · Nathan Korda · Fariba Yousefi · Andrei Paleyes · Stephen Roberts -
2021 Poster: Think Global and Act Local: Bayesian Optimisation over High-Dimensional Categorical and Mixed Search Spaces »
Xingchen Wan · Vu Nguyen · Huong Ha · Binxin Ru · Cong Lu · Michael A Osborne -
2021 Poster: Optimal Transport Kernels for Sequential and Parallel Neural Architecture Search »
Vu Nguyen · Tam Le · Makoto Yamada · Michael A Osborne -
2021 Poster: Augmented World Models Facilitate Zero-Shot Dynamics Generalization From a Single Offline Environment »
Philip Ball · Cong Lu · Jack Parker-Holder · Stephen Roberts -
2021 Spotlight: Think Global and Act Local: Bayesian Optimisation over High-Dimensional Categorical and Mixed Search Spaces »
Xingchen Wan · Vu Nguyen · Huong Ha · Binxin Ru · Cong Lu · Michael A Osborne -
2021 Spotlight: Augmented World Models Facilitate Zero-Shot Dynamics Generalization From a Single Offline Environment »
Philip Ball · Cong Lu · Jack Parker-Holder · Stephen Roberts -
2021 Spotlight: Optimal Transport Kernels for Sequential and Parallel Neural Architecture Search »
Vu Nguyen · Tam Le · Makoto Yamada · Michael A Osborne -
2020 : Contributed Talk 1: Provably Efficient Online Hyperparameter Optimization with Population-Based Bandits »
Jack Parker-Holder · Vu Nguyen · Stephen Roberts -
2020 : Open Problems Panel »
Alessandra Tosi · Nathan Korda · Yuzhui Liu · Zhenwen Dai · Zhenwen Dai · Alexander Lavin · Erick Galinkin · Camylle Lanteigne -
2020 Workshop: Challenges in Deploying and Monitoring Machine Learning Systems »
Alessandra Tosi · Nathan Korda · Neil Lawrence -
2020 : Opening remarks »
Alessandra Tosi · Nathan Korda -
2020 Poster: Knowing The What But Not The Where in Bayesian Optimization »
Vu Nguyen · Michael A Osborne -
2020 Poster: Ready Policy One: World Building Through Active Learning »
Philip Ball · Jack Parker-Holder · Aldo Pacchiano · Krzysztof Choromanski · Stephen Roberts -
2020 Poster: Bayesian Optimisation over Multiple Continuous and Categorical Inputs »
Binxin Ru · Ahsan Alvi · Vu Nguyen · Michael A Osborne · Stephen Roberts -
2019 : Poster discussion »
Roman Novak · Maxime Gabella · Frederic Dreyer · Siavash Golkar · Anh Tong · Irina Higgins · Mirco Milletari · Joe Antognini · Sebastian Goldt · Adín Ramírez Rivera · Roberto Bondesan · Ryo Karakida · Remi Tachet des Combes · Michael Mahoney · Nicholas Walker · Stanislav Fort · Samuel Smith · Rohan Ghosh · Aristide Baratin · Diego Granziol · Stephen Roberts · Dmitry Vetrov · Andrew Wilson · César Laurent · Valentin Thomas · Simon Lacoste-Julien · Dar Gilboa · Daniel Soudry · Anupam Gupta · Anirudh Goyal · Yoshua Bengio · Erich Elsen · Soham De · Stanislaw Jastrzebski · Charles H Martin · Samira Shabanian · Aaron Courville · Shorato Akaho · Lenka Zdeborova · Ethan Dyer · Maurice Weiler · Pim de Haan · Taco Cohen · Max Welling · Ping Luo · zhanglin peng · Nasim Rahaman · Loic Matthey · Danilo J. Rezende · Jaesik Choi · Kyle Cranmer · Lechao Xiao · Jaehoon Lee · Yasaman Bahri · Jeffrey Pennington · Greg Yang · Jiri Hron · Jascha Sohl-Dickstein · Guy Gur-Ari -
2019 Poster: On the Limitations of Representing Functions on Sets »
Edward Wagstaff · Fabian Fuchs · Martin Engelcke · Ingmar Posner · Michael A Osborne -
2019 Oral: On the Limitations of Representing Functions on Sets »
Edward Wagstaff · Fabian Fuchs · Martin Engelcke · Ingmar Posner · Michael A Osborne -
2019 Poster: Automated Model Selection with Bayesian Quadrature »
Henry Chai · Jean-Francois Ton · Michael A Osborne · Roman Garnett -
2019 Poster: AReS and MaRS - Adversarial and MMD-Minimizing Regression for SDEs »
Gabriele Abbati · Philippe Wenk · Michael A Osborne · Andreas Krause · Bernhard Schölkopf · Stefan Bauer -
2019 Poster: Asynchronous Batch Bayesian Optimisation with Improved Local Penalisation »
Ahsan Alvi · Binxin Ru · Jan-Peter Calliess · Stephen Roberts · Michael A Osborne -
2019 Oral: Automated Model Selection with Bayesian Quadrature »
Henry Chai · Jean-Francois Ton · Michael A Osborne · Roman Garnett -
2019 Oral: AReS and MaRS - Adversarial and MMD-Minimizing Regression for SDEs »
Gabriele Abbati · Philippe Wenk · Michael A Osborne · Andreas Krause · Bernhard Schölkopf · Stefan Bauer -
2019 Oral: Asynchronous Batch Bayesian Optimisation with Improved Local Penalisation »
Ahsan Alvi · Binxin Ru · Jan-Peter Calliess · Stephen Roberts · Michael A Osborne -
2019 Poster: Fingerprint Policy Optimisation for Robust Reinforcement Learning »
Supratik Paul · Michael A Osborne · Shimon Whiteson -
2019 Oral: Fingerprint Policy Optimisation for Robust Reinforcement Learning »
Supratik Paul · Michael A Osborne · Shimon Whiteson -
2018 Poster: Fast Information-theoretic Bayesian Optimisation »
Binxin Ru · Michael A Osborne · Mark Mcleod · Diego Granziol -
2018 Poster: Optimization, fast and slow: optimally switching between local and Bayesian optimization »
Mark McLeod · Stephen Roberts · Michael A Osborne -
2018 Oral: Optimization, fast and slow: optimally switching between local and Bayesian optimization »
Mark McLeod · Stephen Roberts · Michael A Osborne -
2018 Oral: Fast Information-theoretic Bayesian Optimisation »
Binxin Ru · Michael A Osborne · Mark Mcleod · Diego Granziol