Timezone: »
The success of Convolutional Neural Networks (CNNs) in computer vision is mainly driven by their strong inductive bias, which is strong enough to allow CNNs to solve vision-related tasks with random weights, meaning without learning. Similarly, Long Short-Term Memory (LSTM) has a strong inductive bias towards storing information over time. However, many real-world systems are governed by conservation laws, which lead to the redistribution of particular quantities — e.g.in physical and economical systems. Our novel Mass-Conserving LSTM (MC-LSTM) adheres to these conservation laws by extending the inductive bias of LSTM to model the redistribution of those stored quantities. MC-LSTMs set a new state-of-the-art for neural arithmetic units at learning arithmetic operations, such as addition tasks,which have a strong conservation law, as the sum is constant over time. Further, MC-LSTM is applied to traffic forecasting, modeling a pendulum, and a large benchmark dataset in hydrology, where it sets a new state-of-the-art for predicting peak flows. In the hydrology example, we show that MC-LSTM states correlate with real world processes and are therefore interpretable.
Author Information
Pieter-Jan Hoedt (LIT AI Lab, Institute for Machine Learning, Johannes Kepler University Linz, Austria)
Frederik Kratzert (LIT AI Lab, Institute for Machine Learning, Johannes Kepler University Linz, Austria)
Daniel Klotz (Ellis Unit / University Linz)
Christina Halmich (LIT AI Lab, Institute for Machine Learning, Johannes Kepler University Linz, Austria)
Markus Holzleitner (LIT AI Lab / University Linz)
Grey Nearing (Google Research)
Sepp Hochreiter (LIT AI Lab, Institute for Machine Learning, Johannes Kepler University Linz, Austria)

Sepp Hochreiter is heading the Institute for Machine Learning, the ELLIS Unit Linz, the LIT AI Lab at the JKU Linz and is director of private research institute IARAI. He is a pioneer of Deep Learning as he discovered the famous problem of vanishing or exploding gradients and invented the long short-term memory (LSTM).
Günter Klambauer (Johannes Kepler University Linz Austria)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Poster: MC-LSTM: Mass-Conserving LSTM »
Tue. Jul 20th 04:00 -- 06:00 PM Room
More from the Same Authors
-
2023 Poster: Enhancing Activity Prediction Models in Drug Discovery with the Ability to Understand Human Language »
Philipp Seidl · Andreu Vall · Sepp Hochreiter · Günter Klambauer -
2022 Poster: History Compression via Language Models in Reinforcement Learning »
Fabian Paischer · Thomas Adler · Vihang Patil · Angela Bitto-Nemling · Markus Holzleitner · Sebastian Lehner · Hamid Eghbal-zadeh · Sepp Hochreiter -
2022 Spotlight: History Compression via Language Models in Reinforcement Learning »
Fabian Paischer · Thomas Adler · Vihang Patil · Angela Bitto-Nemling · Markus Holzleitner · Sebastian Lehner · Hamid Eghbal-zadeh · Sepp Hochreiter -
2022 Poster: Align-RUDDER: Learning From Few Demonstrations by Reward Redistribution »
Vihang Patil · Markus Hofmarcher · Marius-Constantin Dinu · Matthias Dorfer · Patrick Blies · Johannes Brandstetter · Jose A. Arjona-Medina · Sepp Hochreiter -
2022 Oral: Align-RUDDER: Learning From Few Demonstrations by Reward Redistribution »
Vihang Patil · Markus Hofmarcher · Marius-Constantin Dinu · Matthias Dorfer · Patrick Blies · Johannes Brandstetter · Jose A. Arjona-Medina · Sepp Hochreiter