JFB: Jacobian-Free Backpropagation for Implicit Networks

Authors

  • Samy Wu Fung Colorado School of Mines
  • Howard Heaton Typal Research
  • Qiuwei Li Damo Academy, Alibaba Group US
  • Daniel Mckenzie University of California, Los Angeles
  • Stanley Osher University of California, Los Angeles
  • Wotao Yin Damo Academy, Alibaba Group US

DOI:

https://doi.org/10.1609/aaai.v36i6.20619

Keywords:

Machine Learning (ML)

Abstract

A promising trend in deep learning replaces traditional feedforward networks with implicit networks. Unlike traditional networks, implicit networks solve a fixed point equation to compute inferences. Solving for the fixed point varies in complexity, depending on provided data and an error tolerance. Importantly, implicit networks may be trained with fixed memory costs in stark contrast to feedforward networks, whose memory requirements scale linearly with depth. However, there is no free lunch --- backpropagation through implicit networks often requires solving a costly Jacobian-based equation arising from the implicit function theorem. We propose Jacobian-Free Backpropagation (JFB), a fixed-memory approach that circumvents the need to solve Jacobian-based equations. JFB makes implicit networks faster to train and significantly easier to implement, without sacrificing test accuracy. Our experiments show implicit networks trained with JFB are competitive with feedforward networks and prior implicit networks given the same number of parameters.

Downloads

Published

2022-06-28

How to Cite

Fung, S. W., Heaton, H., Li, Q., Mckenzie, D., Osher, S., & Yin, W. (2022). JFB: Jacobian-Free Backpropagation for Implicit Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 36(6), 6648-6656. https://doi.org/10.1609/aaai.v36i6.20619

Issue

Section

AAAI Technical Track on Machine Learning I