Robust Learning with Implicit Residual Networks

Viktor Reshniak, Clayton G. Webster

Research output: Contribution to journalArticlepeer-review

5 Scopus citations

Abstract

In this effort, we propose a new deep architecture utilizing residual blocks inspired by implicit discretization schemes. As opposed to the standard feed-forward networks, the outputs of the proposed implicit residual blocks are defined as the fixed points of the appropriately chosen nonlinear transformations. We show that this choice leads to the improved stability of both forward and backward propagations, has a favorable impact on the generalization power, and allows for control the robustness of the network with only a few hyperparameters. In addition, the proposed reformulation of ResNet does not introduce new parameters and can potentially lead to a reduction in the number of required layers due to improved forward stability. Finally, we derive the memory-efficient training algorithm, propose a stochastic regularization technique, and provide numerical results in support of our findings.

Original languageEnglish
Pages (from-to)34-55
Number of pages22
JournalMachine Learning and Knowledge Extraction
Volume3
Issue number1
DOIs
StatePublished - Mar 2021

Funding

This research was funded by the U.S. Department of Energy, Office of Science, Early Career Research Program under award number ERKJ314; U.S. Department of Energy, Office of Advanced Scientific Computing Research under award numbers ERKJ331 and ERKJ345; the National Science Foundation, Division of Mathematical Sciences, Computational Mathematics program under contract number DMS1620280; and the Behavioral Reinforcement Learning Lab at Lirio LLC.

Keywords

  • ResNet
  • robust
  • stability

Fingerprint

Dive into the research topics of 'Robust Learning with Implicit Residual Networks'. Together they form a unique fingerprint.

Cite this