Kronecker-factored approximation (KFAC) of the Laplace-GGN for Continual Learning

By Nadhir Hassen in Research

May 12, 2021

PDF

Abstract

One major challenge in Continual Learning is that the model forgets how to solve earlier tasks commonly known as Catastrophic Forgetting. Extensive work has been recently made in this direction relies of weight regularisation but may not be effective for complex deep neural network. Recently functional regularisation was developed to directly regularise the network output, although computationally expensive, is expected to perform better. However this technique limits the posterior expressiveness of fully-factorized Gaussian assumption Cremer et al. (2018) of the inference model. Functional regularisation (Pan et al., 2021) has been developed to combine parameter and function space regularisation but does not take into account structured parameter spaces where have complicated inverse Fisher-matrix computations, which give poor predictives estimates. The method applies a full covariance Gaussian posterior approximation and in case of diagonal posterior approximation it fails to model meaningful relationships. In this work we address this issue without alternating the complexity cost. We propose a Laplace Gauss-Newton approximation to combine local parametrization and function space learning using Kronecker-factored approximation (KFAC) of the Laplace-Gauss-Newton (GGN). We use KFAC (Martens & Grosse, 2020a) posterior approximation for image classification to illustrate how we can model parameter covariances per layer without altering scalability in a Continual Learning setting.

Posted on:
May 12, 2021
Length:
1 minute read, 205 words
Categories:
Research
Tags:
Continual Learning Bayesian Deep Learning Optimization Approximate Inference Gaussian Processes
See Also:
Approximate Bayesian Neural Networks
Approximate Bayesian Optimisation for Neural Networks
Orion-Asynchronous Distributed Hyperparameter Optimization