Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning

As the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear...

Full description

Bibliographic Details
Main Authors: Yongguang Wang, Shuzhen Yao, Tian Xu
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9364973/
id doaj-dc22e6654cd74d58b41c2fc6743097d0
record_format Article
spelling doaj-dc22e6654cd74d58b41c2fc6743097d02021-03-30T15:31:59ZengIEEEIEEE Access2169-35362021-01-019362413625110.1109/ACCESS.2021.30627479364973Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep LearningYongguang Wang0https://orcid.org/0000-0002-0470-9178Shuzhen Yao1https://orcid.org/0000-0001-7785-9698Tian Xu2https://orcid.org/0000-0001-6108-7494School of Computer Science and Engineering, Beihang University, Beijing, ChinaSchool of Computer Science and Engineering, Beihang University, Beijing, ChinaSchool of Computer Science and Engineering, Beihang University, Beijing, ChinaAs the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear subspace feature extractor, and it cannot effectively represent the nonlinearly high-dimensional parameter space of deep neural networks (DNNs). Firstly, in this paper, in order to solve the main problem of the linear characteristics of PCA in high-dimensional space, we apply kernel PCA to extract higher-order statistical information in parameter space of DNNs. Secondly, to improve the efficiency of subsequent computation, we propose a strictly ordered incremental kernel PCA (InKPCA) subspace of parameter space within stochastic gradient descent (SGD) trajectories. In the proposed InKPCA subspace, we employ two approximation inference methods: elliptical slice sampling (ESS) and variational inference (VI). Finally, to further improve the memory efficiency of computing the kernel matrix, we apply Nyström approximation to determine the suitable size of subsets in the original datasets. The novelty of this paper is that it is the first time to apply the proposed InKPCA subspace with Nyström approximation for Bayesian inference in DNNs, and the results show that it can produce more accurate predictions and well-calibrated predictive uncertainty in regression and classification tasks of deep learning.https://ieeexplore.ieee.org/document/9364973/Bayesian deep learningincremental kernel PCAelliptical slice samplingvariational inferenceNyström approximation
collection DOAJ
language English
format Article
sources DOAJ
author Yongguang Wang
Shuzhen Yao
Tian Xu
spellingShingle Yongguang Wang
Shuzhen Yao
Tian Xu
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
IEEE Access
Bayesian deep learning
incremental kernel PCA
elliptical slice sampling
variational inference
Nyström approximation
author_facet Yongguang Wang
Shuzhen Yao
Tian Xu
author_sort Yongguang Wang
title Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
title_short Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
title_full Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
title_fullStr Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
title_full_unstemmed Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
title_sort incremental kernel principal components subspace inference with nyström approximation for bayesian deep learning
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2021-01-01
description As the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear subspace feature extractor, and it cannot effectively represent the nonlinearly high-dimensional parameter space of deep neural networks (DNNs). Firstly, in this paper, in order to solve the main problem of the linear characteristics of PCA in high-dimensional space, we apply kernel PCA to extract higher-order statistical information in parameter space of DNNs. Secondly, to improve the efficiency of subsequent computation, we propose a strictly ordered incremental kernel PCA (InKPCA) subspace of parameter space within stochastic gradient descent (SGD) trajectories. In the proposed InKPCA subspace, we employ two approximation inference methods: elliptical slice sampling (ESS) and variational inference (VI). Finally, to further improve the memory efficiency of computing the kernel matrix, we apply Nyström approximation to determine the suitable size of subsets in the original datasets. The novelty of this paper is that it is the first time to apply the proposed InKPCA subspace with Nyström approximation for Bayesian inference in DNNs, and the results show that it can produce more accurate predictions and well-calibrated predictive uncertainty in regression and classification tasks of deep learning.
topic Bayesian deep learning
incremental kernel PCA
elliptical slice sampling
variational inference
Nyström approximation
url https://ieeexplore.ieee.org/document/9364973/
work_keys_str_mv AT yongguangwang incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning
AT shuzhenyao incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning
AT tianxu incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning
_version_ 1724179347371196416