Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning
As the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9364973/ |
id |
doaj-dc22e6654cd74d58b41c2fc6743097d0 |
---|---|
record_format |
Article |
spelling |
doaj-dc22e6654cd74d58b41c2fc6743097d02021-03-30T15:31:59ZengIEEEIEEE Access2169-35362021-01-019362413625110.1109/ACCESS.2021.30627479364973Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep LearningYongguang Wang0https://orcid.org/0000-0002-0470-9178Shuzhen Yao1https://orcid.org/0000-0001-7785-9698Tian Xu2https://orcid.org/0000-0001-6108-7494School of Computer Science and Engineering, Beihang University, Beijing, ChinaSchool of Computer Science and Engineering, Beihang University, Beijing, ChinaSchool of Computer Science and Engineering, Beihang University, Beijing, ChinaAs the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear subspace feature extractor, and it cannot effectively represent the nonlinearly high-dimensional parameter space of deep neural networks (DNNs). Firstly, in this paper, in order to solve the main problem of the linear characteristics of PCA in high-dimensional space, we apply kernel PCA to extract higher-order statistical information in parameter space of DNNs. Secondly, to improve the efficiency of subsequent computation, we propose a strictly ordered incremental kernel PCA (InKPCA) subspace of parameter space within stochastic gradient descent (SGD) trajectories. In the proposed InKPCA subspace, we employ two approximation inference methods: elliptical slice sampling (ESS) and variational inference (VI). Finally, to further improve the memory efficiency of computing the kernel matrix, we apply Nyström approximation to determine the suitable size of subsets in the original datasets. The novelty of this paper is that it is the first time to apply the proposed InKPCA subspace with Nyström approximation for Bayesian inference in DNNs, and the results show that it can produce more accurate predictions and well-calibrated predictive uncertainty in regression and classification tasks of deep learning.https://ieeexplore.ieee.org/document/9364973/Bayesian deep learningincremental kernel PCAelliptical slice samplingvariational inferenceNyström approximation |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Yongguang Wang Shuzhen Yao Tian Xu |
spellingShingle |
Yongguang Wang Shuzhen Yao Tian Xu Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning IEEE Access Bayesian deep learning incremental kernel PCA elliptical slice sampling variational inference Nyström approximation |
author_facet |
Yongguang Wang Shuzhen Yao Tian Xu |
author_sort |
Yongguang Wang |
title |
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning |
title_short |
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning |
title_full |
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning |
title_fullStr |
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning |
title_full_unstemmed |
Incremental Kernel Principal Components Subspace Inference With Nyström Approximation for Bayesian Deep Learning |
title_sort |
incremental kernel principal components subspace inference with nyström approximation for bayesian deep learning |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
As the state-of-the-art technology of Bayesian inference, based on low-dimensional principal components analysis (PCA) subspace inference methods can provide approximately accurate predictive distribution and well calibrated uncertainty. However, the main problem of PCA method is that it is a linear subspace feature extractor, and it cannot effectively represent the nonlinearly high-dimensional parameter space of deep neural networks (DNNs). Firstly, in this paper, in order to solve the main problem of the linear characteristics of PCA in high-dimensional space, we apply kernel PCA to extract higher-order statistical information in parameter space of DNNs. Secondly, to improve the efficiency of subsequent computation, we propose a strictly ordered incremental kernel PCA (InKPCA) subspace of parameter space within stochastic gradient descent (SGD) trajectories. In the proposed InKPCA subspace, we employ two approximation inference methods: elliptical slice sampling (ESS) and variational inference (VI). Finally, to further improve the memory efficiency of computing the kernel matrix, we apply Nyström approximation to determine the suitable size of subsets in the original datasets. The novelty of this paper is that it is the first time to apply the proposed InKPCA subspace with Nyström approximation for Bayesian inference in DNNs, and the results show that it can produce more accurate predictions and well-calibrated predictive uncertainty in regression and classification tasks of deep learning. |
topic |
Bayesian deep learning incremental kernel PCA elliptical slice sampling variational inference Nyström approximation |
url |
https://ieeexplore.ieee.org/document/9364973/ |
work_keys_str_mv |
AT yongguangwang incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning AT shuzhenyao incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning AT tianxu incrementalkernelprincipalcomponentssubspaceinferencewithnystrx00f6mapproximationforbayesiandeeplearning |
_version_ |
1724179347371196416 |