Alpha-divergence minimization for deep Gaussian processes
نویسندگان
چکیده
This paper proposes the minimization of α-divergences for approximate inference in context deep Gaussian processes (DGPs). The proposed method can be considered as a generalization variational (VI) and expectation propagation (EP), two previously used methods DGPs. Both VI EP are based on Kullback-Leibler divergence. is scalable version power propagation, that introduces an extra parameter α specifies targeted α-divergence to optimized. In particular, such recover solution when α→0 α→1. An exhaustive experimental evaluation shows via feasible DGPs choosing intermediate values between 0 1 give better results some problems. means one improve training Importantly, allows stochastic optimization techniques, making it able address datasets with several millions instances.
منابع مشابه
Deep Gaussian Processes
In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief network based on Gaussian process mappings. The data is modeled as the output of a multivariate GP. The inputs to that Gaussian process are then governed by another GP. A single layer model is equivalent to a standard GP or the GP latent variable model (GP-LVM). We perform inference in the model by approxim...
متن کاملThe Rate of Entropy for Gaussian Processes
In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...
متن کاملRandom Feature Expansions for Deep Gaussian Processes
The composition of multiple Gaussian Processes as a Deep Gaussian Process (DGP) enables a deep probabilistic nonparametric approach to flexibly tackle complex machine learning problems with sound quantification of uncertainty. Existing inference approaches for DGP models have limited scalability and are notoriously cumbersome to construct. In this work we introduce a novel formulation of DGPs b...
متن کاملContrastive Divergence In Gaussian Diffusion Processes
Contrastive Divergence (CD) is a recent learning rule found to work very well in practice despite its still unclear theoretical underpinnings (Carreira-Perpiñan and Hinton, 2005; Hinton, 2002; Hinton and Salakhutdinov, 2006; Hyvärinen, 2006; MacKay, 2001; Roth and Black, 2005; Williams and Agakov, 2002; Yuille, 2004). This paper presents a theoretical analysis of CD in Gaussian Diffusion Proces...
متن کاملDeep Multi-fidelity Gaussian Processes
We develop a novel multi-fidelity framework that goes far beyond the classical AR(1) Co-kriging scheme of Kennedy and O’Hagan (2000). Our method can handle general discontinuous cross-correlations among systems with different levels of fidelity. A combination of multi-fidelity Gaussian Processes (AR(1) Co-kriging) and deep neural networks enables us to construct a method that is immune to disco...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: International Journal of Approximate Reasoning
سال: 2022
ISSN: ['1873-4731', '0888-613X']
DOI: https://doi.org/10.1016/j.ijar.2022.08.003