Variational Bayesian multinomial probit regression with gaussian process priors

被引:141
作者
Girolami, Mark [1 ]
Rogers, Simon [1 ]
机构
[1] Univ Glasgow, Dept Comp Sci, Glasgow G12 8QQ, Lanark, Scotland
关键词
D O I
10.1162/neco.2006.18.8.1790
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is well known in the statistics literature that augmenting binary and polychotomous response models with gaussian latent variables enables exact Bayesian analysis via Gibbs sampling from the parameter posterior. By adopting such a data augmentation strategy, dispensing with priors over regression coefficients in favor of gaussian process (GP) priors over functions, and employing variational approximations to the full posterior, we obtain efficient computational methods for GP classification in the multiclass setting.(1) The model augmentation with additional latent variables ensures full a posteriori class coupling while retaining the simple a priori independent GP covariance structure from which sparse approximations, such as multiclass informative vector machines (IVM), emerge in a natural and straightforward manner. This is the first time that a fully variational Bayesian treatment for multiclass GP classification has been developed without having to resort to additional explicit approximations to the nongaussian likelihood term. Empirical comparisons with exact analysis use Markov Chain Monte Carlo (MCMC) and Laplace approximations illustrate the utility of the variational approximation as a computationally economic alternative to full MCMC and it is shown to be more accurate than the Laplace approximation.
引用
收藏
页码:1790 / 1817
页数:28
相关论文
共 29 条
  • [1] BAYESIAN-ANALYSIS OF BINARY AND POLYCHOTOMOUS RESPONSE DATA
    ALBERT, JH
    CHIB, S
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1993, 88 (422) : 669 - 679
  • [2] Beal M.J., 2003, Variational algorithms for approximate Bayesian inference
  • [3] Chu W, 2005, J MACH LEARN RES, V6, P1019
  • [4] Csató L, 2000, ADV NEUR IN, V12, P251
  • [5] Sparse on-line Gaussian processes
    Csató, L
    Opper, M
    [J]. NEURAL COMPUTATION, 2002, 14 (03) : 641 - 668
  • [6] Duan KB, 2005, LECT NOTES COMPUT SC, V3541, P278
  • [7] Gibbs MN, 2000, IEEE T NEURAL NETWOR, V11, P1458, DOI 10.1109/72.883477
  • [8] Girolami M., 2005, Proceedings of the 22nd International Conference on Machine Learning, P241, DOI DOI 10.1145/1102351.1102382
  • [9] An introduction to variational methods for graphical models
    Jordan, MI
    Ghahramani, Z
    Jaakkola, TS
    Saul, LK
    [J]. MACHINE LEARNING, 1999, 37 (02) : 183 - 233
  • [10] KIM HC, 2005, THESIS POHANG U SCI