Probabilistic classifiers with a generalized Gaussian scale mixture prior

Article


Liu, G., Wu, J. and Zhou, S. 2013. Probabilistic classifiers with a generalized Gaussian scale mixture prior. Pattern Recognition. 46 (1), pp. 332-345. https://doi.org/10.1016/j.patcog.2012.07.016
TypeArticle
TitleProbabilistic classifiers with a generalized Gaussian scale mixture prior
AuthorsLiu, G., Wu, J. and Zhou, S.
Abstract

Most of the existing probabilistic classifiers are based on sparsity-inducing modeling. However, we show that sparsity is not always desirable in practice, and only an appropriate degree of sparsity is profitable. In this work, we propose a flexible probabilistic model using a generalized Gaussian scale mixture (GGSM) prior that can provide an appropriate degree of sparsity for its model parameters, and yield either sparse or non-sparse estimates according to the intrinsic sparsity of features in a dataset. Model learning is carried out by an efficient modified maximum a posterior (MAP) estimate. We also show relationships of the proposed model to existing probabilistic classifiers as well as iteratively re-weighted l1 and l2 minimizations. We then study different types of likelihood working with the GGSM prior in kernel-based setup, based on which an improved kernel-based GGIG is presented. Experiments demonstrate that the proposed method has better or comparable performances in linear classifiers as well as in kernel-based classification.

KeywordsClassification; Prior distribution; Generalized Gaussian scale mixture; Likelihood function
PublisherElsevier Science
JournalPattern Recognition
ISSN0031-3203
Publication dates
Print2013
Publication process dates
Deposited22 Nov 2013
Output statusPublished
Digital Object Identifier (DOI)https://doi.org/10.1016/j.patcog.2012.07.016
LanguageEnglish
Permalink -

https://repository.mdx.ac.uk/item/84912

  • 16
    total views
  • 0
    total downloads
  • 0
    views this month
  • 0
    downloads this month

Export as