Practical Algorithms for Latent Variable Models

Practical Algorithms for Latent Variable Models
Title Practical Algorithms for Latent Variable Models PDF eBook
Author Gregory W. Gundersen
Publisher
Pages 0
Release 2021
Genre
ISBN

Download Practical Algorithms for Latent Variable Models Book in PDF, Epub and Kindle

Latent variables allow researchers and engineers to encode assumptions into their statistical models. A latent variable might, for example, represent an unobserved covariate, measurement error, or a missing class label. Inference is challenging because one must account for the conditional dependence structure induced by these variables, and marginalization is often intractable. In this thesis, I present several practical algorithms for inferring latent structure in probabilistic models used in computational biology, neuroscience, and time-series analysis.First, I present a multi-view framework that combines neural networks and probabilistic canonical correlation analysis to estimate shared and view-specific latent structure of paired samples of histological images and gene expression levels. The model is trained end-to-end to estimate all parameters simultaneously, and we show that the latent variables capture interpretable structure, such as tissue-specific and morphological variation. Next, I present a family of nonlinear dimension-reduction models that use random features to support non-Gaussian data likelihoods. By approximating a nonlinear relationship between the latent variables and observations with a function that is linear with respect to random features, we induce closed-form gradients of the posterior distribution with respect to the latent variables. This allows for gradient-based nonlinear dimension-reduction models for a variety of data likelihoods. Finally, I discuss lowering the computational cost of online Bayesian filtering of time series with abrupt changes in structure, called changepoints. We consider settings in which a time series has multiple data sources, each with an associated cost. We trade the cost of a data source against the quality or fidelity of that source and how its fidelity affects the estimation of changepoints. Our framework makes cost-sensitive decisions about which data source to use based on minimizing the information entropy of the posterior distribution over changepoints.

Handbook of Latent Variable and Related Models

Handbook of Latent Variable and Related Models
Title Handbook of Latent Variable and Related Models PDF eBook
Author
Publisher Elsevier
Pages 458
Release 2011-08-11
Genre Mathematics
ISBN 0080471269

Download Handbook of Latent Variable and Related Models Book in PDF, Epub and Kindle

This Handbook covers latent variable models, which are a flexible class of models for modeling multivariate data to explore relationships among observed and latent variables. - Covers a wide class of important models - Models and statistical methods described provide tools for analyzing a wide spectrum of complicated data - Includes illustrative examples with real data sets from business, education, medicine, public health and sociology. - Demonstrates the use of a wide variety of statistical, computational, and mathematical techniques.

Bayesian Nonparametric Latent Variable Models

Bayesian Nonparametric Latent Variable Models
Title Bayesian Nonparametric Latent Variable Models PDF eBook
Author Patrick Dallaire
Publisher
Pages 146
Release 2016
Genre
ISBN

Download Bayesian Nonparametric Latent Variable Models Book in PDF, Epub and Kindle

One of the important problems in machine learning is determining the complexity of the model to learn. Too much complexity leads to overfitting, which finds structures that do not actually exist in the data, while too low complexity leads to underfitting, which means that the expressiveness of the model is insufficient to capture all the structures present in the data. For some probabilistic models, the complexity depends on the introduction of one or more latent variables whose role is to explain the generative process of the data. There are various approaches to identify the appropriate number of latent variables of a model. This thesis covers various Bayesian nonparametric methods capable of determining the number of latent variables to be used and their dimensionality. The popularization of Bayesian nonparametric statistics in the machine learning community is fairly recent. Their main attraction is the fact that they offer highly flexible models and their complexity scales appropriately with the amount of available data. In recent years, research on Bayesian nonparametric learning methods have focused on three main aspects: the construction of new models, the development of inference algorithms and new applications. This thesis presents our contributions to these three topics of research in the context of learning latent variables models. Firstly, we introduce the Pitman-Yor process mixture of Gaussians, a model for learning infinite mixtures of Gaussians. We also present an inference algorithm to discover the latent components of the model and we evaluate it on two practical robotics applications. Our results demonstrate that the proposed approach outperforms, both in performance and flexibility, the traditional learning approaches. Secondly, we propose the extended cascading Indian buffet process, a Bayesian nonparametric probability distribution on the space of directed acyclic graphs. In the context of Bayesian networks, this prior is used to identify the presence of latent variables and the network structure among them. A Markov Chain Monte Carlo inference algorithm is presented and evaluated on structure identification problems and as well as density estimation problems. Lastly, we propose the Indian chefs process, a model more general than the extended cascading Indian buffet process for learning graphs and orders. The advantage of the new model is that it accepts connections among observable variables and it takes into account the order of the variables. We also present a reversible jump Markov Chain Monte Carlo inference algorithm which jointly learns graphs and orders. Experiments are conducted on density estimation problems and testing independence hypotheses. This model is the first Bayesian nonparametric model capable of learning Bayesian learning networks with completely arbitrary graph structures.

Learning Latent Variable Models : Efficient Algorithms and Applications

Learning Latent Variable Models : Efficient Algorithms and Applications
Title Learning Latent Variable Models : Efficient Algorithms and Applications PDF eBook
Author Matteo Ruffini
Publisher
Pages 185
Release 2019
Genre
ISBN

Download Learning Latent Variable Models : Efficient Algorithms and Applications Book in PDF, Epub and Kindle

Learning latent variable models is a fundamental machine learning problem, and the models belonging to this class - which include topic models, hidden Markov models, mixture models and many others - have a variety of real-world applications, like text mining, clustering and time series analysis. For many practitioners, the decade-old Expectation Maximization method (EM) is still the tool of choice, despite its known proneness to local minima and long running times. To overcome these issues, algorithms based on the spectral method of moments have been recently proposed. These techniques recover the parameters of a latent variable model by solving - typically via tensor decomposition - a system of non-linear equations relating the low-order moments of the observable data with the parameters of the model to be learned. Moment-based algorithms are in general faster than EM as they require a single pass over the data, and have provable guarantees of learning accuracy in polynomial time. Nevertheless, methods of moments have room for improvements: their ability to deal with real-world data is often limited by a lack of robustness to input perturbations. Also, almost no theory studies their behavior when some of the model assumptions are violated by the input data. Extending the theory of methods of moments to learn latent variable models and providing meaningful applications to real-world contexts is the focus of this thesis. ssuming data to be generated by a certain latent variable model, the standard approach of methods of moments consists of two steps: first, finding the equations that relate the moments of the observable data with the model parameters and then, to solve these equations to retrieve estimators of the parameters of the model. In Part I of this thesis we will focus on both steps, providing and analyzing novel and improved model-specific moments estimators and techniques to solve the equations of the moments. In both the cases we will introduce theoretical results, providing guarantees on the behavior of the proposed methods, and we will perform experimental comparisons with existing algorithms. In Part II, we will analyze the behavior of methods of moments when data violates some of the model assumptions performed by a user. First, we will observe that in this context most of the theoretical infrastructure underlying methods of moments is not valid anymore, and consequently we will develop a theoretical foundation to methods of moments in the misspecified setting, developing efficient methods, guaranteed to provide meaningful results even when some of the model assumptions are violated. During all the thesis, we will apply the developed theoretical results to challenging real-world applications, focusing on two main domains: topic modeling and healthcare analytics. We will extend the existing theory of methods of moments to learn models that are traditionally used to do topic modeling – like the single-topic model and Latent Dirichlet Allocation – providing improved learning techniques and comparing them with existing methods, which we prove to outperform in terms of speed and learning accuracy. Furthermore, we will propose applications of latent variable models to the analysis of electronic healthcare records, which, similarly to text mining, are very likely to become massive datasets; we will propose a method to discover recurrent phenotypes in populations of patients and to cluster them in groups with similar clinical profiles - a task where the efficiency properties of methods of moments will constitute a competitive advantage over traditional approaches.

Variational Methods for Machine Learning with Applications to Deep Networks

Variational Methods for Machine Learning with Applications to Deep Networks
Title Variational Methods for Machine Learning with Applications to Deep Networks PDF eBook
Author Lucas Pinheiro Cinelli
Publisher Springer Nature
Pages 173
Release 2021-05-10
Genre Technology & Engineering
ISBN 3030706796

Download Variational Methods for Machine Learning with Applications to Deep Networks Book in PDF, Epub and Kindle

This book provides a straightforward look at the concepts, algorithms and advantages of Bayesian Deep Learning and Deep Generative Models. Starting from the model-based approach to Machine Learning, the authors motivate Probabilistic Graphical Models and show how Bayesian inference naturally lends itself to this framework. The authors present detailed explanations of the main modern algorithms on variational approximations for Bayesian inference in neural networks. Each algorithm of this selected set develops a distinct aspect of the theory. The book builds from the ground-up well-known deep generative models, such as Variational Autoencoder and subsequent theoretical developments. By also exposing the main issues of the algorithms together with different methods to mitigate such issues, the book supplies the necessary knowledge on generative models for the reader to handle a wide range of data types: sequential or not, continuous or not, labelled or not. The book is self-contained, promptly covering all necessary theory so that the reader does not have to search for additional information elsewhere. Offers a concise self-contained resource, covering the basic concepts to the algorithms for Bayesian Deep Learning; Presents Statistical Inference concepts, offering a set of elucidative examples, practical aspects, and pseudo-codes; Every chapter includes hands-on examples and exercises and a website features lecture slides, additional examples, and other support material.

Scalable Algorithms for Latent Variable Models in Machine Learning

Scalable Algorithms for Latent Variable Models in Machine Learning
Title Scalable Algorithms for Latent Variable Models in Machine Learning PDF eBook
Author Hsiang-Fu Yu
Publisher
Pages 546
Release 2016
Genre
ISBN

Download Scalable Algorithms for Latent Variable Models in Machine Learning Book in PDF, Epub and Kindle

Latent variable modeling (LVM) is a popular approach in many machine learning applications, such as recommender systems and topic modeling, due to its ability to succinctly represent data, even in the presence of several missing entries. Existing learning methods for LVMs, while attractive, are infeasible for the large-scale datasets required in modern big data applications. In addition, such applications often come with various types of side information such as the text description of items and the social network among users in a recommender system. In this thesis, we present scalable learning algorithms for a wide range of latent variable models such as low-rank matrix factorization and latent Dirichlet allocation. We also develop simple but effective techniques to extend existing LVMs to exploit various types of side information and make better predictions in many machine learning applications such as recommender systems, multi-label learning, and high-dimensional time-series prediction. In addition, we also propose a novel approach for the maximum inner product search problem to accelerate the prediction phase of many latent variable models.

Sampling-based Bayesian Latent Variable Regression Methods with Applications in Process Engineering

Sampling-based Bayesian Latent Variable Regression Methods with Applications in Process Engineering
Title Sampling-based Bayesian Latent Variable Regression Methods with Applications in Process Engineering PDF eBook
Author Hongshu Chen
Publisher
Pages 180
Release 2007
Genre Bayesian statistical decision theory
ISBN

Download Sampling-based Bayesian Latent Variable Regression Methods with Applications in Process Engineering Book in PDF, Epub and Kindle

Abstract: Latent variable methods, such as Principal Component Analysis and Partial Least Squares Regression, can handle collinearity among variables by projecting the original data into a lower dimensional space. They are widely applied to build empirical models of chemical and biological processes. With the development of modern experimental and analytical technology, data sets from those processes are getting bigger and more heterogeneous. The increasing complexity of data sets causes traditional latent variables methods to often fail to provide satisfactory modeling results. Meanwhile, prior information about processes and data usually exist in different sources, such as expert knowledge, historical data etc. However, traditional latent variable methods are ill-suited to incorporate such information. Bayesian latent variable methods, such as Bayesian Latent Variable Regression (BLVR) and Bayesian Principal Component Analysis (BPCA) can combine prior information and data via a rigorous probabilistic framework. Since they make use of more information, they can provide models with better quality. However, BPCA and BLVR are optimization-based, which restricts them from modeling high dimensional data sets or providing error bars. They also make restrictive assumptions to make them suitable for the optimization routines. Because of those pitfalls, they have very limited applications in practice. This dissertation addresses the challenges of making Bayesian latent variable methods practical by developing novel algorithms and a toolbox of sampling-based methods, including a sampling-based BLVR (BLVR-S). BLVR-S is computationally efficient and is able to model high dimensional data sets. It can also readily provide confidence intervals for estimates. An iterative modeling procedure is proposed to deal with hybrid data sets with both continuous and discrete variables. An extended version of BLVR-S is developed to address lack of information about measurement noise in modeling. A generalized BLVR-S is developed to relax the restrictive assumptions of prior distributions. Those methods tackle some practical challenges in Bayesian modeling. The advantages of those Bayesian latent variable regression methods are illustrated in various case studies. Some practical aspects of applying Bayesian latent variable methods are also explored. Through those efforts, the Bayesian latent variable methods are expected to have more practical applications in building empirical models in process engineering.