Publication:
Scaling MCMC Inference and Belief Propagation to Large, Dense Graphical Models

dc.contributor.advisorAndrew McCallum
dc.contributor.advisorBenjamin Marlin
dc.contributor.advisorDavid Jensen
dc.contributor.authorSingh, Sameer
dc.contributor.departmentUniversity of Massachusetts Amherst
dc.date2024-03-27 17:03:32
dc.date.accessioned2024-04-26T15:27:57Z
dc.date.available2024-04-26T15:27:57Z
dc.date.issued2014
dc.date.submittedFebruary
dc.date.submitted2014
dc.description.abstractWith the physical constraints of semiconductor-based electronics becoming increasingly limiting in the past decade, single-core CPUs have given way to multi-core and distributed computing platforms. At the same time, access to large data collections is progressively becoming commonplace due to the lowering cost of storage and bandwidth. Traditional machine learning paradigms that have been designed to operate sequentially on single processor architectures seem destined to become obsolete in this world of multi-core, multi-node systems and massive data sets. Inference for graphical models is one such example for which most existing algorithms are sequential in nature and are difficult to scale using parallel computations. Further, modeling large datasets leads to an escalation in the number of variables, factors, domains, and the density of the models, all of which have a substantial impact on the computational and storage complexity of inference. To achieve scalability, existing techniques impose strict independence assumptions on the model, resulting in tractable inference at the expense of expressiveness, and therefore of accuracy and utility, of the model. Motivated by the need to scale inference to large, dense graphical models, in this thesis we explore approximations to Markov chain Monte Carlo (MCMC) and belief propagation (BP) that induce dynamic sparsity in the model to utilize parallelism. In particular, since computations over some factors, variables, and values are more important than over others at different stages of inference, proposed approximations that prioritize and parallelize such computations facilitate efficient inference. First, we show that a synchronously distributed MCMC algorithm that uses dynamic partitioning of the model achieves scalable inference. We then identify bottlenecks in the synchronous architecture, and demonstrate that a collection of MCMC techniques that use asynchronous updates are able to address these drawbacks. For large domains and high-order factors, we find that dynamically inducing sparsity in variable domains, results in scalable belief propagation that enables joint inference. We also show that formulating distributed BP and joint inference as generalized BP on cluster graphs, and by using cluster message approximations, provides significantly lower communication cost and running time. With these tools for inference in hand, we are able to tackle entity tagging, relation extraction, entity resolution, cross-document coreference, joint inference, and other information extraction tasks over large text corpora.
dc.description.degreeDoctor of Philosophy (PhD)
dc.description.departmentComputer Science
dc.identifier.doihttps://doi.org/10.7275/hhc2-rg97
dc.identifier.orcidN/A
dc.identifier.urihttps://hdl.handle.net/20.500.14394/17647
dc.relation.urlhttps://scholarworks.umass.edu/cgi/viewcontent.cgi?article=1105&context=dissertations_2&unstamped=1
dc.source.statuspublished
dc.subjectmachine learning
dc.subjectgraphical models
dc.subjectmarkov chain monte carlo
dc.subjectbelief propagation
dc.subjectapproximate inference
dc.subjectdistributed computing
dc.subjectparallel computing
dc.subjectArtificial Intelligence and Robotics
dc.subjectOther Statistics and Probability
dc.titleScaling MCMC Inference and Belief Propagation to Large, Dense Graphical Models
dc.typeopenaccess
dc.typedissertation
digcom.contributor.authorisAuthorOfPublication|email:sameer@cs.umass.edu|institution:University of Massachusetts Amherst|Singh, Sameer
digcom.identifierdissertations_2/143
digcom.identifier.contextkey5460245
digcom.identifier.submissionpathdissertations_2/143
dspace.entity.typePublication
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
sameer_thesis.pdf
Size:
8.66 MB
Format:
Adobe Portable Document Format
Collections