Show simple item record

dc.contributor.advisorJohn W. Fisher III and Oren Freifeld.en_US
dc.contributor.authorYu, Angelen_US
dc.contributor.otherMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science.en_US
dc.date.accessioned2018-02-08T15:57:49Z
dc.date.available2018-02-08T15:57:49Z
dc.date.copyright2016en_US
dc.date.issued2016en_US
dc.identifier.urihttp://hdl.handle.net/1721.1/113440
dc.descriptionThesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2016.en_US
dc.descriptionThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.en_US
dc.descriptionCataloged from student-submitted PDF version of thesis.en_US
dc.descriptionIncludes bibliographical references (pages 71-72).en_US
dc.description.abstractMachine learning algorithms are often computationally intensive and operate on large datasets. Being able to eciently learn models on large datasets holds the future of machine learning. As the speed of serial computation stalls, it is necessary to utilize the power of parallel computing in order to better scale with the growing complexity of algorithms and the growing size of datasets. In this thesis, we explore the use of Julia, a fairly new high level programming language that lends itself to easy parallelization over multiple CPU cores as well as multiple machines, on Markov chain Monte Carlo (MCMC) inference algorithms. First, we take existing algorithms and implement them in Julia. We focus on MCMC inference using Continuous Piecewise-Affine Based (CPAB) transformations and a parallel MCMC sampler for Dirichlet Process Mixture Models (DPMM). Instead of parallelizing over multiple cores on a single machine, our Julia implementations extend existing implementations by parallelizing over multiple machines. We compare our implementation with these existing implementations written in more traditional programming languages. Next, we develop a model Projections Dirichlet Process Gaussian Mixture Model (PDP-GMM) which relaxes the assumption that the draws from a Dirichlet Process Gaussian Mixture Model (DP-GMM) are directly observed. We extend our DPMM Julia implementation and present a few applications of this model.en_US
dc.description.statementofresponsibilityby Angel Yu.en_US
dc.format.extent72 pagesen_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsMIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectElectrical Engineering and Computer Science.en_US
dc.titleParallel and distributed MCMC inference using Juliaen_US
dc.title.alternativeParallel and distributed Markov chain Monte Carlo inference using Juliaen_US
dc.typeThesisen_US
dc.description.degreeM. Eng.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.identifier.oclc1020069264en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record