Show simple item record

dc.contributor.authorChoi, Myung Jin
dc.contributor.authorWillsky, Alan S.
dc.contributor.authorTan, Vincent Y. F.
dc.contributor.authorAnandkumar, Animashree
dc.date.accessioned2012-06-28T12:30:26Z
dc.date.available2012-06-28T12:30:26Z
dc.date.issued2011-05
dc.date.submitted2011-02
dc.identifier.issn1532-4435
dc.identifier.issn1533-7928
dc.identifier.urihttp://hdl.handle.net/1721.1/71241
dc.description.abstractWe study the problem of learning a latent tree graphical model where samples are available only from a subset of variables. We propose two consistent and computationally efficient algorithms for learning minimal latent trees, that is, trees without any redundant hidden nodes. Unlike many existing methods, the observed nodes (or variables) are not constrained to be leaf nodes. Our algorithms can be applied to both discrete and Gaussian random variables and our learned models are such that all the observed and latent variables have the same domain (state space). Our first algorithm, recursive grouping, builds the latent tree recursively by identifying sibling groups using so-called information distances. One of the main contributions of this work is our second algorithm, which we refer to as CLGrouping. CLGrouping starts with a pre-processing procedure in which a tree over the observed variables is constructed. This global step groups the observed nodes that are likely to be close to each other in the true latent tree, thereby guiding subsequent recursive grouping (or equivalent procedures such as neighbor-joining) on much smaller subsets of variables. This results in more accurate and efficient learning of latent trees. We also present regularized versions of our algorithms that learn latent tree approximations of arbitrary distributions. We compare the proposed algorithms to other methods by performing extensive numerical experiments on various latent tree graphical models such as hidden Markov models and star graphs. In addition, we demonstrate the applicability of our methods on real-world data sets by modeling the dependency structure of monthly stock returns in the S&P index and of the words in the 20 newsgroups data set.en_US
dc.language.isoen_US
dc.publisherCrossRef test prefixen_US
dc.relation.isversionofhttp://jmlr.csail.mit.edu/papers/v12/choi11b.htmlen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceMIT Pressen_US
dc.titleLearning Latent Tree Graphical Modelsen_US
dc.typeArticleen_US
dc.identifier.citationChoi, Myung Jin et al. "Learning Latent Tree Graphical Models" Journal of Machine Learning Research 12 (2011). © JMLR 2011en_US
dc.contributor.departmentMassachusetts Institute of Technology. Laboratory for Information and Decision Systemsen_US
dc.contributor.approverWillsky, Alan S.
dc.contributor.mitauthorChoi, Myung Jin
dc.contributor.mitauthorWillsky, Alan S.
dc.relation.journalJournal of Machine Learning Researchen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dspace.orderedauthorsChoi, Myung Jin; Tan, Vincent Y. F.; Anandkumar, Animashree; Willsky, Alan S.
dc.identifier.orcidhttps://orcid.org/0000-0003-0149-5888
mit.licensePUBLISHER_POLICYen_US
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record