Show simple item record

dc.contributor.advisorJoshua Tenenbaum.en_US
dc.contributor.authorPerfors, Amy (Amy Francesca)en_US
dc.contributor.otherMassachusetts Institute of Technology. Dept. of Brain and Cognitive Sciences.en_US
dc.date.accessioned2009-06-25T20:33:06Z
dc.date.available2009-06-25T20:33:06Z
dc.date.copyright2008en_US
dc.date.issued2008en_US
dc.identifier.urihttp://hdl.handle.net/1721.1/45601
dc.descriptionThesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Brain and Cognitive Sciences, 2008.en_US
dc.descriptionThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.en_US
dc.descriptionIncludes bibliographical references (p. 225-243).en_US
dc.description.abstractWithin the metaphor of the "mind as a computation device" that dominates cognitive science, understanding human cognition means understanding learnability not only what (and how) the brain learns, but also what data is available to it from the world. Ideal learnability arguments seek to characterize what knowledge is in theory possible for an ideal reasoner to acquire, which illuminates the path towards understanding what human reasoners actually do acquire. The goal of this thesis is to exploit recent advances in machine learning to revisit three common learnability arguments in language acquisition. By formalizing them in Bayesian terms and evaluating them given realistic, real-world datasets, we achieve insight about what must be assumed about a child's representational capacity, learning mechanism, and cognitive biases. Exploring learnability in the context of an ideal learner but realistic (rather than ideal) datasets enables us to investigate what could be learned in practice rather than noting what is impossible in theory. Understanding how higher-order inductive constraints can themselves be learned permits us to reconsider inferences about innate inductive constraints in a new light. And realizing how a learner who evaluates theories based on a simplicity/goodness-of-fit tradeoff can handle sparse evidence may lead to a new perspective on how humans reason based on the noisy and impoverished data in the world. The learnability arguments I consider all ultimately stem from the impoverishment of the input either because it lacks negative evidence, it lacks a certain essential kind of positive evidence, or it lacks suffcient quantity of evidence necessary for choosing from an infinite set of possible generalizations.en_US
dc.description.abstract(cont.) I focus on these learnability arguments in the context of three major topics in language acquisition: the acquisition of abstract linguistic knowledge about hierarchical phrase structure, the acquisition of verb argument structures, and the acquisition of word leaning biases.en_US
dc.description.statementofresponsibilityby Amy Perfors.en_US
dc.format.extent243 p.en_US
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsM.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582en_US
dc.subjectBrain and Cognitive Sciences.en_US
dc.titleLearnability, representation, and language : a Bayesian approachen_US
dc.typeThesisen_US
dc.description.degreePh.D.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciences
dc.identifier.oclc315893755en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record