Think! Evidence

Learnability, representation, and language : a Bayesian approach

Show simple item record

dc.contributor Joshua Tenenbaum.
dc.contributor Massachusetts Institute of Technology. Dept. of Brain and Cognitive Sciences.
dc.contributor Massachusetts Institute of Technology. Dept. of Brain and Cognitive Sciences.
dc.creator Perfors, Amy (Amy Francesca)
dc.date 2009-06-25T20:33:06Z
dc.date 2009-06-25T20:33:06Z
dc.date 2008
dc.date 2008
dc.identifier http://hdl.handle.net/1721.1/45601
dc.identifier 315893755
dc.description Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Brain and Cognitive Sciences, 2008.
dc.description This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
dc.description Includes bibliographical references (p. 225-243).
dc.description Within the metaphor of the "mind as a computation device" that dominates cognitive science, understanding human cognition means understanding learnability not only what (and how) the brain learns, but also what data is available to it from the world. Ideal learnability arguments seek to characterize what knowledge is in theory possible for an ideal reasoner to acquire, which illuminates the path towards understanding what human reasoners actually do acquire. The goal of this thesis is to exploit recent advances in machine learning to revisit three common learnability arguments in language acquisition. By formalizing them in Bayesian terms and evaluating them given realistic, real-world datasets, we achieve insight about what must be assumed about a child's representational capacity, learning mechanism, and cognitive biases. Exploring learnability in the context of an ideal learner but realistic (rather than ideal) datasets enables us to investigate what could be learned in practice rather than noting what is impossible in theory. Understanding how higher-order inductive constraints can themselves be learned permits us to reconsider inferences about innate inductive constraints in a new light. And realizing how a learner who evaluates theories based on a simplicity/goodness-of-fit tradeoff can handle sparse evidence may lead to a new perspective on how humans reason based on the noisy and impoverished data in the world. The learnability arguments I consider all ultimately stem from the impoverishment of the input either because it lacks negative evidence, it lacks a certain essential kind of positive evidence, or it lacks suffcient quantity of evidence necessary for choosing from an infinite set of possible generalizations.
dc.description (cont.) I focus on these learnability arguments in the context of three major topics in language acquisition: the acquisition of abstract linguistic knowledge about hierarchical phrase structure, the acquisition of verb argument structures, and the acquisition of word leaning biases.
dc.description by Amy Perfors.
dc.description Ph.D.
dc.format 243 p.
dc.format application/pdf
dc.language eng
dc.publisher Massachusetts Institute of Technology
dc.rights M.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission.
dc.rights http://dspace.mit.edu/handle/1721.1/7582
dc.subject Brain and Cognitive Sciences.
dc.title Learnability, representation, and language : a Bayesian approach
dc.type Thesis


Files in this item

Files Size Format View
315893755-MIT.pdf 1.790Mb application/pdf View/Open

Files in this item

Files Size Format View
315893755-MIT.pdf 1.790Mb application/pdf View/Open

Files in this item

Files Size Format View
315893755-MIT.pdf 1.790Mb application/pdf View/Open

This item appears in the following Collection(s)

Show simple item record

Search Think! Evidence


Browse

My Account