| dc.contributor.advisor | Patrick H. Winston. | en_US |
| dc.contributor.author | Barnwell, Jake A. (Jake Allen) | en_US |
| dc.contributor.other | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science. | en_US |
| dc.date.accessioned | 2018-12-11T20:39:35Z | |
| dc.date.available | 2018-12-11T20:39:35Z | |
| dc.date.copyright | 2018 | en_US |
| dc.date.issued | 2018 | en_US |
| dc.identifier.uri | http://hdl.handle.net/1721.1/119546 | |
| dc.description | Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2018. | en_US |
| dc.description | This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections. | en_US |
| dc.description | Cataloged from student-submitted PDF version of thesis. | en_US |
| dc.description | Includes bibliographical references (page 63). | en_US |
| dc.description.abstract | If we are to understand how we can build artificial intelligence machines that are able to organically process and acquire knowledge like humans do, we must first understand how to teach a human intelligence system to model and learn generic concepts without resorting to mechanistic hand-coding of such concepts. Such an achievement in human intelligence is an important stride towards the realization of Minsky's hypothetical "baby machine." Genesis is a story-understanding system that reads and analyzes stories in simple English, as humans do. Genesis can recognize certain sequences of events as particular manifestations of some user-defined concepts, such as revenge or physical retaliation. However, both the structure and definition of these high-level concepts must be explicitly provided by the user, and must be exactly matched using concept patterns against sequences of events in a story. I propose that this paradigm is unlike how humans learn such concepts, and instead, such concepts are learned by example, not by explicit definition. To introduce this organic, example-driven paradigm to Genesis, I have created STUDENT, a system that ingests a small series of positive and negative examples of concepts and builds an internal model for these concepts. By aligning similar examples of concepts together, I have constructed refined concept patterns which encompass multiple, different manifestations of the concepts, without requiring a human to explicitly define them. Adding this capability to Genesis enables it to learn concepts by example from a teacher. This behavior emulates how maturing humans learn concepts. Achieving this ability to emulate a human's learning process for a new idea brings us one step closer to modeling human intelligence and story understanding. | en_US |
| dc.description.statementofresponsibility | by Jake A. Barnwell. | en_US |
| dc.format.extent | 63 pages | en_US |
| dc.language.iso | eng | en_US |
| dc.publisher | Massachusetts Institute of Technology | en_US |
| dc.rights | MIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission. | en_US |
| dc.rights.uri | http://dspace.mit.edu/handle/1721.1/7582 | en_US |
| dc.subject | Electrical Engineering and Computer Science. | en_US |
| dc.title | Using near misses to teach concepts to a human intelligence system | en_US |
| dc.type | Thesis | en_US |
| dc.description.degree | M. Eng. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | |
| dc.identifier.oclc | 1076272866 | en_US |