Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01pg15bh93t
Title: A Memory-Augmented Neural Network Model of Abstract Rule Learning
Authors: Sinha, Ishan
Advisors: Cohen, Jonathan
Department: Computer Science
Class Year: 2020
Abstract: The ability to extrapolate knowledge from familiar to novel domains is a defining feature of human intelligence. Contemporary neural network techniques, however, are primarily limited to interpolation among data in their training experience. In this work, we focus on neural networks’ capacity for arbitrary role-filler binding, the ability to associate abstract “roles” to context-specific “fillers,” which is a capacity that many have argued is an important mechanism underlying the ability to extrapolate. Using a simplified version of Raven’s Progressive Matrices, a hallmark test of human intelligence, we introduce a sequential formulation of a visual problem-solving task that requires this form of binding. Further, we introduce the Arbitrary Binding Network, a recurrent neural network model augmented with an external memory, and empirically demonstrate that it successfully learns the underlying abstract rule structure of our task and perfectly generalizes this rule structure to novel fillers.
URI: http://arks.princeton.edu/ark:/88435/dsp01pg15bh93t
Type of Material: Princeton University Senior Theses
Language: en
Appears in Collections:Computer Science, 1988-2020

Files in This Item:
File Description SizeFormat 
SINHA-ISHAN-THESIS.pdf5.41 MBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.