Modelling learning and decision making under information processing constraints

Malloy, Tyler
Thumbnail Image
Other Contributors
Sims, Christopher Robert
Fajen, Brett R.
Van Heuveln, Bram
Issue Date
Cognitive science
Terms of Use
Attribution-NonCommercial-NoDerivs 3.0 United States
This electronic version is a licensed copy owned by Rensselaer Polytechnic Institute, Troy, NY. Copyright of original work retained by author.
Full Citation
This thesis explores the impact of information processing constraints on models of human learning and decision making. This is achieved through altering existing methods within the fields of economic decision making and reinforcement learning, with inspiration from information theory and the rational inattention economic framework. The result is two models, one of human decision making and one of human learning, which seek to represent the way that differences in individual information processing abilities impact learning and decision making. Data from human responses in a learning task is used to compare the accuracy of this model against existing methods. Results from experimentation show that these models achieve a high degree of accuracy while accounting for the impact of differences in information processing capacity utilized by participants during the learning task. These results further the understanding of how cognitive limitations impact human learning and decision making, and suggest ways in which this type of model could potentially benefit current approaches in artificial intelligence, by incorporating more human-like learning strategies.
December 2020
School of Humanities, Arts, and Social Sciences
Dept. of Cognitive Science
Rensselaer Polytechnic Institute, Troy, NY
Rensselaer Theses and Dissertations Online Collection
CC BY-NC-ND. Users may download and share copies with attribution in accordance with a Creative Commons Attribution-Noncommercial-No Derivative Works 3.0 License. No commercial use or derivatives are permitted without the explicit approval of the author.