Mutual Reinforcement Learning

dc.contributor.advisorMukhopadhyay, Snehasis
dc.contributor.authorReid, Cameron
dc.contributor.otherMohler, George
dc.contributor.otherTuceryan, Mihran
dc.date.accessioned2021-05-18T12:28:05Z
dc.date.available2021-05-18T12:28:05Z
dc.date.issued2021-05
dc.degree.date2021en_US
dc.degree.disciplineComputer & Information Science
dc.degree.grantorPurdue Universityen_US
dc.degree.levelM.S.en_US
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en_US
dc.description.abstractMutual learning is an emerging field in intelligent systems which takes inspiration from naturally intelligent agents and attempts to explore how agents can communicate and coop- erate to share information and learn more quickly. While agents in many biological systems have little trouble learning from one another, it is not immediately obvious how artificial agents would achieve similar learning. In this thesis, I explore how agents learn to interact with complex systems. I further explore how these complex learning agents may be able to transfer knowledge to one another to improve their learning performance when they are learning together and have the power of communication. While significant research has been done to explore the problem of knowledge transfer, the existing literature is concerned ei- ther with supervised learning tasks or relatively simple discrete reinforcement learning. The work presented here is, to my knowledge, the first which admits continuous state spaces and deep reinforcement learning techniques. The first contribution of this thesis, presented in Chapter 2, is a modified version of deep Q-learning which demonstrates improved learning performance due to the addition of a mutual learning term which penalizes disagreement between mutually learning agents. The second contribution, in Chapter 3, is a presentation work which describes effective communication of agents which use fundamentally different knowledge representations and systems of learning (model-free deep Q learning and model- based adaptive dynamic programming), and I discuss how the agents can mathematically negotiate their trust in one another to achieve superior learning performance. I conclude with a discussion of the promise shown by this area of research and a discussion of problems which I believe are exciting directions for future research.en_US
dc.identifier.urihttps://hdl.handle.net/1805/25957
dc.identifier.urihttp://dx.doi.org/10.7912/C2/11
dc.language.isoen_USen_US
dc.rightsCC0 1.0 Universal*
dc.rights.urihttps://creativecommons.org/publicdomain/zero/1.0*
dc.subjectreinforcement learningen_US
dc.subjectartificial intelligenceen_US
dc.subjectmachine learningen_US
dc.titleMutual Reinforcement Learningen_US
dc.typeThesisen
thesis.degree.disciplineComputer & Information Scienceen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
thesis.pdf
Size:
2.86 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.99 KB
Format:
Item-specific license agreed upon to submission
Description: