Decentralized and Partially Decentralized Multi-Agent Reinforcement Learning

dc.contributor.advisorMukhopadhyay, Snehasis
dc.contributor.authorTilak, Omkar Jayant
dc.contributor.otherSi, Luo
dc.contributor.otherNeville, Jennifer
dc.contributor.otherRaje, Rajeev
dc.contributor.otherTuceryan, Mihran
dc.contributor.otherGorman, William J.
dc.date.accessioned2013-08-22T19:42:15Z
dc.date.available2013-08-22T19:42:15Z
dc.date.issued2013-08-22
dc.degree.date2012en_US
dc.degree.disciplineDepartment of Computer and Information Scienceen_US
dc.degree.grantorPurdue Universityen_US
dc.degree.levelPh.D.en_US
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en_US
dc.description.abstractMulti-agent systems consist of multiple agents that interact and coordinate with each other to work towards to certain goal. Multi-agent systems naturally arise in a variety of domains such as robotics, telecommunications, and economics. The dynamic and complex nature of these systems entails the agents to learn the optimal solutions on their own instead of following a pre-programmed strategy. Reinforcement learning provides a framework in which agents learn optimal behavior based on the response obtained from the environment. In this thesis, we propose various novel de- centralized, learning automaton based algorithms which can be employed by a group of interacting learning automata. We propose a completely decentralized version of the estimator algorithm. As compared to the completely centralized versions proposed before, this completely decentralized version proves to be a great improvement in terms of space complexity and convergence speed. The decentralized learning algorithm was applied; for the first time; to the domains of distributed object tracking and distributed watershed management. The results obtained by these experiments show the usefulness of the decentralized estimator algorithms to solve complex optimization problems. Taking inspiration from the completely decentralized learning algorithm, we propose the novel concept of partial decentralization. The partial decentralization bridges the gap between the completely decentralized and completely centralized algorithms and thus forms a comprehensive and continuous spectrum of multi-agent algorithms for the learning automata. To demonstrate the applicability of the partial decentralization, we employ a partially decentralized team of learning automata to control multi-agent Markov chains. More flexibility, expressiveness and flavor can be added to the partially decentralized framework by allowing different decentralized modules to engage in different types of games. We propose the novel framework of heterogeneous games of learning automata which allows the learning automata to engage in disparate games under the same formalism. We propose an algorithm to control the dynamic zero-sum games using heterogeneous games of learning automata.en_US
dc.identifier.urihttps://hdl.handle.net/1805/3462
dc.identifier.urihttp://dx.doi.org/10.7912/C2/2305
dc.language.isoen_USen_US
dc.subjectLearning Automata, MARL, Decentralized Learningen_US
dc.subject.lcshComputer systems -- Analysisen_US
dc.subject.lcshMarkov processesen_US
dc.subject.lcshComputer gamesen_US
dc.subject.lcshFuzzy algorithmsen_US
dc.subject.lcshComputer simulationen_US
dc.subject.lcshSwarm intelligence -- Analysisen_US
dc.subject.lcshMachine theoryen_US
dc.subject.lcshComputational complexityen_US
dc.subject.lcshIntelligent agents (Computer software)en_US
dc.titleDecentralized and Partially Decentralized Multi-Agent Reinforcement Learningen_US
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Omkar_Tilak_ProQuestApprovedThesis.pdf
Size:
2.39 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.88 KB
Format:
Item-specific license agreed upon to submission
Description: