Novel deep reinforcement learning algorithms for continuous control
buir.advisor | Kozat, Süleyman Serdar | |
dc.contributor.author | Sağlam, Baturay | |
dc.date.accessioned | 2023-07-06T11:03:55Z | |
dc.date.available | 2023-07-06T11:03:55Z | |
dc.date.copyright | 2023-06 | |
dc.date.issued | 2023-06 | |
dc.date.submitted | 2023-06-23 | |
dc.description | Cataloged from PDF version of article. | |
dc.description | Thesis (Master's): Bilkent University, Department of Electrical and Electronics Engineering, İhsan Doğramacı Bilkent University, 2023. | |
dc.description | Includes bibliographical references (leaves 67-74). | |
dc.description.abstract | Continuous control deep reinforcement learning (RL) algorithms are capable of learning complex and high-dimensional policies directly from raw sensory inputs. However, they often face challenges related to sample efficiency and exploration, which limit their practicality for real-world applications. In light of this, we introduce two novel techniques that enhance the performance of continuous control deep RL algorithms by refining their experience replay and exploration mechanisms. The first technique introduces a novel framework for sampling experiences in actor-critic methods. Specifically designed to stabilize and prevent divergence caused by Prioritized Experience Replay (PER), our framework effectively trains both actor and critic networks by striking a balance between temporal-difference error and policy gradient. Through both theoretical analysis and empirical investigations, we demonstrate that our framework is effective in improving the performance of continuous control deep RL algorithms. The second technique encompasses a directed exploration strategy that relies on intrinsic motivation. Drawing inspiration from established theories on animal motivational systems and adapting them to the actor-critic setting, our strategy showcases its effectiveness by generating exploratory behaviors that are both informative and diverse. It achieves this by maximizing the error of the value function and unifying the ex-isting intrinsic exploration objectives in the literature. We evaluate the presented methods on various continuous control benchmarks and demonstrate that they outperform state-of-the-art methods while achieving new levels of performance in deep RL. | |
dc.description.provenance | Made available in DSpace on 2023-07-06T11:03:55Z (GMT). No. of bitstreams: 1 B162166.pdf: 47730379 bytes, checksum: 6f6e44fcf070b9da3148a8907badd615 (MD5) Previous issue date: 2023-06 | en |
dc.description.statementofresponsibility | by Baturay Sağlam | |
dc.format.extent | xv, 85 leaves : illustrations ; 30 cm. | |
dc.identifier.itemid | B162166 | |
dc.identifier.uri | https://hdl.handle.net/11693/112370 | |
dc.language.iso | English | |
dc.rights | info:eu-repo/semantics/openAccess | |
dc.subject | Deep reinforcement learning | |
dc.subject | Continuous control | |
dc.subject | Off-policy learning | |
dc.subject | Exploitation-exploration | |
dc.title | Novel deep reinforcement learning algorithms for continuous control | |
dc.title.alternative | Sürekli kontrol için yeni derin pekiştirmeli öğrenme algoritmaları | |
dc.type | Thesis | |
thesis.degree.discipline | Electrical and Electronic Engineering | |
thesis.degree.grantor | Bilkent University | |
thesis.degree.level | Master's | |
thesis.degree.name | MS (Master of Science) |