Advancements in Deep Reinforcement Learning: A Comprehensive Survey on Policy Optimization Techniques
DOI:
https://doi.org/10.63282/3050-9246.IJETCSIT-V1I2P101Keywords:
Deep Reinforcement Learning, Policy Optimization, Policy Gradient Methods, Actor-Critic, Trust Region Methods, Proximal Policy Optimization, Model-Based Reinforcement Learning, Sample Efficiency, Exploration Strategies, Generalization in DRLAbstract
Deep Reinforcement Learning (DRL) has emerged as a powerful paradigm for solving complex decision-making problems in various domains, including robotics, gaming, and autonomous systems. At the core of DRL lies the optimization of policies that map states to actions, enabling agents to learn optimal behaviors through interaction with their environment. This paper provides a comprehensive survey of recent advancements in policy optimization techniques in DRL. We categorize and discuss the key methods, including policy gradient methods, actor-critic algorithms, and model-based approaches. We also explore the challenges and future directions in the field, highlighting the integration of DRL with other machine learning techniques and the application of DRL in real-world scenarios. The paper aims to serve as a valuable resource for researchers and practitioners interested in the latest developments in DRL
Downloads
References
[1] Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction. MIT Press.
[2] Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., & Silver, D. (2015). Continuous control with deep
reinforcement learning. arXiv preprint arXiv:1509.02971.
[3] Schulman, J., Levine, S., Abbeel, P., Jordan, M., & Moritz, P. (2015). Trust region policy optimization. In International
Conference on Machine Learning (pp
[4] https://escholarship.org/content/qt9z908523/qt9z908523.pdf?t=otc2ko
[5] https://par.nsf.gov/servlets/purl/10321727
[6] https://arxiv.org/html/2502.06869v1
[7] https://jmlr.org/papers/volume20/18-476/18-476.pdf
[8] https://www.mdpi.com/1424-8220/23/7/3762
_in_Go
[10] https://ieeexplore.ieee.org/document/8103164
_and_emerging_trends