高级检索

    蔡林逸, 冯翔, 虞慧群. 基于组稀疏优化的强化学习稀疏表征[J]. 华东理工大学学报(自然科学版). DOI: 10.14135/j.cnki.1006-3080.20231223001
    引用本文: 蔡林逸, 冯翔, 虞慧群. 基于组稀疏优化的强化学习稀疏表征[J]. 华东理工大学学报(自然科学版). DOI: 10.14135/j.cnki.1006-3080.20231223001
    CAI Linyi, FENG Xiang, YU Huiqun. Reinforcement Learning with Sparse Representation via Sparse Overlapping Group Lasso[J]. Journal of East China University of Science and Technology. DOI: 10.14135/j.cnki.1006-3080.20231223001
    Citation: CAI Linyi, FENG Xiang, YU Huiqun. Reinforcement Learning with Sparse Representation via Sparse Overlapping Group Lasso[J]. Journal of East China University of Science and Technology. DOI: 10.14135/j.cnki.1006-3080.20231223001

    基于组稀疏优化的强化学习稀疏表征

    Reinforcement Learning with Sparse Representation via Sparse Overlapping Group Lasso

    • 摘要: 由于强化学习具有出色的数据效率和快速学习的能力,许多实际问题开始应用其学习复杂策略。但是高维环境中的强化学习常常受限于维度灾难或者灾难性干扰,性能表现不佳甚至学习失败。围绕表征学习,提出了一种符合Lasso类型优化的稀疏卷积深度强化学习方法。首先,对稀疏表征的理论和优势进行综述,从而创新性地将稀疏卷积方法引入深度强化学习中,提出了一种新的稀疏表征方法。其次,对由稀疏卷积编码定义的可微优化层进行了数学推导并给出了优化算法。此外,为了验证新的稀疏表征方法的有效性,将其应用于相关文献中常见的基准环境中进行测试。实验结果表明,应用稀疏卷积编码的算法具有更好的性能和鲁棒性,在降低了50%以上的模型开销的前提下,取得了相当甚至更优的性能。此外,还研究了稀疏程度对算法性能的影响,结果显示适当的稀疏度能获得更优的性能。

       

      Abstract: Due to the outstanding data efficiency and learning rapidly, many real-world problems have begun to apply reinforcement learning to learn complex strategies. However, reinforcement learning in high-dimensional environments is often limited by the curse of dimensionality or catastrophic interference, leading to bad performance or even failure. This paper proposes a sparse convolutional deep reinforcement learning method around representation learning, which conforming to Lasso-type optimization. First, the theory and advantages of sparse representation are reviewed, and the sparse convolution method is innovatively introduced into deep reinforcement learning, proposing a new sparse representation method. Second, this paper derives the differentiable optimization layer defined by sparse convolutional encoding mathematically and give an optimization algorithm. In addition, in order to verify the effectiveness of the new sparse representation method, we tested it in benchmark environments commonly considered within the related literature. Experimental results show that the algorithm using sparse convolutional encoding has better performance and robustness, achieving comparable or even superior performance with more than 50% reduction in model cost. In addition, we also studied the impact of sparsity on algorithm performance, and the results show that an appropriate degree of sparsity can achieve better performance.

       

    /

    返回文章
    返回