Lemos, Marcelo Luiz Harry Diniz and Vieira, Ronaldo e Silva and Tavares, Anderson Rocha and Marcolino, Leandro Soriano and Chaimowicz, Luiz (2025) Enhancing deep reinforcement learning for scale flexibility in real-time strategy games. Entertainment Computing, 52: 100843. ISSN 1875-9521
ENTCOM_Scale_invariant_Reinforcement_learning.pdf - Accepted Version
Available under License Creative Commons Attribution.
Download (1MB)
Abstract
Real-time strategy (RTS) games present a unique challenge for AI agents due to the combination of several fundamental AI problems. While Deep Reinforcement Learning (DRL) has shown promise in the development of autonomous agents for the genre, existing architectures often struggle with games featuring maps of varying dimensions. This limitation hinders the agent’s ability to generalize its learned strategies across different scenarios. This paper proposes a novel approach that overcomes this problem by incorporating Spatial Pyramid Pooling (SPP) within a DRL framework. We leverage the GridNet architecture’s encoder–decoder structure and integrate an SPP layer into the critic network of the Proximal Policy Optimization (PPO) algorithm. This SPP layer dynamically generates a standardized representation of the game state, regardless of the initial observation size. This allows the agent to effectively adapt its decision-making process to any map configuration. Our evaluations demonstrate that the proposed method significantly enhances the model’s flexibility and efficiency in training agents for various RTS game scenarios, albeit with some discernible limitations when applied to very small maps. This approach paves the way for more robust and adaptable AI agents capable of excelling in sequential decision problems with variable-size observations.