Multiobjective Reinforcement Learning Based Energy Consumption in C-RAN enabled Massive MIMO

S. Sharma, W. Yoon

Multiobjective Reinforcement Learning Based Energy Consumption in C-RAN enabled Massive MIMO

Číslo: 1/2022
Periodikum: Radioengineering Journal
DOI: 10.13164/re.2022.0155

Klíčová slova: Convergence, energy consumption, reinforcement learning, reward, optimization.

Pro získání musíte mít účet v Citace PRO.

Přečíst po přihlášení

Anotace: Multiobjective optimization has become a suitable method to resolve conflicting objectives and enhance the performance evaluation of wireless networks. In this study, we consider a multiobjective reinforcement learning (MORL) approach for the resource allocation and energy consumption in C-RANs. We propose the MORL method with two conflicting objectives. Herein, we define the state and action spaces, and reward for the MORL agent. Furthermore, we develop a Q-learning algorithm that controls the ON-OFF action of remote radio heads (RRHs) depending on the position and nearby users with goal of selecting the best single policy that optimizes the trade-off between EE and QoS. We analyze the performance of our Q-learning algorithm by comparing it with simple ON-OFF scheme and heuristic algorithm. The simulation results demonstrated that normalized ECs of simple ON-OFF, heuristic and Q-learning algorithm were 0.99, 0.85, and 0.8 respectively. Our proposed MORL-based Q-learning algorithm achieves superior EE performance compared with simple ON-OFF scheme and heuristic algorithms.