Your browser doesn't support javascript.
loading
Joint Beamforming, Power Allocation, and Splitting Control for SWIPT-Enabled IoT Networks with Deep Reinforcement Learning and Game Theory.
Liu, JainShing; Lin, Chun-Hung Richard; Hu, Yu-Chen; Donta, Praveen Kumar.
Afiliación
  • Liu J; Department of Computer Science and Information Engineering, Providence University, Taichung 43301, Taiwan.
  • Lin CR; Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung 80424, Taiwan.
  • Hu YC; Department of Computer Science and Information Management, Providence University, Taichung 43301, Taiwan.
  • Donta PK; Research Unit of Distributed Systems, TU Wien, 1040 Vienna, Austria.
Sensors (Basel) ; 22(6)2022 Mar 17.
Article en En | MEDLINE | ID: mdl-35336499
ABSTRACT
Future wireless networks promise immense increases on data rate and energy efficiency while overcoming the difficulties of charging the wireless stations or devices in the Internet of Things (IoT) with the capability of simultaneous wireless information and power transfer (SWIPT). For such networks, jointly optimizing beamforming, power control, and energy harvesting to enhance the communication performance from the base stations (BSs) (or access points (APs)) to the mobile nodes (MNs) served would be a real challenge. In this work, we formulate the joint optimization as a mixed integer nonlinear programming (MINLP) problem, which can be also realized as a complex multiple resource allocation (MRA) optimization problem subject to different allocation constraints. By means of deep reinforcement learning to estimate future rewards of actions based on the reported information from the users served by the networks, we introduce single-layer MRA algorithms based on deep Q-learning (DQN) and deep deterministic policy gradient (DDPG), respectively, as the basis for the downlink wireless transmissions. Moreover, by incorporating the capability of data-driven DQN technique and the strength of noncooperative game theory model, we propose a two-layer iterative approach to resolve the NP-hard MRA problem, which can further improve the communication performance in terms of data rate, energy harvesting, and power consumption. For the two-layer approach, we also introduce a pricing strategy for BSs or APs to determine their power costs on the basis of social utility maximization to control the transmit power. Finally, with the simulated environment based on realistic wireless networks, our numerical results show that the two-layer MRA algorithm proposed can achieve up to 2.3 times higher value than the single-layer counterparts which represent the data-driven deep reinforcement learning-based algorithms extended to resolve the problem, in terms of the utilities designed to reflect the trade-off among the performance metrics considered.
Palabras clave

Texto completo: 1 Banco de datos: MEDLINE Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: Taiwán

Texto completo: 1 Banco de datos: MEDLINE Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: Taiwán