LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

Efficient Exploration Through Bootstrapped and Bayesian Deep Q-Networks for Joint Power Control and Beamforming in mmWave Networks

Photo by ldxcreative from unsplash

Efficient exploration is critical in deploying Deep Reinforcement Learning (DRL) for joint power control and beamforming in mmWave networks. This letter proposes Bootstrapped and Bayesian Deep Q-Network (B2DQN), a DRL… Click to show full abstract

Efficient exploration is critical in deploying Deep Reinforcement Learning (DRL) for joint power control and beamforming in mmWave networks. This letter proposes Bootstrapped and Bayesian Deep Q-Network (B2DQN), a DRL algorithm based on both Bootstrap Sampling (BS) and Thompson Sampling (TS). BS induces diversity to prevent exploration from being trapped in local optima, and TS allows for targeted exploration of actions with higher estimated returns with higher probability. B2DQN synthesizes the advantages of BS and TS by building a Bayesian linear regression model on each bootstrapped function to improve the diversity of BS and enhance targeted exploration with the diversity. The experiment results demonstrate that B2DQN outperforms the DQN variants based only on BS or TS, often learning policies that attain 78%-360% higher final performance measured by the convergence episodes in an open-source simulated 5G mmWave network.

Keywords: efficient exploration; power control; beamforming mmwave; control beamforming; joint power; exploration

Journal Title: IEEE Communications Letters
Year Published: 2023

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.