Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Reinforcement Learning-based Wi-Fi Contention Window Optimization

Version 1 : Received: 29 October 2022 / Approved: 1 November 2022 / Online: 1 November 2022 (02:39:05 CET)

How to cite: Cruz, S.C.D.S.; Ahmed Ouameur, M.; Figueiredo, F.A.P.D. Reinforcement Learning-based Wi-Fi Contention Window Optimization. Preprints 2022, 2022110011. https://doi.org/10.20944/preprints202211.0011.v1 Cruz, S.C.D.S.; Ahmed Ouameur, M.; Figueiredo, F.A.P.D. Reinforcement Learning-based Wi-Fi Contention Window Optimization. Preprints 2022, 2022110011. https://doi.org/10.20944/preprints202211.0011.v1

Abstract

The collision avoidance mechanism adopted by the IEEE 802.11 standard is not optimal. The mechanism employs a binary exponential backoff (BEB) algorithm in the medium access control (MAC) layer. Such an algorithm increases the backoff interval whenever a collision is detected to minimize the probability of subsequent collisions. However, the expansion of the backoff interval causes degradation of the radio spectrum utilization (i.e., bandwidth wastage). That problem worsens when the network has to manage the channel access to a dense number of stations, leading to a dramatic decrease in network performance. Furthermore, a wrong backoff setting increases the probability of collisions such that the stations experience numerous collisions before achieving the optimal backoff value. Therefore, to mitigate bandwidth wastage and, consequently, maximize the network performance, this work proposes using reinforcement learning (RL) algorithms, namely Deep Q Learning (DQN) and Deep Deterministic Policy Gradient (DDPG), to tackle such an optimization problem. As for the simulations, the NS-3 network simulator is used along with a toolkit known as NS3-gym, which integrates a reinforcement-learning (RL) framework into NS-3. The results demonstrate that DQN and DDPG have much better performance than BEB for static and dynamic scenarios, regardless of the number of stations. Moreover, the performance difference is amplified as the number of stations increases, with DQN and DDPG showing a 27% increase in throughput with 50 stations compared to BEB. Furthermore, DQN and DDPG presented similar performances.

Keywords

Wi-Fi; contention-based access scheme; channel utilization optimization; machine learning; reinforcement learning; NS-3, NS3-gym

Subject

Computer Science and Mathematics, Mathematics

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.