Submitted:
10 June 2025
Posted:
11 June 2025
You are already at the latest version
Abstract
Keywords:
I. Introduction
II. Related Work
III. Methods
- A.
- System model

- B.
- Proposed algorithm flowchart

- C.
- Mathematical development
-
Algorithm t=0Loop
- Depths from 0 to K-1:
-
𝒜[0,1] selects a channel by randomly sampling as per its channel probability vector [p{1,1}(t), p{1,2}(t)]. We denote j1(t) as the chosen channel at depth 0 with j1(t)ϵ{1,2}.𝒜{1, j1(t)}, chooses a channel and activates the next LA at depth «2».The process continues until K-1, which is the level that chooses the channel.
- 2
- Depth K:
- ✧
-
The index of the channel chosen at depth K is denotedjK(t) ϵ {1, . . . 2K}.
- ✧
- Update the estimated chance of reward based on the response received from the environment at leaf depth K:
-
u{K,Jk(t)}(t + 1) = u{K,Jk(t)}(t) + (1 − β(t))v{K,Jk(t)}(t + 1) = v{K,Jk(t)}(t) + 1
- 3
- Define the reward estimate recursively for all subsequent channels along the path to the root, k ϵ {0, . . K − 1}, where 𝒜 at any one level inherits the feedback from the 𝒜 at the level below:
- 4
- Update the channel probability vectors along the path to the leaf with the current maximum reward estimate:
-
Each 𝒜 j ϵ {1, . . . , 2k} at depth k where k ϵ {0, . . . , K − 1} has two channels α {k + 1,2j − 1} and α {k + 1,2j}.We denote the larger element between andand the lower reward estimate as .
- ✧
- Update P{k+1 ,jkℎ+1(t)} and using the estimate and for all
- 5
- For each Learning Automata, if either of its channel selection probabilities surpasses a threshold T, with T being a positive number close to unity, the channel probability will stop updating, meaning the convergence is achieved.
- 6
- t= t+1
- End Loop
- D.
- Software environment
- E.
- System simulation
- F.
- Simulation Variable
| Variable | Symbol | Description |
| Number of channels |
N | Total number of available channels in the network. |
| Initial channel probability | P(0) | Initial probability vector for channel selection, P(0) = [p1, p2, … , pN] |
| Reward | R | Reward metric for successful transmission on a channel, such as PDR, SNR. |
| Learning Rate | δ | Step size for probability updates. |
| Hierarchical levels | L | Number of levels in the HDPA hierarchy. |
| Convergence threshold | B | The threshold for convergence, indicating when the algorithm has likely found the optimal channel. |
| Maximum iteration |
T | Maximum number of iterations for the simulation. |
| Action selection probability | Pi | Probability of selecting channels at iteration. |
| Reward estimate | di | Estimate of the reward for channel i. |
| Channel State | S | The state of each channel is either idle or busy. |
IV. Results and Discussion
| A1 | A2 | A3 | A4 | A5 | A6 | A7 | A8 |
| 0.199 | 0.282 | 0.394 | 0.499 | 0.681 | 0.698 | 0.971 | 0.999 |
| Parameter | HDPA |
|---|---|
| Mean | 6,279.64 |
| Std | 131.36 |
| Accuracy | 98.78% |
| Learning parameter | 8.7e−4 |


| Parameters | HDPA | HCPA |
|---|---|---|
| Mean | 6,279.64 | 6,778.34 |
| STD | 131.36 | 117.12 |
| Accuracy | 98.78% | 93.89% |
| Learning parameter | 8.7e−4 | 6.9e−4 |
V. Conclusion, and Future Work
- A.
- Conclusion
- B.
- Future works
References
- I. Cheikh, E. Sabir, R. Aouami, M. Sadik and S. Roy, "Throughput-Delay Tradeoffs for Slotted-Aloha-based LoRaWAN Networks. in 2021 International Wireless Communications and Mobile Computing (IWCMC), Harbin City, China, 2021. [CrossRef]
- Honggang Wang,Peidong Pei,Ruoyu Pan,Kai Wu,Yu Zhang,Jinchao Xiao and Jingfeng Yang, "A Collision Reduction Adaptive Data Rate Algorithm Based on the FSVM for a Low-Cost LoRa Gateway. Mathematics 2022, 10, 3920. [CrossRef]
- X. Zhang, L. X. Zhang, L. Jiao, O.-C. Granmo and B. J. Oommen, "Channel selection in cognitive radio networks: A switchable Bayesian learning automata approach. in IEEE 24th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC). IEEE 2013. [CrossRef]
- DIANE, Ass, DIALLO, Ousmane, et NDOYE, El Hadji Malick. A systematic and comprehensive review on low power wide area network: characteristics, architecture, applications and research challenges. Discover Internet of Things 2025, 5, 7. [Google Scholar] [CrossRef]
- R. C. a. Y. J. Hui Bai, "Evolutionary reinforcement learning: A survey. Intelligent Computing 2023, 2, 0025. [Google Scholar] [CrossRef]
- R. O. Omslandseter, L. Jiao, X. Zhang, A. Yazidi and B. J. Oommen, "The hierarchical discrete pursuit learning automaton: a novel scheme with fast convergence and epsilon-optimality. IEEE Transactions on Neural Networks and Learning System 2022, 35, 8278–8292. [Google Scholar] [CrossRef]
- A. Yazidi, X. Zhang, L. Jiao and B. J. Oommen, "The hierarchical continuous pursuit learning automation: a novel scheme for environments with large numbers of actions. IEEE transactions on neural networks and learning systems 2019, 31, 512–526. [CrossRef] [PubMed]
- A. Prakash, N. Choudhury, A. Hazarika and A. Gorrela, "Effective Feature Selection for Predicting Spreading Factor with ML in Large LoRaWAN-based Mobile IoT Networks. in 2025 National Conference on Communications (NCC), New Delhi, India, 2025. [CrossRef]
- S. Lavdas, N. S. Lavdas, N. Bakas, K. Vavousis, Ala’Khalifeh, W. E. Hajj and Z. Zinonos, "Evaluating LoRaWAN Network Performance in Smart City Environments Using Machine Learning. IEEE Internet of Things Journal, 2025; 1–1. [Google Scholar] [CrossRef]
- D. Garlisi, A. Pagano, F. Giuliano, D. Croce and I. Tinnirello, "Interference Analysis of LoRaWAN and Sigfox in Large-Scale Urban IoT Networks. IEEE Access 2025, 13, 44836–44848. [CrossRef]
- G. M. E. R. a. A. W. Hossein Keshmiri, "LoRa Resource Allocation Algorithm for Higher Data Rates. Sensors 2025, 25, 518. [CrossRef]
- A. Li, M. Fujisawa, I. Urabe, R. Kitagawa, S.-J. Kim and M. Hasegawa, "A lightweight decentralized reinforcement learning based channel selection approach for high-density LoRaWAN. in 2021 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), Los Angeles, CA, USA, 2021. [CrossRef]
- Stephen, S. Oyewobi,,Gerhard P. Hancke,Adnan M. Abu-Mahfouz andAdeiza J. Onumanyi, "An effective spectrum handoff based on reinforcement learning for target channel selection in the industrial Internet of Things. Sensors 2019, 19, 1395. [Google Scholar] [CrossRef]
- S. Hasegawa, S.-J. S. Hasegawa, S.-J. Kim, Y. Shoji and M. Hasegawa, "Performance evaluation of machine learning based channel selection algorithm implemented on IoT sensor devices in coexisting IoT networks. in 2020 IEEE 17th Annual Consumer Communications & Networking Conference (CCNC)., Las Vegas, NV, USA, 2020. [CrossRef]
- F. Loh, N. Mehling, S. Geißler and T. Hoßfeld, "Simulative performance study of slotted aloha for lorawan channel access. in NOMS 2022-2022 IEEE/IFIP Network Operations and Management Symposium, Budapest, Hungary, 2022. [CrossRef]
- L. Yurii, L. L. Yurii, L. Anna and S. Stepan, "Research on the Throughput Capacity of LoRaWAN Communication Channel. in 2023 IEEE East-West Design & Test Symposium (EWDTS), Batumi, Georgia, 2023. [CrossRef]
- G. Gaillard and C. Pham, "CANL LoRa: Collision Avoidance by Neighbor Listening for Dense LoRa Networks. in 2023 IEEE Symposium on Computers and Communications (ISCC), Gammarth, Tunisia, 2023. [CrossRef]


Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).