Submitted:
09 June 2025
Posted:
10 June 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
- A robust and non-contact driver monitoring framework is developed based on mmWave radar, featuring strong resistance to interference and enhanced privacy protection. Unlike conventional methods that focus solely on either behavioral or physiological data, the proposed system integrates both types of information through parallel processing pipelines, enabling more comprehensive driver state perception.
- A deep learning architecture named Radar-based Temporal-Spatio Fusion Network (RTSFN) is proposed, which effectively fuses temporal and spatial features extracted from radar signals to identify risky driving behaviors.
- The system is designed with a balance between detection accuracy and computational efficiency, allowing real-time inference on edge devices and supporting practical deployment in in-vehicle environments.
2. FMCW Radar Principles and Hardware Setup
2.1. Overview of FMCW Millimeter-Wave Radar Technology
2.1.1. Range Estimation
2.1.2. Velocity Estimation
2.1.3. Angle Estimation
2.2. Radar Hardware Setup and Configuration
3. System Overview
4. Human Motion Detection: RTSFN-based Driver Action Detection
4.1. Input Data of Radar Signals
4.2. Input Data Pre-Processing
4.3. Radar-based Temporal-Spatio Fusion Network (RTSFN)
- Temporal Encoder (Gated TCN): Captures long-range motion patterns from Range-Doppler sequences using time-distributed convolution and gated temporal modeling.
- Spatial Encoder (SE-CNN): Encodes static posture features from Range Profile and Noise Profile inputs via squeeze-and-excitation enhanced convolutional layers.
- Cross-Gated Fusion: Enhances the complementarity between temporal and spatial features using shared gating mechanisms and residual integration.
- Adapter Module: Applies residual transformation and dimensional compression to stabilize training and improve feature expressiveness.
- Multi-Task Output: Outputs results for two tasks — binary risk detection and multi-class classification of seven distinct driving behaviors.
4.3.1. Temporal Modality Encoder
4.3.2. Spatial Modality Encoder
4.3.3. Feature Buffering Mechanism
4.3.4. Spatial-Temporal Fusion Module
4.4. Dataset Labeling
4.5. Model Training
- Task-Specific Losses: We compute cross-entropy losses for both classification heads:where y is the ground truth label and is the predicted probability.
- Uncertainty-Based Weighting: Inspired by the work of Kendall et al. [14], we introduce two learnable log-variance parameters and to adaptively balance the tasks. The uncertainty-weighted total task loss is:
- Cosine Alignment Loss (Optional): To encourage consistent feature representations across modalities (e.g., range and Doppler), we apply a cosine similarity penalty between the latent vectors and :
5. Physiological Signals Monitoring: Heart and Respiration Signal Analysis
5.1. Heart Rate Monitoring
5.1.1. Construct Phase Matrix from TLV
5.1.2. Unwrap Phase and Select Optimal Range Bin
5.1.3. Phase Signal Denoising and Template Matching
5.1.4. Envelope Extraction and Bandpass Filtering
5.1.5. Chirp Z-Transform Frequency Estimation
5.1.6. Heart Rate Calculation
5.2. Respiration Rate Monitoring
5.2.1. Phase Matrix Construction from TLV:
5.2.2. Phase Unwrapping and Averaging:
5.2.3. Low-Rank and Sparse Decomposition (RPCA):
5.2.4. Sparse Spectrum Recovery via FISTA:
5.2.5. Respiration Frequency Estimation and BPM Calculation:
6. Results
6.1. Experimental Setup
6.1.1. AWR1642BOOST Millimeter-Wave Radar
6.1.2. NVIDIA Jetson Orin Nano
6.2. Dataset
6.3. Experimental Result
6.4. Comparisons with Other Methods
7. Discussion
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- World Health Organization. 2023. Global Status Report on Road Safety 2023. Geneva: World Health Organization.
- Jeon, Y.; Kim, B.; Baek, Y. Ensemble CNN to Detect Drowsy Driving with In-Vehicle Sensor Data. Sensors, 21(8), 2372, 2021. [CrossRef]
- Ersal, T.; Fuller, H.J.A.; Tsimhoni, O.; Stein, J.L.; Fathy, H.K. Model-Based Analysis and Classification of Driver Distraction Under Secondary Tasks. IEEE Transactions on Intelligent Transportation Systems, 11(3), 692–701, 2010.
- Dong, Z.; Zhang, M.; Sun, J.; Cao, T.; Liu, R.; Wang, Q. A Fatigue Driving Detection Method Based on Frequency Modulated Continuous Wave Radar. IEEE International Conference on Consumer Electronics and Computer Engineering (ICCECE), 670–675, 2021.
- Freitas, A.; Almeida, R.; Gonçalves, H.; Conceição, G.; Freitas, A. Monitoring Fatigue and Drowsiness in Motor Vehicle Occupants Using Electrocardiogram and Heart Rate—A Systematic Review. Transportation Research Part F: Traffic Psychology and Behaviour, 103, 586–607, 2024.
- Luo, H.; Qiu, T.; Liu, C.; Huang, P. Research on Fatigue Driving Detection Using Forehead EEG Based on Adaptive Multi-Scale Entropy. Biomedical Signal Processing and Control, 51, 50–58, 2019.
- Ren, B.; Guan, W.; Zhou, Q.; Wang, Z. EEG-Based Driver Fatigue Monitoring Within a Human–Ship–Environment System: Implications for Ship Braking Safety. Sensors, 23(10), 4644, 2023.
- Jiao, Y.; Zhang, C.; Chen, X.; Fu, L.; Jiang, C.; Wen, C. Driver Fatigue Detection Using Measures of Heart Rate Variability and Electrodermal Activity. IEEE Transactions on Intelligent Transportation Systems, 25(6), 5510–5524, 2023.
- Lian, Z.; Xu, T.; Yuan, Z.; Li, J.; Thakor, N.; Wang, H. Driving Fatigue Detection Based on Hybrid Electroencephalography and Eye Tracking. IEEE Journal of Biomedical and Health Informatics, 2024.
- Zhu, T.; Zhang, C.; Wu, T.; Ouyang, Z.; Li, H.; Na, X.; Liang, J.; Li, W. Research on a Real-Time Driver Fatigue Detection Algorithm Based on Facial Video Sequences. Applied Sciences, 12(4), 2224, 2022.
- Kır, S.; Becerikli, Y. Behavior-based driver fatigue detection system with deep belief network. Neural Computing and Applications, 34(16), 14053–14065, 2022.
- Venon, A.; Dupuis, Y.; Vasseur, P.; Merriaux, P. Millimeter Wave FMCW RADARs for Perception, Recognition and Localization in Automotive Applications: A Survey. IEEE Transactions on Intelligent Vehicles, 7(3), 533–555, 2022. [CrossRef]
- Warwick, B.; Symons, N.; Chen, X.; Xiong, K. Detecting Driver Drowsiness Using Wireless Wearables. In Proceedings of the 2015 IEEE 12th International Conference on Mobile Ad Hoc and Sensor Systems (MASS), Dallas, TX, USA, 19–22 October 2015; pp. 585–588. [CrossRef]
- Kendall, A.; Gal, Y.; Cipolla, R. Multi-Task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7482–7491. [CrossRef]
- Jung, J.; Lim, S.; Kim, B.K.; Lee, S. CNN-based driver monitoring using millimeter-wave radar sensor. IEEE Sensors Letters 2021, 5, 1–4. [CrossRef]
- Sen, A.; Mandal, A.; Karmakar, P.; Das, A.; Chakraborty, S. mmDrive: mmWave sensing for live monitoring and on-device inference of dangerous driving. In Proceedings of the 2023 IEEE International Conference on Pervasive Computing and Communications (PerCom), Orlando, FL, USA, 13–17 March 2023; pp. 2–11. IEEE. [CrossRef]
- Sengar, N.; Kumari, I.; Lee, J.; Har, D. PoseViNet: Distracted driver action recognition framework using multi-view pose estimation and vision transformer. arXiv 2023, arXiv:2312.14577. Available online: https://arxiv.org/abs/2312.14577.
- Shariff, W.; Kielty, P.; Lemley, J.; Corcoran, P. Spiking-DD: Neuromorphic event camera based driver distraction detection with spiking neural network. In Proceedings of the IET Conference Proceedings CP887, Stevenage, UK, August 2024; Volume 2024, No. 10, pp. 71–78. [CrossRef]
- Guo, Z.; Liu, Q.; Zhang, L.; Li, Z.; Li, G. L-TLA: A lightweight driver distraction detection method based on three-level attention mechanisms. IEEE Transactions on Reliability 2024, 73, 1731–1742. [CrossRef]














| Parameter | Value |
|---|---|
| Frequency | 77 GHz |
| Platform | xWR16xx |
| Scene Classifier | best_range_res |
| Azimuth Resolution | 15° |
| Range Resolution | 0.047 m |
| Maximum Unambiguous Range | 2.42 m |
| Maximum Radial Velocity | 1 m/s |
| Radial Velocity Resolution | 0.13 m/s |
| Frame Duration | 250 ms |
| RF Calibration Data | None |
| Range Detection Threshold | 15 dB |
| Doppler Detection Threshold | 15 dB |
| Range Peak Grouping | Enabled |
| Doppler Peak Grouping | Enabled |
| Static Clutter Removal | Disabled |
| Angle of Arrival FoV | Full FoV |
| Range FoV | Full FoV |
| Doppler FoV | Full FoV |
| Specification | AWR1642BOOST |
|---|---|
| Frequency Band | 77 GHz |
| Processing Unit | DSP + Hardware Accelerators |
| Interface | UART, SPI, I2C |
| Application | Radar Signal Processing |
| Specification | Jetson Orin Nano |
|---|---|
| Processing Unit | 6-core ARM Cortex-A78AE CPU |
| AI Acceleration | 1024-core NVIDIA Ampere GPU + 32 Tensor Cores |
| Memory | 8GB LPDDR5 |
| Power Consumption | 7W – 15W |
| Interface | USB, PCIe, GPIO, I2C, SPI |
| Application | Edge AI Computing |
| Method | Sen et al. [16] | Jung et al. [15] | Sengar et al. [17] | Shariff et al. [18] | Guo et al. [19] | Our Method |
|---|---|---|---|---|---|---|
| Sensor Type | FMCW Radar | FMCW Radar | Multi-view Camera | Event Camera | Depth Camera | FMCW Radar |
| Nodding | 93% | 80% | - | - | - | 96% |
| Yawning | 96% | - | 86% | - | - | 98% |
| Using Phone | 97% | - | 89% | - | 90% | 95% |
| Picking Object | 97% | - | 85% | - | - | 97% |
| Turning Back | 97% | - | 92% | - | 86.67% | 95% |
| Drinking | 99% | - | 92% | - | 93.33% | 99% |
| Fetching Forward | 87% | - | 89% | - | 84.44% | 96% |
| Fatigue Driving | - | - | - | 94.4% | - | -% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).