Submitted:
04 March 2025
Posted:
05 March 2025
You are already at the latest version
Abstract
Keywords:
I. Introduction
- A.
- Background and Importance of Object Detection in Autonomous Vehicles
- B. Objectives of the Analysis
- Explore and evaluate state-of-the-art object detection techniques used in autonomous vehicles, including both traditional methods and modern deep learning-based approaches.
- Identify the strengths and limitations of current object detection methods, particularly in the context of real-world driving conditions.
- Discuss future directions for improving object detection in AVs, such as the integration of multi-sensor data, advancements in model architectures, and the adoption of edge computing for real-time processing.
II. Overview of Object Detection in Autonomous Vehicles
- A.
- Key Requirements for Object Detection in Autonomous Driving
- Real-time processing: The system must process data and detect objects with minimal latency to enable timely decision-making and control, especially in dynamic driving environments.
- High accuracy and robustness: Detection systems must achieve high precision and recall rates while being resilient to challenges such as varying lighting conditions, occlusions, and adverse weather.
- Ability to detect multiple object classes: The system should be capable of identifying and classifying a wide range of objects, including pedestrians, vehicles, cyclists, traffic signs, and other road users, to ensure comprehensive situational awareness.
- B. Types of Objects Detected
- Static objects: These include traffic lights, road signs, lane markings, and other stationary elements that provide critical information for navigation and decision-making.
- Dynamic objects: These encompass vehicles, pedestrians, cyclists, and other movus vehicles rely on a combination of sensor modalities to achieve robust and accurate object detection:
- LiDAR: Light Detection and Ranging (LiDAR) sensors generate precise 3D point clouds, offering accurate distance measurements and object localization, even in low-light conditions.
- Radar: Radar systems are effective for detecting objects at long ranges and in adverse weather conditions, providing reliable speed and distance measurements.
- Sensor fusion techniques: Combining data from multiple sensors (e.g., cameras, LiDAR, and radar) through sensor fusion techniques enhances detection accuracy, robustness, and redundancy, addressing the limitations of individual sensors.
III. Traditional Object Detection Techniques
- A.
- Feature-Based Methods
- Histogram of Oriented Gradients (HOG): HOG extracts gradient orientation histograms from an image to capture edge and texture information, which is then used to detect objects. It is widely used for pedestrian detection due to its ability to capture shape information.
- Scale-Invariant Feature Transform (SIFT): SIFT identifies key points in an image and computes descriptors that are invariant to scale, rotation, and illumination changes. It is particularly useful for detecting objects in varying conditions but is computationally intensive.
- B. Machine Learning Approaches
- Support Vector Machines (SVM): SVMs are used to classify objects by finding the optimal hyperplane that separates different object classes in a high-dimensional feature space. They are effective for binary classification tasks but struggle with multi-class detection.
- AdaBoost: AdaBoost is an ensemble learning technique that combines multiple weak classifiers to create a strong classifier. It is commonly used in conjunction with Haar-like features for object detection, such as in face detection systems.
- C. Limitations of Traditional Methods
- Lack of robustness in complex environments: These methods often fail to generalize well to diverse and dynamic real-world scenarios, such as varying lighting conditions, occlusions, and cluttered backgrounds.
- Limited scalability for real-time applications: Traditional techniques are computationally expensive and struggle to meet the real-time processing requirements of autonomous driving systems, especially as the complexity of the environment increases.
IV. Deep Learning-Based Object Detection Techniques
- A.
- Convolutional Neural Networks (CNNs)
- ●
- Overview of CNN architecture: CNNs are composed of multiple layers, including convolutional layers, pooling layers, and fully connected layers. These layers work together to automatically extract hierarchical features from input images, enabling effective object detection.
- ●
- Role of CNNs in feature extraction and classification: CNNs excel at learning spatial hierarchies of features, making them highly effective for tasks like object localization and classification. They eliminate the need for handcrafted features, improving robustness and accuracy.
- ●
- B. Two-Stage Detectors
- ●
- Region-based CNN (R-CNN): R-CNN generates region proposals using selective search and then classifies and refines these regions using CNNs. While accurate, it is computationally expensive.
- ●
- Fast R-CNN and Faster R-CNN: Fast R-CNN improves efficiency by sharing convolutional features across region proposals. Faster R-CNN introduces a Region Proposal Network (RPN) to further speed up the process.
- ●
- Mask R-CNN: An extension of Faster R-CNN, Mask R-CNN adds a branch for pixel-level object segmentation, making it suitable for tasks requiring precise object boundaries.
- ●
- C. Single-Stage Detectors
- ●
- You Only Look Once (YOLO) series (YOLOv1 to YOLOv8): YOLO frameworks perform object detection in a single forward pass, achieving high speed and real-time performance. Each iteration (YOLOv1 to YOLOv8) introduces improvements in accuracy and efficiency.
- ●
- Single Shot MultiBox Detector (SSD): SSD predicts object categories and bounding boxes at multiple scales directly from feature maps, balancing speed and accuracy.
- ●
- RetinaNet: RetinaNet addresses the class imbalance problem in single-stage detectors using a focal loss function, achieving accuracy comparable to two-stage detectors while maintaining high speed.
- ●
- D. Transformer-Based Detectors
- ●
- Vision Transformers (ViTs): ViTs apply transformer architectures to image data, leveraging self-attention mechanisms to capture global context and improve object detection performance.
- ●
- DETR (DEtection TRansformer): DETR uses transformers to directly predict object bounding boxes and classes in an end-to-end manner, eliminating the need for handcrafted components like anchor boxes.
- ●
- E. 3D Object Detection Techniques
- ●
- PointNet and PointNet++ for LiDAR data: These methods process raw point cloud data from LiDAR sensors, enabling accurate 3D object detection by capturing spatial relationships between points.
- ●
- Voxel-based methods: These approaches convert point clouds into voxel grids, allowing the use of 3D CNNs for feature extraction and object detection.
- ●
- Frustum-based methods: These techniques combine 2D object detections from cameras with 3D point clouds from LiDAR, focusing on regions of interest to improve efficiency and accuracy.
V. Evaluation Metrics for Object Detection
- A.
- Common Metrics
- A.
- Precision, Recall, and F1-Score:
- ●
- Precision: Measures the proportion of correctly detected objects out of all detected objects (true positives / (true positives + false positives)).
- ●
- Recall: Measures the proportion of correctly detected objects out of all ground truth objects (true positives / (true positives + false negatives)).
- ●
- F1-Score: The harmonic mean of precision and recall, providing a balanced measure of a model’s accuracy.
- ●
- Intersection over Union (IoU):
- ●
- IoU quantifies the overlap between a predicted bounding box and the ground truth bounding box. It is calculated as the area of intersection divided by the area of union. A higher IoU indicates better localization accuracy.
- ●
- Mean Average Precision (mAP):
- B. Challenges in Evaluation
- Handling class imbalance:
- Evaluating performance in diverse environmental conditions:
VI. Comparative Analysis of Object Detection Techniques
- A.
- Performance Comparison
- A.
- Traditional Methods (e.g., HOG, SIFT, SVM):
- ●
- Strengths: Simple, interpretable, and effective in controlled environments with limited object classes.
- ●
- Weaknesses: Poor performance in complex, dynamic environments; limited scalability for real-time applications.
- ●
- Performance: Low to moderate accuracy, especially in scenarios with occlusions, varying lighting, or cluttered backgrounds.
- ●
- Deep Learning-Based Methods:
- ●
- Two-Stage Detectors (e.g., R-CNN, Faster R-CNN, Mask R-CNN):
- ●
- Strengths: High accuracy, especially for small and occluded objects; excellent for procalization and segmentation.
- ●
- Weaknesses: Computationally intensive, slower inference speeds compared to single-stage detectors.
- ●
- Performance: State-of-the-art accuracy on benchmark datasets but less suitable for real-time applications.
- ●
- Single-Stage Detectors (e.g., YOLO, SSD, RetinaNet):
- ●
- Strengths: High speed and efficiency, suitable for real-time applications; good balance between accuracy and speed.
- ●
- Weaknesses: Slightly lower accuracy compared to two-stage detectors, especially for small objects.
- ●
- Performance: Excellent for real-time object detection with competitive accuracy on benchmark datasets.
- ●
- Transformer-Based Detectors (e.g., DETR, ViTs):
- ●
- Strengths: Strong global context modeling, end-to-end training, and high accuracy.
- ●
- Weaknesses: High computational requirements and slower training times compared to CNNs.
- ●
- Performance: Competitive accuracy, especially in complex scenes, but still evolving for real-time applications.
- 3D
- Object Detection Techniques (e.g., PointNet, Voxel-based methods):
- ●
- Strengths: Accurate 3D localization and detection, essential for autonomous driving.
- ●
- Weaknesses: Computationally expensive and dependent on high-quality LiDAR or depth data.
- ●
- Performance: High accuracy in 3D object detection tasks but slower compared to 2D methods.
- ●
- B. Computational Efficiency
- ●
- Traditional Methods: Low computational efficiency due to reliance on handcrafted features and limited scalability.
- ●
- Two-Stage Detectors: Moderate to high computational cost due to region proposal generation and refinement.
- ●
- Single-Stage Detectors: High computational efficiency, optimized for real-time performance.
- ●
- Transformer-Based Detectors: High computational cost during training and inference, but advancements are improving efficiency.
- ●
- 3D Object Detection Techniques: High computational cost due to the complexity of processing 3D point cloud data.
- ●
- C. Strengths and Weaknesses of Each Technique
- ●
- Traditional Methods:
- ●
- Strengths: Simplicity and interpretability.
- ●
- Weaknesses: Limited accuracy and robustness in complex environments.
- ●
- Two-Stage Detectors:
- ●
- Strengths: High accuracy and precise localization
- ●
- Weaknesses: Slower inference speeds and higher computational cost.
- ●
- Single-Stage Detectors:
- ●
- Strengths: High speed and efficiency, suitable for real-time applications.
- ●
- Weaknesses: Slightly lower accuracy for small or occluded objects.
- ●
- Transformer-Based Detectors:
- ●
- Strengths: Strong global context modeling and high accuracy.
- ●
- Weaknesses: High computational requirements and slower training times.
- 3D
- Object Detection Techniques:
- ●
- Strengths: Accurate 3D localization, essential for autonomous driving.
- ●
- Weaknesses: Computationally expensive and dependent on high-quality sensor data.
VII.Challengesand Open Problems
- A.
- Environmental and Operational Challenges
- ●
- Varying Lighting Conditions: Object detection systems must perform reliably in diverse lighting scenarios, such as bright sunlight, nighttime, and shadows, which can significantly impact visibility and accuracy.
- ●
- Adverse Weather Conditions: Rain, snow, fog, and dust can obscure sensors and reduce detection performance, posing a challenge for robust operation.
- ●
- Dynamic and Complex Environments: Urban environments with high traffic density, occlusions, and unpredictable pedestrian behavior require highly adaptable detection systems.
- ●
- Sensor Limitations: Cameras, LiDAR, and radar each have limitations (e.g., cameras struggle in low light, LiDAR is affected by weather, radar has low resolution), necessitating effective sensor fusion.
- ●
- B. Technical Challenges
- ●
- Real-Time Processing: Achieving low-latency object detection while maintaining high accuracy is critical for autonomous driving but remains computationally demanding.
- ●
- Scalability: Developing detection systems that can scale to handle large datasets and diverse driving scenarios without compromising performance.
- ●
- Generalization: Ensuring models trained on specific datasets generalize well to unseen environments and conditions.
- ●
- 3D Object Detection: Accurately detecting and localizing objects in 3D space using LiDAR or stereo vision data is computationally intensive and requires advanced algorithms.
- ●
- Edge Cases: Handling rare or unexpected scenarios, such as unusual vehicle shapes, partially occluded objects, or novel road conditions, remains a significant challenge.
- ●
- C. Ethical and Safety Concerns
- ●
- Safety and Reliability: Ensuring object detection systems are fail-safe and can operate reliably in all conditions is critical to prevent accidents and ensure public trust.
- ●
- Bias and Fairness: Addressing potential biases in training datasets that could lead to unequal performance across different demographics or object types.
- ●
- Privacy Concerns: Cameras and sensors used in autonomous vehicles raise privacy issues, as they may capture and process sensitive information about pedestrians and other road users.
- ●
- Regulatory Compliance: Meeting evolving regulatory standards for autonomous vehicles, including certification of object detection systems, is a complex and ongoing challenge.
- ●
- Ethical Decision-Making: Developing frameworks for ethical decision-making in scenarios where accidents are unavoidable (e.g., choosing between two harmful outcomes) remains an open problem.
VIII. Future Directions and Emerging Trends
- A.
- Advancements in Deep Learning Architectures
- ●
- Transformer-Based Models: Vision Transformers (ViTs) and DETR are gaining traction for their ability to model global context and improve object detection accuracy. Future advancements may focus on optimizing these models for real-time applications.
- ●
- Lightweight Architectures: Developing efficient neural networks (e.g., MobileNet, EfficientNet) that maintain high accuracy while reducing computational requirements for edge devices.
- ●
- Continual Learning: Enabling models to learn incrementally from new data without forgetting previously learned knowledge, improving adaptability to changing environments.
- ●
- Self-Supervised Learning: Reducing reliance on labeled data by leveraging unlabeled data for pre-training, making object detection systems more scalable and cost-effective.
- ●
- B. Integration of Multi-Sensor Data
- ●
- Sensor Fusion Techniques: Advanced fusion methods (e.g., deep learning-based fusion) to combine data from cameras, LiDAR, radar, and other sensors for more robust and accurate object detection.
- ●
- Cross-Modal Learning: Developing models that can effectively learn and transfer knowledge across different sensor modalities, enhancing performance in diverse conditions.
- ●
- Unified Architectures: Designing end-to-end frameworks that process multi-sensor data seamlessly, improving efficiency and reducing latency.
- ●
- C. Role of Simulation and Synthetic Data
- ●
- High-Fidelity Simulators: Using advanced simulation platforms (e.g., CARLA, NVIDIA DRIVE Sim) to generate realistic training data and test object detection systems in diverse scenarios.
- ●
- Synthetic Data Generation: Leveraging generative models (e.g., GANs) to create synthetic datasets that complement real-world data, addressing data scarcity and diversity issues.
- ●
- Domain Adaptation: Developing techniques to bridge the gap between synthetic and real-world data, ensuring models trained in simulation perform well in real-world environments.
- ●
- D. Explainability and Transparency
- ●
- Interpretable Models: Designing object detection systems that provide clear explanations for their predictions, enhancing trust and accountability.
- ●
- Visualization Tools: Developing tools to visualize model decision-making processes, such as attention maps and feature visualizations, to better understand model behavior.
- ●
- Ethical AI Frameworks: Establishing guidelines and frameworks to ensure object detection systems are transparent, fair, and free from biases.
- ●
- Human-in-the-Loop Systems: Integrating human oversight to validate and interpret model predictions, particularly in critical or ambiguous scenarios.
IX. Case Studies and Real-World Applications
- A.
- Industry Leaders in Autonomous Vehicle Object Detection
- ●
- Tesla:
- ●
- Tesla’s Autopilot and Full Self-Driving (FSD) systems rely heavily on object detection using a camera-centric approach. Their neural networks process data from multiple cameras to detect and classify objects like vehicles, pedestrians, and traffic signs in real time.
- ●
- Key Innovation: Tesla’s use of deep learning and over-the-air updates allows continuous improvement of their object detection models based on real-world driving data.
- ●
- Waymo:
- ●
- Waymo, a subsidiary of Alphabet, uses a combination of LiDAR, radar, and cameras for object detection. Their systems are designed to operate in complex urban environments, leveraging high-resolution 3D maps and advanced sensor fusion techniques.
- ●
- Key Innovation: Waymo’s focus on safety and redundancy ensures robust object detection even in challenging conditions like heavy rain or fog.
- ●
- Mobileye (Intel):
- ●
- Mobileye specializes in vision-based object detection systems for advanced driver-assistance systems (ADAS) and autonomous vehicles. Their EyeQ chips process camera data to detect and track objects with high accuracy.
- ●
- Key Innovation: Mobileye’s proprietary algorithms and hardware-software integration optimize performance for real-time applications.
- ●
- NVIDIA:
- ●
- NVIDIA’s DRIVE platform provides end-to-end solutions for autonomous driving, including state-of-the-art object detection using deep learning. Their platforms support multi-sensor fusion and are widely used by automotive manufacturers and researchers.
- ●
- Key Innovation: NVIDIA’s focus on scalable and efficient AI platforms enables rapid development and deployment of object detection systems.
- ●
- B. Academic and Research Contributions
- ●
- KITTI Dataset and Benchmark:
- ●
- The KITTI dataset, developed by the Karlsruhe Institute of Technology and Toyota Technological Institute at Chicago, is a widely used benchmark for object detection in autonomous driving. It includes labeled data from cameras, LiDAR, and GPS, enabling researchers to evaluate and compare object detection algorithms.
- ●
- Impact: KITTI has driven significant advancements in object detection techniques, particularly in 3D object detection and sensor fusion.
- ●
- Berkeley DeepDrive (BDD) Dataset:
- ●
- The BDD dataset, created by UC Berkeley, contains diverse driving scenarios captured across different times of day, weather conditions, and locations. It has been instrumental in developing robust object detection models that generalize well to real-world conditions.
- ●
- Impact: BDD has facilitated research into domain adaptation, multi-task learning, and robustness in object detection systems.
- ●
- Stanford Autonomous Driving Research:
- ●
- Stanford University’s research focuses on developing explainable and ethical AI systems for autonomous driving. Their work includes advancements in interpretable object detection models and safety-critical decision-making frameworks.
- ●
- Impact: Stanford’s contributions have highlighted the importance of transparency and safety in autonomous vehicle technologies.
- ●
- CMU’s TartanDrive Dataset:
- ●
- Carnegie Mellon University’s TartanDrive dataset includes multi-modal sensor data (cameras, LiDAR, IMU) collected in off-road environments. It supports research into object detection in unstructured and challenging terrains.
- ●
- Impact: TartanDrive has expanded the scope of object detection research to include non-urban and off-road applications.
X. Conclusion
- A.
- Summary of Key Findings
- ●
- Importance of Object Detection: Object detection is a cornerstone technology for autonomous vehicles, enabling safe and efficient navigation by accurately identifying and localizing objects such as pedestrians, vehicles, and traffic signs.
- ●
- Evolution of Techniques: Traditional methods like HOG and SIFT have been largely replaced by deep learning-based approaches, including CNNs, YOLO, SSD, and transformer-based models, which offer superior accuracy and robustness.
- ●
- Challenges: Real-time processing, varying environmental conditions, sensor limitations, and ethical concerns remain significant challenges for object detection systems in autonomous vehicles.
- ●
- Advancements: Emerging trends such as sensor fusion, simulation, synthetic data, and explainable AI are driving innovation and addressing current limitations.
- ●
- Industry and Academic Contributions: Industry leaders like Tesla, Waymo, and NVIDIA, along with academic research initiatives, have significantly advanced the field, providing datasets, benchmarks, and cutting-edge technologies.
- ●
- B. Final Thoughts
References
- M. Islam et al., “A Comprehensive Review on Object Detection in the Context of Autonomous Driving,” 2024 4th International Conference on Ubiquitous Computing and Intelligent Information Systems (ICUIS), Gobichettipalayam, India, 2024, pp. 1860-1864. [CrossRef]
- Islam, M. M., Chowdhury, I. J., Mahboob, T. Z., Mazumder, M. S. J., Hossain, M. J., Biswas, M. S., & Rone, P. D. (2024, December). A Comprehensive Review on Object Detection in the Context of Autonomous Driving. In 2024 4th International Conference on Ubiquitous Computing and Intelligent Information Systems (ICUIS) (pp. 1860-1864). IEEE. [CrossRef]
- Suraj, P. (2024). SYNERGIZING ROBOTICS AND ARTIFICIAL INTELLIGENCE: TRANSFORMING MANUFACTURING AND AUTOMATION FOR INDUSTRY 5.0. Synergy: Cross-Disciplinary Journal of Digital Investigation, 2(11), 69-75.
- Raju, O. N., Rakesh, D., & SubbaReddy, K. (2012). SRGM with imperfect debugging using capability analysis of log-logistic model. Int J Comput Technol, 2, 30-33. [CrossRef]
- Dasari, R., Prasanth, Y., & NagaRaju, O. (2017). An analysis of most effective virtual machine image encryption technique for cloud security. International Journal of Applied Engineering Research, 12(24), 15501-15508.
- Islam, M. S., Rony, M. A. T., Saha, P., Ahammad, M., Alam, S. M. N., & Rahman, M. S. (2023, December). Beyond words: unraveling text complexity with novel dataset and a classifier application. In 2023 26th International Conference on Computer and Information Technology (ICCIT) (pp. 1-6). IEEE. [CrossRef]
- Islam, M. M., Chowdhury, I. J., Mahboob, T. Z., Mazumder, M. S. J., Hossain, M. J., Biswas, M. S., & Rone, P. D. (2024, December). A Comprehensive Review on Object Detection in the Context of Autonomous Driving. In 2024 4th International Conference on Ubiquitous Computing and Intelligent Information Systems (ICUIS) (pp. 1860-1864). IEEE. [CrossRef]
- Amarnath Immadisetty. (2024). MASTERING DATA PLATFORM DESIGN: INDUSTRY-AGNOSTIC PATTERNS FOR SCALE. INTERNATIONAL JOURNAL OF RESEARCH IN COMPUTER APPLICATIONS AND INFORMATION TECHNOLOGY (IJRCAIT), 7(2), 2259-2270. https://ijrcait.com/index.php/home/article/view/IJRCAIT_07_02_164.
- Immadisetty, A. (2024). SUSTAINABLE INNOVATION IN DIGITAL TECHNOLOGIES: A SYSTEMATIC REVIEW OF ENERGY-EFFICIENT COMPUTING AND CIRCULAR DESIGN PRACTICES. INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING AND TECHNOLOGY, 15(06), 1056-1066.
- Anjum, Kazi Nafisa, and Ayuns Luz. “Investigating the Role of Internet of Things (IoT) Sensors in Enhancing Construction Site Safety and Efficiency.”.
- Chinta, Purna Chandra Rao, Niharika Katnapally, Krishna Ja, Varun Bodepudi, Suneel Babu, and Manikanth Sakuru Boppana. “Exploring the role of neural networks in big data-driven ERP systems for proactive cybersecurity management.” Kurdish Studies (2022). [CrossRef]
- Singh, J. (2022). The Ethics of Data Ownership in Autonomous Driving: Navigating Legal, Privacy, and Decision-Making Challenges in a Fully Automated Transport System. Australian Journal of Machine Learning Research & Applications, 2(1), 324-366.
- Singh, J. (2024). Autonomous Vehicles and Smart Cities: Integrating AI to Improve Traffic Flow, Parking, and Environmental Impact. Journal of AI-Assisted Scientific Discovery, 4(2), 65-105.
- Krishna Madhav, J., Varun, B., Niharika, K., Srinivasa Rao, M., & Laxmana Murthy, K. (2023). Optimising Sales Forecasts in ERP Systems Using Machine Learning and Predictive Analytics. J Contemp Edu Theo Artific Intel: JCETAI-104. [CrossRef]
- Singh, J. (2024). AI-Driven Path Planning in Autonomous Vehicles: Algorithms for Safe and Efficient Navigation in Dynamic Environments. Journal of AI-Assisted Scientific Discovery, 4(1), 48-88.
- Mmaduekwe, U., and E. Mmaduekwe. “Cybersecurity and Cryptography: The New Era of Quantum Computing.” Current Journal of Applied Science and Technology 43, no. 5. [CrossRef]
- Singh, J. (2024). Robust AI Algorithms for Autonomous Vehicle Perception: Fusing Sensor Data from Vision, LiDAR, and Radar for Enhanced Safety. Journal of AI-Assisted Scientific Discovery, 4(1), 118-157.
- Singh, J. (2022). Deepfakes: The Threat to Data Authenticity and Public Trust in the Age of AI-Driven Manipulation of Visual and Audio Content. Journal of AI-Assisted Scientific Discovery, 2(1), 428-467.
- Routhu, Kishankumar, Varun Bodepudi, Krishna Madhav Jha, and Purna Chandra Rao Chinta. “A Deep Learning Architectures for Enhancing Cyber Security Protocols in Big Data Integrated ERP Systems.” Available at SSRN 5102662 (2020). [CrossRef]
- Bodepudi, V., & Chinta, P. C. R. (2024). Enhancing Financial Predictions Based on Bitcoin Prices using Big Data and Deep Learning Approach. Available at SSRN 5112132. [CrossRef]
- Chinta, P. C. R., Moore, C. S., Karaka, L. M., Sakuru, M., Bodepudi, V., & Maka, S. R. (2025). Building an Intelligent Phishing Email Detection System Using Machine Learning and Feature Engineering. European Journal of Applied Science, Engineering and Technology, 3(2), 41-54. [CrossRef]
- Moore, C. (2024). Enhancing Network Security With Artificial Intelligence Based Traffic Anomaly Detection In Big Data Systems. Available at SSRN 5103209. [CrossRef]
- Krishna Madhav, J., Varun, B., Niharika, K., Srinivasa Rao, M., & Laxmana Murthy, K. (2023). Optimising Sales Forecasts in ERP Systems Using Machine Learning and Predictive Analytics. J Contemp Edu Theo Artific Intel: JCETAI-104. [CrossRef]
- Singh, J. (2023). Advancements in AI-Driven Autonomous Robotics: Leveraging Deep Learning for Real-Time Decision Making and Object Recognition. Journal of Artificial Intelligence Research and Applications, 3(1), 657-697.
- Sadaram, G., Karaka, L. M., Maka, S. R., Sakuru, M., Boppana, S. B., & Katnapally, N. (2024). AI-Powered Cyber Threat Detection: Leveraging Machine Learning for Real-Time Anomaly Identification and Threat Mitigation. MSW Management Journal, 34(2), 788-803.
- Chinta, Purna Chandra Rao. “The Art of Business Analysis in Information Management Projects: Best Practices and Insights.” DOI 10 (2023). [CrossRef]
- Azuikpe, P. F., Fabuyi, J. A., Balogun, A. Y., Adetunji, P. A., Peprah, K. N., Mmaduekwe, E., & Ejidare, M. C. (2024). The necessity of artificial intelligence in fintech for SupTech and RegTech supervisory in banks and financial organizations. International Journal of Science and Research Archive, 12(2), 2853-2860. [CrossRef]
- Chinta, P. C. R., & Katnapally, N. (2021). Neural Network-Based Risk Assessment for Cybersecurity in Big Data-Oriented ERP Infrastructures. Neural Network-Based Risk Assessment for Cybersecurity in Big Data-Oriented ERP Infrastructures. [CrossRef]
- Singh, J. (2019). Sensor-Based Personal Data Collection in the Digital Age: Exploring Privacy Implications, AI-Driven Analytics, and Security Challenges in IoT and Wearable Devices. Distributed Learning and Broad Applications in Scientific Research, 5, 785-809.
- Singh, J. (2021). The Rise of Synthetic Data: Enhancing AI and Machine Learning Model Training to Address Data Scarcity and Mitigate Privacy Risks. Journal of Artificial Intelligence Research and Applications, 1(2), 292-332.
- Katnapally, N., Chinta, P. C. R., Routhu, K. K., Velaga, V., Bodepudi, V., & Karaka, L. M. (2021). Leveraging Big Data Analytics and Machine Learning Techniques for Sentiment Analysis of Amazon Product Reviews in Business Insights. American Journal of Computing and Engineering, 4(2), 35-51. [CrossRef]
- Sadaram, Gangadhar, Manikanth Sakuru, Laxmana Murthy Karaka, Mohit Surender Reddy, Varun Bodepudi, Suneel Babu Boppana, and Srinivasa Rao Maka. “Internet of Things (IoT) Cybersecurity Enhancement through Artificial Intelligence: A Study on Intrusion Detection Systems.” Universal Library of Engineering Technology Issue (2022). [CrossRef]
- Katnapally, N., Chinta, P. C. R., Routhu, K. K., Velaga, V., Bodepudi, V., & Karaka, L. M. (2021). Leveraging Big Data Analytics and Machine Learning Techniques for Sentiment Analysis of Amazon Product Reviews in Business Insights. American Journal of Computing and Engineering, 4(2), 35-51. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).