Preprint
Article

This version is not peer-reviewed.

SIFT-SNN for Traffic-Flow Infrastructure Safety: A Real-Time Context-Aware Anomaly Detection Framework

A peer-reviewed article of this preprint also exists.

Submitted:

02 December 2025

Posted:

03 December 2025

You are already at the latest version

Abstract
Automated anomaly detection in transportation infrastructures is essential for enhancing safety and reducing the operational costs associated with manual inspection protocols. This study presents an improved neuromorphic vision system, which extends the prior SIFT-SNN (Scale-Invariant Feature Transform–Spiking Neural Network) proof-of-concept by incorporating temporal feature aggregation for context-aware and sequence-stable detection. Analysis of classical stitching-based pipelines exposed sensitivity to motion and lighting variations, motivating the proposed temporally smoothed neuromorphic design. SIFT keypoints are encoded into latency-based spike trains and classified using a leaky integrate-and-fire (LIF) spiking neural network implemented in PyTorch. Evaluated across three hardware configurations: an NVIDIA RTX 4060 GPU, an Intel i7 CPU, and a simulated Jetson Nano; the system achieves 92.3% accuracy and a macro F1-score of 91.0% under five-fold cross-validation. Inference latencies were measured at 9.5 ms, 26.1 ms, and ~48.3 ms per frame, respectively. Memory footprints were under 290 MB and power consumption estimated to be between 5 and 65 W. The classifier distinguishes between safe, partially dislodged, and fully dislodged barrier pins, which are critical failure modes for the Auckland Harbour Bridge’s Movable Concrete Barrier (MCB) system. Temporal smoothing further improves recall for ambiguous cases. Achieving a compact model size (2.9 MB), low-latency inference, and minimal power demands, the proposed framework offers a deployable, interpretable, and energy-efficient alternative to conventional CNN-based inspection tools. Future work will be focused on exploring generalisability and transferability of the work presented, additional input sources, as well as human computer interaction paradigms for various deployment infrastructures and advancement.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

1.1. Background

Traffic-related fatalities remain one of the most urgent public health concerns worldwide, with over 1.25 million deaths and between 20 and 50 million injuries reported annually [1,2,3]. The World Health Organisation forecasts that by 2030, road accidents will become the fifth leading cause of death globally [4]. Beyond the human cost, the global economic burden of road traffic injuries is projected to exceed USD 1.8 trillion from 2015 to 2030, equivalent to an annual tax of 0.12% on world GDP [5]. Poor road infrastructure, including degraded surfaces, missing safety barriers, and inadequate design, plays a significant role in global road safety failures. According to the UNECE (2023), over 1.19 million people die annually in traffic accidents, 90% of whom are from low- and middle-income countries, with 54% of fatalities involving vulnerable road users such as pedestrians, cyclists, and motorcyclists, groups especially exposed due to unsafe or poorly maintained infrastructure [6].
The degrading conditions of various traffic infrastructures are amplified by the growing complexity of modern infrastructure, which includes over 64 million kilometres of roads and more than one million bridges worldwide [3]. Recent high-profile collapses, such as the Morandi Bridge in Genoa (2018) and the Gujarat pedestrian bridge in India (2022), have highlighted the global vulnerability of aging or mismanaged infrastructure [7,8].
Currently, the predominant model for structural health monitoring is periodic manual inspection, with cycles ranging from six months to five years, depending on regional regulations [9]. While historically effective, this approach is increasingly unfit for modern demands due to high labour and safety-associated costs, subjective variability, exposure risks, and its inability to detect sudden or progressive failures between inspection windows.
In response to the limitations of manual inspections, there has been growing interest in Automated Road Defect and Anomaly Detection (ARDAD) systems that integrate computer vision (CV), machine learning (ML), and deep learning (DL) to automate large-scale inspection workflows [10,11]. A common objective to guide automation in ARDAD is the intention to reduce human dependency, enable real-time alerting, and enhance coverage. However, the current ARDAD landscape remains disproportionately focused on surface-level features, such as potholes, cracks, and asphalt delamination, while neglecting embedded or mechanical anomalies that may not manifest through traditional image cues, yet often precede catastrophic structural failures [11].

1.2. Structural Anomalies in Context

Contextual or structural anomalies such as displaced manhole covers, missing safety bolts, or modular component failure often lack prominent surface indicators yet pose a disproportionate threat to structural integrity and user safety. For example, the common issues of investigation usually involve out-of-place components within highly regular, load-bearing structures.
One critical example is the Movable Concrete Barrier (MCB) system installed on the Auckland Harbour Bridge (AHB), a vital multi-span transport corridor that carries more than 170,000 vehicles per day [12]. The bridge employs a Barrier Transfer Machine (BTM) to reposition concrete segments across lanes based on traffic flow demand [13]. Mechanical metal pins secure each segment, and failure or partial dislodgement of any of barrier components may result in severe misalignment or progressive collapse of the barrier sections.
Currently, inspections of barrier pins are conducted manually, and they are only performed during off-peak hours. No embedded or automated system exists for continuous monitoring, creating potential blind spots in infrastructure safety. Figure 1 illustrates the operational layout of the BTM system and its dependence on the structural integrity of the MCB units.
Given the mechanical and spatial regularity of such systems, context-aware detection approaches, those capable of understanding expected structural configurations across time, are essential for identifying subtle, high-risk deviations.

1.3. Motivational Gaps in Existing Systems

Although recent advances in image processing, computer vision (CV) and general AI-enabled inspections have yielded promising results in surface-level defect detection, significant limitations persist when applied to complex transport infrastructure such as the Auckland Harbour Bridge. The following shortcomings remain particularly unresolved:
  • Lack of subsurface and structural sensitivity: The majority of Automated Road Defect and Anomaly Detection (ARDAD) systems are designed to detect visible surface irregularities such as cracks and potholes. ARDAD models are not equipped to identify embedded or mechanical anomalies, such as displaced pins or loose joints, that may not be visually salient but pose critical safety risks.
  • Dependence on extensive training data: Deep learning-based solutions typically require large, annotated datasets that capture diverse operational scenarios. Domain shift and environmental variability (e.g., lighting, motion blur, weather) often reduce model generalisability in real-world deployment.
  • Limited temporal reasoning: Most systems operate on a frame-by-frame basis, ignoring temporal continuity. As a result, they are incapable of detecting progressive or transient anomalies that unfold over time, a key requirement for dynamic infrastructure such as movable barriers.
  • High computational complexity: State-of-the-art CNN architectures, while accurate, are computationally intensive and often require GPU resources. This makes them unsuitable for embedded or roadside units that must operate in real time under constrained power and latency budgets. Spatiotemporal models developed for video anomaly detection further highlight this limitation, as they typically operate on dense frame sequences and rely on heavy CNN or transformer backbones, which hinders real-time deployment on resource-constrained embedded hardware [15,16].
  • Low granularity in classification: Many ARDAD frameworks offer binary classification outputs (defect/no defect), which are insufficient for infrastructure systems that require graduated assessments of failure states (e.g., safe, partial displacement, total failure).
While some attempts have been made to preserve spatial context, such as frame aggregation under motion to merge adjacent frames, commonly used methods are unreliable under inspection scenarios involving high-speed motion, parallax distortions, or structural occlusions (Table 1). Recent infrastructure-inspection efforts have attempted dual-sensor image stitching and fusion to cover large spans of structural assets, combining visible and infrared modalities to improve defect visibility [17,18]. However, these approaches rely on consistent overlap and stable imaging conditions, making them vulnerable to motion blur, parallax, and lighting variability, conditions common in movable barrier inspections. As such, they fail to provide stable or coherent representations in contexts like the Barrier Transfer Machine (BTM) traversal along the Auckland Harbour Bridge.
To summarise, existing systems lack the spatiotemporal awareness, structural interpretability, and computational efficiency required for real-time, context-aware safety monitoring of mechanical infrastructure.

1.4. Preliminary Work

This study extends our earlier work presented in [19], where combining the Scale-Invariant Feature Transform (SIFT) with Spiking Neural Networks (SNN) approaches into a hybrid SIFT–SNN model was introduced as proof-of-concept for binary anomaly detection of mechanical pin dislodgement on the Auckland Harbour Bridge (AHB). While the initial proof-of-concept demonstrated the feasibility of combining classical descriptors with neuromorphic classification, the proof-of-concept was limited in three key aspects:
(i)
The absence of temporal modelling,
(ii)
a binary decision space that could not capture intermediate or ambiguous states, and
(iii)
lack of evaluation under hardware-constrained or real-time operating conditions.
To address the initial scope limitations and progress towards a deployable prototype, we now present a substantially extended system with the following contributions:
  • A temporal enhancement layer for modelling inter-frame dependencies.
  • A multi-class classification scheme capable of distinguishing between fully safe, partially dislodged, fully dislodged, and ambiguous mechanical pin states.
  • A hardware-aware performance evaluation, demonstrating low-latency classification using CPU-only and edge-constrained configurations.
Experience with panorama stitching for spatial aggregation revealed that conventional image-compositing approaches are unstable under motion, parallax, and illumination variation. This limitation motivated the transition from purely spatial aggregation to spatiotemporal feature modelling. The extended system presented here incorporates temporal reasoning, hardware-aware design, and multi-level diagnostic outputs into a unified framework for neuromorphic anomaly detection in smart infrastructure monitoring.

1.5. Objectives and Research Questions

This study addresses the limitations of the-state-of-the-art (Section 1.3) by advancing a previously validated SIFT-SNN anomaly detection pipeline into a context-aware, deployable system suitable for real-time infrastructure monitoring. The key objectives of the study are:
  • To incorporate temporal feature aggregation into the SIFT-SNN framework, enabling spatiotemporal context awareness across sequential inspection frames.
  • To extend the anomaly detection task from binary classification to a multi-class recognition scheme, enabling nuanced differentiation between partial, full, and ambiguous defect states.
  • To empirically benchmark the system’s latency, diagnostic accuracy, and suitability for edge deployment under constrained hardware deployment conditions.
  • To reflect on unsuccessful spatial context integration approaches, including panorama stitching, and clarify their limitations in dynamic transport inspection environments.
The study objectives reflect a multidimensional expansion of the original framework, encompassing the spatial, temporal, and operational dimensions of anomaly detection. By addressing both the methodological enhancements and their real-world constraints, the study aligns technical innovation with practical applicability in safety-critical infrastructure scenarios.
As a contribution to general computer vision and AI, the primary research questions (RQ) to guide the experimental work supporting the study are:
  • RQ1: Does the integration of temporal features across frame sequences improve anomaly detection performance over static SIFT-SNN pipelines?
  • RQ2: Can the enhanced SIFT-SNN framework operate with real-time latency and maintain diagnostic precision under edge deployment constraints?
  • RQ3: To what extent does multi-class classification improve safety diagnostics in infrastructure systems with graded failure modes?
For completeness, we note that early exploration of panorama stitching highlighted challenges in motion-heavy inspection scenarios. These observations motivated the emphasis on temporal modelling within the main research questions, rather than forming a standalone objective.

2. Related Work

Advancements in automated infrastructure monitoring have gained significant traction over the past decade and a half (2010-2025), particularly in the domains of road defect detection and structural health diagnostics. A substantial body of research has explored computer vision and deep learning approaches for identifying surface-level anomalies such as cracks, potholes, and material degradation. However, limited attention has been paid to the detection of structural or modular anomalies, particularly those associated with movable or load-bearing components such as mechanical joints, articulating hinges, or safety pins. To identify critical limitations that motivate the present study, this section reviews existing literature across four key themes: (1) automated road and bridge inspections, (2) structural health monitoring of modular systems, (3) spiking neural networks (SNNs), and (4) temporal modelling and contextual awareness.

2.1. Automated Road and Bridge Inspections

Automated Road Defect and Anomaly Detection (ARDAD) systems have evolved significantly over the past two decades, progressing from rule-based image processing and classical computer vision (CV) algorithms to machine learning (ML) and deep learning (DL) frameworks [11]. Traditional approaches such as edge detection, Hough transforms, and texture analysis were once dominant in detecting linear cracks and surface disruptions on pavements and bridges. However, traditional approaches for infrastructure anomaly detection (for safety monitoring) often lacked robustness under varying lighting, occlusion, and weather conditions [20].
The rise of convolutional neural networks (CNNs) ushered in a new generation of ARDAD systems capable of learning high-level visual features directly from annotated data. Studies such as Zhang, et al. [21] demonstrated the superior performance of deep CNNs in identifying fine-grained crack patterns with high accuracy across diverse road textures. Similarly, Maeda, et al. [22] introduced one of the earliest large-scale annotated datasets for road damage detection in Japan, enabling the benchmarking of supervised DL models for crack, pothole, and road mark degradation classification.
More recent research has incorporated drone-based and LiDAR-equipped inspection platforms to improve scene coverage and 3D context capture, particularly for bridges and elevated structures [23,24]. While the most recent advances have significantly improved detection rates and operational scalability, the dominant focus remains on static, surface-visible defects, such as longitudinal cracking, surface rutting, potholes, and material delamination, while deeper structural components are often overlooked [25].
Despite the growing volume of research and the availability of publicly accessible datasets (e.g. Crack500 [21], RDD2020/2022 [22], BDD100K [26], and AigleRN [27]), there remains a notable blind spot in the detection of embedded or structural anomalies. Examples include mechanical joints, fasteners, and movable modules, which often lack visual distinctiveness and require contextual or temporal analysis for reliable detection. Furthermore, the diagnostic emphasis in most existing literature has been on classification of known surface defects, with limited capability to detect out-of-distribution anomalies or to interpret evolving structural conditions that may not exhibit immediate surface cues.
The identified need for context-aware anomaly detection systems that integrate spatial geometry, temporal dynamics, and neuromorphic efficiency capabilities is the key element to justify the experimental work that led to the creation of the SIFT-SNN framework.

2.2. Structural Health Monitoring of Modular Systems

While traditional structural health monitoring (SHM) has primarily targeted large-scale anomalies, such as bridge deck corrosion [28], cable tension loss [29], or girder deformation [30], recent research has begun to address the nuanced challenges posed by modular and movable systems in transport infrastructure [31]. Examples include mechanically interlocked or articulated components, such as expansion joints [32], movable barrier systems [33], hinges, pin assemblies, and rail switches [34], whose failure modes are often subtle and lack distinct visual signatures when observed in isolated frames.
Many conventional SHM methods rely on vibration analysis, strain gauge readings, or acoustic emissions, which, although precise, require embedded sensors and are not always feasible for high-frequency, wide-area inspection [35]. As an alternative, vision-based SHM for modular systems has emerged, leveraging image and video feeds to detect alignment shifts, displacement patterns, or anomalous interactions between moving parts [36]. However, modular systems introduce two persistent challenges that complicate visual diagnostics include:
  • Lack of large, annotated datasets for such component-level defects.
  • Need for context-aware modelling due to geometric regularity and positional dependency.
In the context of movable concrete barrier (MCB) systems such as those on the Auckland Harbour Bridge (AHB), the failure of steel connecting pins may not result in overt visual damage but relatively slight misalignments or motion irregularities, which are difficult to detect with static image classifiers [37]. As such, temporal inspection methods, combining structural priors with dynamic context, are increasingly necessary.
Emerging works have proposed using techniques such as:
  • 3D reconstruction or photogrammetry to assess displacement across time [38]
  • Graph-based modelling of modular units and their interconnections [39]
  • Multi-view visual inspection to resolve ambiguous or occluded parts [40].
To date, however, few of the structural health monitoring approaches have been put into operation for edge-deployable or real-time systems suitable for live traffic scenarios. While mapping data and output classes directly to the traffic scenario seems an intuitive approach, most of the existing systems are not designed to handle multi-class or additional states of degradation (e.g., safe, partially failed, fully failed), which is essential for risk-sensitive components. One of the insights guiding our experimental research is based on a rationale for automated barrier pin position monitoring, where false positive vs false negative errors are not to be treated equally.
To address the gaps identified, this study seeks to build on the preliminary work and insights from anomaly detection in ARDAD systems and extend them with a hybrid neuromorphic vision architecture designed to deliver interpretable, multi-class classification and spatiotemporal reasoning for diagnosing failures in safety-critical structural components.

2.3. Spiking Neural Networks (SNNs) in Vision-Based Anomaly Detection

Spiking Neural Networks (SNNs) constitute the third generation of neural models, incorporating biologically inspired, asynchronous, and event-driven mechanisms for processing neural information. In contrast to conventional deep learning models, such as CNNs or Transformers, which rely on dense, synchronous matrix operations, SNNs transmit information via discrete spikes, emulating the neural dynamics observed in biological systems [41].
Considering increasing expectations for low energy consumption and low computational demands of SNN, this represents particularly attractive research avenue in CV for real-time, low-power applications in domains such as robotics, neuromorphic vision, and edge-based anomaly detection [42]. The Leaky Integrate-and-Fire (LIF) neuron model remains one of the most widely adopted in SNN implementations due to its simplicity and computational efficiency [43].
Key Advantages in Infrastructure Monitoring Contexts
SNNs exhibit intrinsic properties that make them particularly well-suited to safety-critical infrastructure diagnostics:
  • Event-driven efficiency: Only responds to salient input events, drastically reducing computation and energy demands [42].
  • Temporal sensitivity: Intrinsically models time-series or spatiotemporal data, aligning with the needs of progressive or transient anomaly detection [44].
  • Hardware deployability: Compatible with neuromorphic chips (e.g., Loihi, SpiNNaker), making real-time deployment feasible at the edge [45].
The benefits of SNNs have therefore motivated researchers to explore their use in a range of application domains beyond neuroscience and robotics.
Recent Applications in Safety Inspection
Although still in its early stages of development in civil infrastructure monitoring, SNN-based models are being gradually introduced into safety-critical tasks. Several recent studies illustrate their promise:
  • Zhang et al. used SNNs for surface-level defect classification in noisy visual environments [46].
  • Vemuru proposed a biologically inspired spiking neural network (SNN) edge detector employing Gabor-type filters for feature extraction, demonstrating that such hybrid architectures enhance edge saliency in low-exposure conditions and support improved classification in conjunction with CNNs [47].
  • In autonomous driving and dynamic scene analysis, SNNs have been shown to outperform CNNs in terms of latency and energy efficiency under constrained compute budgets [48].
However, most existing studies limit SNNs to static input modalities or binary classification settings. Few, if any, have applied SNNs to:
  • Context-aware multi-class diagnostics in modular infrastructure
  • Progressive fault detection using frame sequences
  • Integration with traditional descriptors (e.g., SIFT) for hybrid pipelines
This paper builds upon our early proof-of-concept that introduced a hybrid SIFT-SNN system for binary anomaly detection on movable barrier pins [19]. In this paper, the initial proof-of-concept is extended through the integration of temporal modelling, spatiotemporal context fusion, and multi-class diagnostic interpretation, hence marking one of the first attempts to deploy such a hybrid SIFT-SNN architecture in structural safety diagnostics under real-world dynamic conditions.

2.4. Temporal Modelling and Contextual Awareness

Contemporary anomaly detection systems, particularly in domains such as surveillance, human activity recognition, and medical diagnostics, have increasingly leveraged temporal cues to improve robustness, context awareness, and failure anticipation [49]. The anomaly detection systems utilise time-series modelling to detect not only the presence of an anomaly, but also its temporal evolution, offering significant benefits in applications where defects emerge progressively or transiently [50].
In computer vision, temporal modelling is typically achieved through 3D convolutional networks (C3D), recurrent neural networks (RNNs), temporal attention modules, or optical flow-based methods. For instance, Tran et al. demonstrated that 3D CNNs significantly outperformed 2D models for human action recognition by capturing spatiotemporal features across frames [51]. Likewise, Feichtenhofer et al. introduced the SlowFast network, which processes both slow and fast temporal dynamics, enabling fine-grained motion interpretation [52].
However, such approaches are rarely translated into infrastructure inspection tasks. Most ARDAD or SHM models continue to operate at a per-frame level, overlooking valuable inter-frame dependencies [31]. This omission is particularly problematic in modular systems like movable concrete barriers (MCBs), where subtle misalignments or component displacement only become evident over time [33].
Efforts to bridge spatial discontinuities across video sequences have explored panorama stitching using keypoint-based methods such as SIFT, SURF, or ORB [53,54]. While effective under controlled conditions, keypoint-based methods often fail in real-world inspection scenarios due to:
  • Rapid egomotion of the camera system
  • Structural parallax between bridge segments
  • Occlusion and lighting variations
  • Perspective distortions and lens warp
The limitations of keypoint-based methods significantly undermine the reliability of stitched spatial representations, especially in edge-deployable systems where compute and latency budgets are constrained, computational resources are limited and real-time performance is crucial [55]. In practice, preliminary tests and visual inspection revealed that these techniques exhibit consistent failure modes when applied to dynamic bridge inspection footage.
Thus, there is a compelling need to develop hybrid approaches that:
  • Aggregate temporal cues without relying solely on optical alignment
  • Provide interpretable outputs across evolving frame sequences
  • Support real-time deployment under hardware constraints.
The above challenges motivate the integration of lightweight temporal modules within a neuromorphic SNN pipeline, enabling both temporal sensitivity and energy efficiency, which in this paper, we consider as hallmarks of robust infrastructure diagnostics. By embedding temporal aggregation directly into a hybrid SIFT-SNN architecture, the proposed system aims to capture evolving structural anomalies while avoiding the computational overhead associated with optical alignment or dense spatiotemporal networks.

2.5. Gap Summary and Positioning

The preceding review reveals critical limitations in current automated infrastructure monitoring systems. While surface-level defect detection has advanced substantially, driven by large annotated datasets and deep convolutional architectures, existing models still fall short in several contexts such as:
  • Neglect of structural and modular components: Most ARDAD and SHM systems are optimised for surface anomalies and lack sensitivity to structural faults in mechanically interlocked or movable elements such as barrier pins and joints.
  • Lack of temporal reasoning: Per-frame classification approaches dominate the literature, often failing to capture progressive or transient anomalies that unfold over time.
  • Limited edge computing deployability: Deep CNNs and optical flow models typically require significant compute resources, making them impractical for real-time, low-power roadside units.
  • Underutilization of neuromorphic architectures: Despite the clear advantages of SNNs in energy efficiency and temporal encoding, their application to modular infrastructure inspection remains underexplored, particularly in multi-class or context-aware diagnostic tasks.
To address the current technology gaps, this study proposes a novel hybrid SIFT-SNN pipeline, extended with:
  • Temporal modelling across frame sequences for progressive anomaly detection;
  • Panoramic stitching evaluation under real-world egomotion and structural parallax conditions;
  • Multi-class classification of mechanical states (e.g., safe, partial, failed, ambiguous);
  • Latency and power benchmarking on edge-deployable neuromorphic hardware platforms.
Table 2 summarises the comparative positioning of the proposed method against the state-of-the-art across five key dimensions.
The comparative summary vs the state-of-the-art positioning underscores the novelty of the proposed framework as one of the first real-time, interpretable, and multi-class SNN-based systems applied to structural anomaly detection in modular infrastructure. Hence, the combined capabilities of the proposed SIFT-SNN (Table 2) shows strong supporting arguments for live deployment scenario on the Auckland Harbour Bridge.

3. Materials and Methods

This section presents the experimental design used to develop and evaluate the proposed context-aware SIFT-SNN anomaly detection system for structural safety diagnostics in modular transport infrastructure. The study leverages both a synthetic dataset, designed to simulate graded pin displacement anomalies under controlled conditions, and real-world video footage of the Auckland Harbour Bridge’s movable concrete barrier (MCB) system, used for qualitative validation and scenario benchmarking.
We first describe the construction of the synthetic dataset, including image generation protocols, defect labelling schema, and class balance strategy. Next, we outline the preprocessing pipeline, which includes grayscale normalisation, SIFT-based feature extraction, and spike encoding for neuromorphic processing. The hybrid pipeline architecture, comprising a handcrafted feature extractor and a Leaky Integrate-and-Fire (LIF) spiking neural network, is then detailed, followed by the temporal modelling extension used to integrate context across sequential frames.
Finally, we describe the anomaly classification and performance evaluation protocol, including multi-class confusion metrics, latency measurements under CPU-only constraints, and qualitative assessments of real-world footage. As part of a broader methodology reflection, we also report on a failed panorama stitching attempt, included here to illustrate the limitations of classical spatial alignment techniques under high-motion, occlusion-prone conditions.

3.1. Data Collection and Ground Truth Annotation

To support the development of a robust and context-aware anomaly detection framework for Auckland Harbour Bridge’s Movable Concrete Barrier (MCB) system, a composite dataset was constructed using real-world inspection footage and synthetically generated defect cases.

3.1.1. Real-World Data Acquisition

Three supervised field deployments were conducted using a custom rig mounted on the Barrier Transfer Machine (BTM). Multi-angle video footage of the barrier’s metal pin interfaces was captured using consumer-grade cameras (GoPro 5/8/9, iPhone 13 Pro with LiDAR, Samsung A7, and iPad 6). Devices were affixed to the BTM’s front and rear arms using vibration-dampening measures and powered externally to support prolonged high-frame-rate recording (up to 240 fps). Camera configurations and deployment conditions are summarised in Table 3.
Environmental variability included:
  • Lighting conditions: early morning, midday sun, overcast, night-time
  • Motion: slow traversal, full-speed operation, stop-start transitions
  • Visual noise: structural shadows, cable occlusions, surface glare.
The data collection recordings (Table 3) were primarily used for qualitative system validation, temporal stability checks, and failure analysis under deployment-like conditions.
Ground Truth Annotation
Individual video frames were manually annotated using LabelMe and internal Python tools. Each instance was labelled into one of four mutually exclusive alignment states:
  • ‘Safe’ … Pins fully seated and flush with block surfaces
  • ‘Partial Misalignment’ … Detectable tilt or partial protrusion
  • ‘Out’ … Significant dislocation or complete disengagement
  • ‘Ambiguous’ … Frames obscured by blur, occlusion, or poor lighting.
Annotations included:
  • Bounding box coordinates of the Region of Interest (ROI)
  • Class label
  • Environmental metadata (lighting, obstruction, session ID).
Annotation reliability was confirmed through dual review, with an inter-rater agreement of greater than 95% on a stratified subset.

3.1.2. Synthetic Frame Augmentation

Due to the natural scarcity of Full Misalignment cases, a procedurally controlled augmentation pipeline was used to enrich this underrepresented class. Based on our previous work, detailed in Rathee, Bačić and Doborjeh [11], this pipeline applied localised affine transformations, motion blur kernels, and pixel intensity perturbations to convert ‘Safe’ frames into realistic ‘Out’ i.e. full misalignment variants. Each synthetic frame was manually reviewed for structural plausibility by bridge inspection experts.
To avoid dataset imbalance or domain shift:
  • Augmentation was limited strictly to the Full Misalignment class
  • Synthetic-to-real ratio was capped at 30% of the total dataset
  • No alterations were applied to other well-represented classes.
This approach preserved statistical integrity while improving model sensitivity to rare but safety-critical anomaly cases.

3.2. Preprocessing Pipeline

The raw field footage and synthetic frame sequences underwent a structured preprocessing pipeline designed to extract stable, discriminative features for neuromorphic inference. The pipeline included temporal segmentation, region-of-interest enhancement, and SIFT-based keypoint extraction, enabling both spatial sparsity and robustness under deployment conditions.

3.2.1. Temporal Windowing and Segmentation

To incorporate short-term motion context and suppress transient noise, video sequences were segmented into overlapping windows of 3 to 5 frames. This technique promotes temporal smoothing, aids in ambiguous state disambiguation, and aligns with event-driven temporal modelling approaches common in neuromorphic vision systems [56].
To incorporate short-term motion context and suppress transient classification noise, each video sequence was segmented into overlapping windows of 3 to 5 frames. Within the windows (containing frame sequence), spatial features were extracted per frame and then aggregated through temporal smoothing mechanisms embedded in the downstream spike-based architecture. The overlapping window method helps to resolve ambiguous or noisy frames, caused by motion blur, occlusion, or lighting disruptions, by anchoring them to nearby, higher-confidence observations.
Figure 2 depicts the effect of ‘Ambiguous’ frames' appearance. For example, in a scenario involving partial occlusion, isolated frame classification fails to confidently assign labels to frames that are blurred or darkened. However, when processed within a 3-frame window, temporal dependencies allow contextual correction, resulting in more stable and accurate anomaly classification. This design mirrors the principles of event-driven neuromorphic processing, where temporal sparsity and continuity are leveraged for robust perception.

3.2.2. Region-of-Interest (ROI) Extraction and Enhancement

Each frame was cropped to a fixed ROI bounding box centred on the pin-hinge junction. Following this, the following image enhancement procedures were applied:
  • Grayscale conversion to reduce computational complexity and ensure SIFT feature stability under varying lighting [57].
  • Contrast Limited Adaptive Histogram Equalisation (CLAHE) to enhance contrast in shaded or overexposed regions locally [58].
  • Gaussian blurring (σ = 1.2–1.5) to reduce high-frequency noise and suppress spurious keypoint activation during SIFT extraction [57].
Figure 3 illustrates preprocessing outputs for Safe and Misaligned classes, highlighting differences in structural texture and keypoint density.

3.2.3. SIFT Keypoint Extraction and Normalisation

Keypoints were computed as local extrema in Difference of Gaussian (DoG) scale-space representations, followed by descriptor extraction as 128-dimensional feature vectors per keypoint [57]. Feature extraction used OpenCV’s cv2.SIFT_create() with the following configuration:
  • Contrast threshold: 0.04 (empirically tuned)
  • Top 100 keypoints per frame retained (sorted by response)
  • Output: 100 × 128 = 12,800-dimensional feature vectors
  • Zero-padding is applied when fewer than 100 keypoints are detected
This sparse, high-salience representation was used as input to the spike encoding module, described in Section 3.3.

3.2.4. Panorama Stitching (Exploratory Evaluation)

As an auxiliary investigation, we explored the feasibility of constructing panoramic sequences from temporally adjacent frames to enhance macro-structural context across sequential pins. This approach draws inspiration from SLAM pipelines [59] and visual corridor tracking [60].
Using SIFT-based keypoint matching followed by RANSAC-based homography estimation (via OpenCV’s cv2.Stitcher), three experiments were conducted as shown in Figure 4.
Despite initial promise, the panorama approach was ultimately excluded from the final pipeline due to the following challenges:
  • Inter-frame motion often exceeded homography constraints due to the BTM’s traversal speed.
  • Parallax effects and occlusions (e.g., barrier chains, dynamic shadows, surrounding vehicles) reduced descriptor stability.
  • The resulting stitched frames showed spatial distortions, high black-pixel dropout, and inconsistent ROI alignment.
  • Stitching failure rate was recorded at 78.2%, with a mean alignment error of 17.3 pixels (normalised).
Despite its exclusion, this experiment provided critical understanding of motion limits, feature drift, and calibration requirements under real-world bridge monitoring conditions. The analysis also informs future work involving synchronised multi-camera rigs or SLAM-integrated pipelines.

3.3. Spike Encoding and SNN Architecture

To leverage the efficiency and temporal precision of neuromorphic computation, the preprocessed SIFT descriptors were converted into latency-coded spike trains and passed through a biologically inspired Spiking Neural Network (SNN) for anomaly classification. The architecture and encoding process were optimised for deployment on edge hardware and aligned with principles of event-driven vision systems.

3.3.1. Spike Encoding Scheme

Each input frame was represented by a fixed 100 × 128 SIFT descriptor matrix, resulting in a 12,800-dimensional feature vector. The descriptors were L2-normalised to the [0, 1] range to ensure uniformity across lighting conditions and feature magnitudes.
A latency-based encoding strategy, specifically Time-to-First-Spike (TTFS), was employed to convert continuous features into sparse spike trains:
  • Features with higher magnitude fired earlier, with spike times inversely proportional to descriptor value.
  • Each descriptor dimension x i [ 0 ,   1 ] was mapped to a spike delay d i as:
d i = τ m a x   · ( 1 x i ) ,
where τ m a x = 10 is the spike window duration in distance timesteps.
  • Non-firing neurons (below threshold) were encoded as silent (no spike), preserving temporal sparsity and computational efficiency.
This strategy enabled the preservation of salience ordering in the spatial domain while maintaining temporal sparsity, a crucial trait for downstream SNN efficiency.
Figure 5 demonstrates the discriminative utility of the TTFS scheme, highlighting characteristic firing patterns across the four alignment classes.

3.3.2. SNN Architecture

The SNN model was implemented using snnTorch, a PyTorch-compatible framework for spiking neural network computations. The architecture comprised two fully connected Leaky Integrate-and-Fire (LIF) layers:
  • Input layer: 12,800 spike-encoded inputs.
  • Hidden Layer: 512 LIF neurons with membrane decay and dynamic threshold-based activation
  • Dropout: A 25% dropout is applied between layers to improve generalisation.
  • Output layer: 4 neurons, corresponding to the four annotation classes (Safe, Partial, Full, Ambiguous).
Neuron membrane potential evolved according to:
τ m d V ( t ) d t =   V ( t ) + I ( t ) ,
A spike is emitted whenever the membrane potential exceeds the threshold:
V ( t ) V t h
And the neuron state is updated according to:
S ( t ) = 1 ,   V ( t )   V r e s e t .
The Surrogate gradient descent was used during training, with a piecewise linear approximation to allow backpropagation through the non-differentiable spiking function.
Spikes were accumulated over the 10-timestep inference window, and the final class prediction was made based on the highest spike count across output neurons (Figure 6).
In Figure 6 input SIFT descriptors are latency-encoded into sparse temporal spike trains across a time window τ m a x = 10 , as illustrated in the raster input matrix (bottom left). The spike trains are propagated to a hidden layer comprising 512 LIF neurons, where membrane potentials evolve according to the differential equation:
d u i ( t ) d t =   u i ( t ) τ m +   w i x i ( t )
Here, u i ( t ) denotes the membrane potential, τ m is the membrane time constant, ​ w i is the synaptic weight, and x i ( t ) represents the incoming spike train. Spikes are generated when the membrane potential exceeds a threshold θ , governed by the spiking rule Θ ( u i ( t )   θ ) , where Θ ( · ) is the Heaviside function.
Dropout ( p = 0.25 ) is applied post-activation in both hidden and output layers to regularise training. The output layer consists of four LIF neurons corresponding to the four annotated classes (Safe, Partial, Out, Ambiguous). During training, backpropagation is enabled via surrogate gradients, using a piecewise linear approximation:
z i u i t =   { 1 γ ,     i f | u i t θ | < y 2 0 ,                                     o t h e r w i s e
This approximation allows for non-zero gradient flow near the firing threshold θ, enabling stable optimisation despite the non-differentiability of the spike function.

3.3.3. Training Configuration

The SNN was trained using the following hyperparameters and learning configuration:
  • Loss Function: Focal loss with γ = 2.0 and class-balancing factor α = 0.75 , to address class imbalance (especially under-represented Full and Ambiguous cases).
  • Optimiser: Adam (initial learning rate = 1e-3) and cosine annealing scheduler for adaptive decay.
  • Batch Size: 32
  • Epochs: 50
  • Cross-Validation: 5-fold stratified cross-validation, ensuring proportional representation of each class in train and validation splits.
The training setup prioritised both anomaly boundary sensitivity and stability under noisy or ambiguous conditions. All models were trained on an NVIDIA RTX 4060 GPU using PyTorch 2.1 and snnTorch 0.7.6. Model profiling was conducted to estimate memory and inference efficiency across CPU and GPU platforms.

3.4. Temporal Aggregation Layer

While the SIFT-SNN model provides high spatial salience, its per-frame predictions can fluctuate due to motion blur, vibration, or lighting transients, which are common in real-world deployments. To mitigate this instability, we introduce temporal aggregation layers that consolidate short-term evidence across frames, improving diagnostic consistency and interpretability.

3.4.1. Majority Voting: Stability Through Discrete Mode Estimation

Although neuromorphic systems excel in spike-based inference, their per-frame predictions may still exhibit temporal fragmentation under noisy conditions. We implemented a majority voting strategy over sliding windows (3–5 frames) to emulate discrete decision persistence, analogous to mode-seeking behaviour in decision-time cortical assemblies. The majority voting strategy over sliding windows, as a method, improves short-term consistency by suppressing transient misclassifications, especially when ambiguous activations momentarily spike. For example, handling of transient misclassification can be particularly effective in stabilising outputs near structural boundaries, such as momentary occlusions from barrier chains or light flares. Due to its simplicity and zero added computational cost, this strategy is ideal for real-time inference on constrained edge computing devices.

3.4.2. Weighted Temporal Smoothing: Bio-Inspired Confidence Accumulation

Inspired by the leaky evidence integration observed in biological perception, we applied an exponential moving average (EMA) filter to the softmax confidence scores to simulate a form of confidence inertia. Let be the softmax output at time t, the smoothed vector p ^ t was updated using:
p ^ =   α · p t + ( 1 α ) · p ^ t 1
With α = 0.6 , empirically selected. This soft fusion retained strong signals while damping volatility from transient events such as camera jitter or momentary blur. The final prediction was taken as p ^ t . Though slightly more expensive than voting, this method offered finer control over decision continuity and improved confidence tracking across frame sequences.

3.4.3. Recurrent Readout: Temporal Context as Memory

We further explored a recurrent readout mechanism via Gated Recurrent Units (GRUs) to learn dynamic patterns in spike-rate evolution. The GRU operated over temporal sequences of spike-accumulated class scores, aiming to model longer-range dependencies. For instance, GRU-based recurrent redout could potentially resolve ambiguous frames based on pre- and post-context, where clear pin alignment transitions occur. Although this method showed promise in preliminary validation (~2% F1 gain on ambiguous cases), the GRU-based recurrent redout was excluded from the final deployment due to increased training complexity and incompatibility with low-power edge inference. Nevertheless, GRU-based processing mechanism establishes a viable path for future extension towards bio-inspired memory-integrated neuromorphic pipelines.
Together, combined aggregation strategies are to enhance the temporal robustness of the SIFT-SNN framework, transforming a frame-level anomaly detector into a temporally aware system capable of sustaining coherent safety assessments under operational noise, visual drift, and motion discontinuities.

3.5. Evaluation Metrics

To assess the effectiveness and deployability of the proposed SIFT-SNN anomaly detection pipeline, we conducted a multidimensional evaluation that combined classification metrics, runtime profiling, and resource analysis. Multidimensional evaluations were intended to reflect both diagnostic performance and real-time feasibility for edge deployment in a dynamic traffic infrastructure context.

3.5.1. Classification Metrics

Model performance and generalisation capability were assessed using 5-fold stratified cross-validation on a curated dataset of 6,000 annotated frames, with 1,200 frames used for held-out testing. The evaluation focused on standard multi-class classification metrics:
  • Accuracy: Overall percentage of correct predictions.
  • Precision, Recall, F1 Score: Reported per class (Safe, Partial Misalignment, Full Misalignment, Ambiguous), with macro-averaging applied to mitigate class imbalance effects, and in particular, for the Partial and Ambiguous classes.
  • Confusion Matrix: Aggregated across folds to highlight systematic misclassification patterns, especially confusion between Safe and Partial Misalignment near boundary conditions.
Figure 7. Four-class confusion matrix summarising model predictions across the Safe, Partial, Full Misalignment, and Ambiguous classes (N = 1,200 test frames).
Figure 7. Four-class confusion matrix summarising model predictions across the Safe, Partial, Full Misalignment, and Ambiguous classes (N = 1,200 test frames).
Preprints 187765 g007
Temporal aggregation strategies described in Section 3.4 were tested in both raw and smoothed inference modes. Notably, exponential moving average fusion improved F1 scores for the Ambiguous and Full Misalignment classes by 2.1% and 1.5%, respectively, demonstrating the value of sequence-aware prediction refinement in dynamic video settings.

3.5.2. Real-Time Inference and Deployment Profiling

Beyond classification accuracy, we evaluated the system’s computational performance to ensure suitability for embedded edge deployment. Inference latency, memory footprint, and model size were benchmarked across the following platforms:
  • NVIDIA RTX 4060 (desktop GPU)
  • Intel Core i7-11800H (CPU-only mode)
  • Jetson Nano (projected via emulated FP16 inference using TensorRT)
The Jetson Nano metrics were empirically projected by constraining the GPU frequency and power on an RTX 4060 to approximate the Nano’s hardware profile. Inference was performed using a TensorRT-converted version of the trained SNN model (via torch2trt) with FP16 acceleration. This setup yielded consistent latency and memory usage within the expected Jetson Nano performance envelope (Table 4).

3.5.3. Performance Summary

The findings on deployment options and power consumption (Table 4) validate the model’s readiness for Phase 2 deployment in live operational environments.
Comprehensive evaluation results (Table 4) confirms that the SIFT-SNN architecture achieves an effective balance between classification accuracy, low-latency performance, and energy efficiency, hence, meeting the constraints required for embedded anomaly detection on systems such as the Barrier Transfer Machine (BTM).

4. Results and Evaluation

This section presents the empirical evaluation of the proposed hybrid anomaly detection framework that integrates SIFT-based feature extraction, spike-based neural inference, and temporal aggregation mechanisms. The assessment focuses on three core aspects:
  • Classification performance across defined structural alignment classes.
  • Temporal stability under real-world environmental variability.
  • Feasibility for real-time, low-power deployment.
Our results demonstrate that biologically inspired computation, when combined with handcrafted descriptors, can outperform or complement conventional deep learning, enabling efficient, interpretable, and deployable safety inspection solutions for critical infrastructure.

4.1. Experimental Setup

To rigorously assess the proposed SIFT-SNN pipeline, we designed a multi-faceted evaluation guided by three principal objectives:
(i) Achieving high classification accuracy across four semantic classes: Safe, Partial Misalignment, Full Misalignment, and Ambiguous;
(ii) Ensuring temporal consistency in the presence of noise, occlusion, or lighting variation;
(iii) Validating computational efficiency on platforms suitable for embedded edge deployment.

4.1.1. Dataset and Preprocessing

Video data were acquired during three real-world field deployments on Auckland Harbour Bridge’s Movable Concrete Barrier (MCB) system. All frames were manually annotated and stratified to ensure balanced class representation. The preprocessing pipeline included:
  • Region of interest (ROI) cropping,
  • Grayscale conversion,
  • CLAHE-based contrast enhancement,
  • SIFT keypoint extraction,
  • Latency-based spike encoding.
The complete training and inference pipeline is illustrated in Figure 8.

4.1.2. Model Architecture and Training

Inference was powered by a two-layer Leaky Integrate-and-Fire (LIF) SNN implemented using snnTorch 0.7.6 on PyTorch 2.1. The network was trained with focal loss (γ = 2.0, α = 0.75), using the Adam optimiser with a cosine annealing scheduler for 50 epochs. To ensure robust generalisation across environmental conditions, we employed a 5-fold stratified cross-validation scheme.
Classification results (Table 5) are reported with macro-averaged scores summarising overall performance.

4.1.3. Deployment Evaluation and Profiling

To assess real-time viability, we profiled inference latency, memory footprint, and power consumption using NVIDIA Nsight Systems and torch.profiler. Benchmarking was conducted across three target platforms:
  • RTX 4060 GPU (desktop)
  • Intel i7-11800H CPU (CPU-only mode)
  • Jetson Nano (simulated using downclocked GPU and FP16 inference on RTX 4060)
Model inference used torch2trt to convert the trained SNN into a TensorRT-accelerated version suitable for embedded inference. Table 6 summarises deployment benchmark results.
The results (Table 6) confirm the model’s feasibility for edge deployment, with minimal degradation in performance when emulated on Jetson Nano.

4.1.4. Temporal Aggregation and Sequence Stability

To enhance robustness under challenging visual conditions (e.g., motion blur, occlusion), we evaluated temporal aggregation strategies:
  • Majority voting
  • Exponential smoothing.
Both methods were applied to sequences of 3–5 frames. Results showed a reduction in transient misclassifications and improvements in temporal stability.
A lightweight GRU-based readout was also explored to capture longer-term dependencies. While the GRU-based redout showed improved consistency in exploratory tests, it was ultimately excluded from the final deployment configuration due to increased training complexity and incompatibility with embedded platforms.

4.2. Temporal Aggregation Performance

While per-frame classification provides a foundation for anomaly detection, real-world deployment demands temporal consistency, particularly in the presence of motion blur, partial occlusion, or vibration-induced noise. To address this, we integrated lightweight temporal aggregation mechanisms aimed at stabilising predictions and reducing spurious anomaly triggers from transient artifacts.
We compared three temporal fusion methods:
  • Majority voting: For each 3–5 frame clip, predicted class labels were aggregated using a simple majority rule. This helped smooth misclassifications in borderline cases, especially for the ‘Ambiguous’ class.
  • Exponential smoothing: A soft fusion technique using an Exponential Moving Average (EMA) of softmax probabilities. For each time step t, the smoothed score p ^ t was calculated as:
p ^ t = α · p t + ( 1 α ) · p ^ { t 1 } ,   w h e r e   α = 0.6
This approach preserved peak activations while filtering fluctuations near class decision boundaries.
  • GRU readout (exploratory): A Gated Recurrent Unit (GRU) model was placed atop spike-accumulated class scores to learn sequence-level patterns. Although promising in smoothing ambiguous transitions, the GRU model was excluded from final deployment due to its increased model complexity and hardware overhead.
Both majority voting and EMA smoothing yielded significant gains in robustness without incurring additional computational complexity. Notably, EMA fusion offered the best trade-off between predictive consistency and deployment efficiency, achieving a macro F1 improvement of 1.8% over frame-wise classification.
The obtained results (Table 7) reinforce the value of integrating temporal context into spike-based classification pipelines. Even simple sequence-aware fusion strategies can markedly enhance the reliability, interpretability, and field-readiness of neuromorphic inspection systems operating under noisy or occluded conditions.

4.3. Ablation Studies

To quantify the contribution of each component in the proposed SIFT-SNN framework, we conducted a series of ablation experiments. The intend of the ablation studies is isolate the role of engineered SIFT descriptors, spike encoding mechanisms, and neuromorphic architecture to justify their inclusion, contribution and evaluate alternatives.

4.3.1. SIFT-Only Baseline (MLP Classifier)

In this experiment, we bypassed spike generation entirely and fed the raw 12,800-dimensional SIFT descriptor vector directly into a two-layer fully connected MLP classifier. The architecture was kept similar to the SNN (Input–512–Dropout–4), but with standard ReLU activations and softmax output. The model was trained under identical conditions using the focal loss and the same data splits for cross-validation.
The resulting F1 scores were consistently 5–7% lower across all classes compared to the full SIFT-SNN pipeline. The model also showed greater overfitting and sensitivity to ambiguous frames. This suggests that while SIFT descriptors contain rich spatial information, their full potential is only unlocked when paired with temporally-aware, event-driven processing.

4.3.2. Random Input + SNN

To assess the importance of structured feature extraction, we tested the SNN using randomly generated Gaussian input vectors (same dimensions as SIFT descriptors). These synthetic vectors were normalised and spike-encoded using the same latency-based scheme.
The resulting model failed to converge meaningfully, achieving macro F1 scores below 50% and demonstrating erratic behaviour across validation folds. This confirms that the SIFT descriptors serve as critical low-level features, encoding edge and texture information that the SNN cannot infer from unstructured input.

4.3.3. Comparison of Spike Encoding Schemes

We further evaluated alternative spike encoding strategies:
  • Latency coding (TTFS): Our proposed method, where spike time is inversely proportional to descriptor magnitude.
  • Rate coding: Input values were mapped to spike counts over a 10-time-step window.
  • Direct analogue input: SIFT vectors were passed into the same LIF-based network without spike conversion (i.e., treating them as continuous inputs in surrogate-SNN mode).
Latency-based encoding (Table 8) produced the sparsest spike trains with the highest accuracy, validating its effectiveness for anomaly detection. Rate coding incurred additional latency and energy cost due to higher firing activity. The analogue input method, although viable, failed to capitalise on the temporal efficiency of event-driven representations.

5. Discussion

5.1. Class-Wise Performance Interpretation

The experimental results presented in Chapter 4 confirm that the proposed SIFT-SNN architecture achieves consistently high performance across all four alignment classes. Under frame-wise evaluation, the system attained a macro F1 score of 89.9% ( Table 5), which further improved to 92.5% with sequence-aware fusion (Table 7). Both metrics demonstrate the model's robustness and suitability for real-world anomaly detection in dynamic traffic infrastructure settings.
Performance was strongest on the ‘Safe’ and Full Misalignment (‘Out’) classes, both of which exhibit distinct visual characteristics that align well with the SIFT feature space. The ‘Safe’ class is characterised by high structural regularity and well-aligned pin patterns, enabling the SNN to extract salient features through reliable early spike activation. In contrast, the ‘Out’ class (that is, typically representing synthetic or extreme misalignments) displays prominent geometric discontinuities (e.g., protruding pins or pin loss), leading to high-magnitude SIFT descriptors and temporally early spiking. This clear feature separation contributed to F1 scores exceeding 90% for both categories.
More challenging were the ‘Partial’ and especially ‘Ambiguous’ classes, which frequently involved motion blur, partial occlusion, or variable lighting conditions that reduce both keypoint stability and temporal spike consistency. Despite such difficulties, the system maintained robust performance, achieving an F1 score of 85.3% on the ‘Ambiguous’ class under frame-wise inference. The obtained resilience of system performance is attributed to the synergistic integration of handcrafted SIFT descriptors (known for their robustness to illumination and scale variations) and the spike-based architecture’s temporal contrast sensitivity, which allows the classifier to focus on dynamic changes rather than static pixel intensities.
The obtained results highlight the complementary strengths of the SIFT-SNN hybrid approach. SIFT descriptors provide interpretable, spatially meaningful representations of structural alignment, while the spiking neural network confers noise tolerance and temporal sparsity. As hybrid approach, the combination of SIFT-SNN enables reliable detection even in visually degraded or ambiguous field conditions, outperforming both conventional MLP baselines and CNN architectures in challenging scenarios.

5.2. Comparison with Traditional and Deep Learning Methods

To contextualise the performance of the proposed SIFT-SNN framework, we benchmark it against traditional CNNs and modern deep architectures commonly employed in visual anomaly detection. Prior evaluations of models such as ResNet-50, MobileNetV2, and EfficientNet-B0 demonstrated strong accuracy but suffered from high memory usage, slower inference, and limited interpretability. Recent transformer-inspired models, such as ConvNeXt-T and ViT-B/16, offer slight gains in accuracy but further exacerbate deployment constraints due to their size and computational demands.
By contrast, the SIFT-SNN system achieves comparable performance (91.1% macro F1) while maintaining a lightweight footprint (8.3 MB) and low inference latency (9.5 ms), making it well-suited for edge deployment. Its hybrid design leverages precomputed SIFT descriptors and spiking neural dynamics, enabling event-driven, low-power operation, which represents a key advantage for embedded platforms such as Jetson Nano or Intel Loihi.
Moreover, the system offers improved transparency, as visual SIFT keypoints and spike timing patterns enable post hoc inspection of classification decisions, thereby addressing the black-box limitations of conventional deep learning. This makes the framework particularly suitable for safety-critical contexts that demand explainability.
A summary of performance, latency, and deployment factors across evaluated models is presented in Table 9, illustrating the SIFT-SNN’s balanced trade-off between accuracy, efficiency, and interpretability.

5.3. Significance of Temporal Aggregation

While frame-wise classification provides a basic mechanism for anomaly detection, real-world deployment demands temporal consistency, particularly under conditions such as motion blur, partial occlusion, or lighting variation. Temporal aggregation addresses this by incorporating contextual cues from adjacent frames, stabilising predictions and improving robustness.
As shown in Table 7, fusion strategies such as majority voting, exponential smoothing (EMA), and GRU-based readout significantly enhance performance, especially for ambiguous or transitional cases. Considering relatively narrow decision boundaries between the classes, it is common to find subtle visual cues that are difficult to discern in isolation (or articulate from a domain expert interview) but could become more apparent over time.
Among the tested methods, EMA proved most effective for deployment, improving the macro F1 score from 89.9% (frame-wise) to 91.7% while maintaining low computational overhead. By smoothing class probabilities, EMA reduces transient misclassifications and prevents erratic label switching.
The GRU-based readout achieved the highest F1 score (92.5%) in exploratory testing, thanks to its ability to model temporal dependencies across longer sequences. However, its training complexity and runtime demands render it unsuitable for edge deployment in its current form.
The results above affirm the importance of temporal modelling in safety-critical systems. Lightweight fusion techniques, such as EMA, offer a practical balance between stability and efficiency, aligning well with the temporal dynamics of spiking architectures. By integrating such mechanisms, the system approximates human-like reasoning in terms of using context and continuity to enhance decision quality.

5.4. Reflections on Spatial Context Integration Attempts

As outlined in Chapter 1, early experimentation with spatial context integration via panorama stitching exposed several practical limitations. Keypoint-based compositing methods frequently failed due to motion blur, parallax effects between independently moving barrier segments, and varying illumination conditions. These factors produced unstable or distorted stitched frames, which undermined the feasibility of spatial aggregation for anomaly detection. Given these limitations and the computational cost of real-time stitching, the study shifted toward temporal aggregation as a more reliable mechanism for capturing contextual cues across sequential frames. This transition aligns with the empirical findings in Section 5.3, where temporal fusion improves robustness under the same conditions that caused spatial methods to fail.
5.5:. Deployment Readiness and Edge Viability
A primary goal of the SIFT-SNN framework is robust deployment on resource-constrained platforms, such as the Barrier Transfer Machine (BTM), which operates on the Auckland Harbour Bridge. Unlike conventional deep learning models that demand GPU-level compute and continuous connectivity, our design emphasises efficiency, modularity, and real-time viability on edge devices.
Deployment benchmarks (Table 6) across three hardware configurations: RTX 4060 GPU, Intel i7 CPU, and a Jetson Nano simulation demonstrate the system’s practicality. The Jetson Nano simulation achieved an inference latency of 28.4 ms per frame, which is well within the 30 ms threshold for real-time monitoring of high-speed MCB movement.
The pipeline’s compact model size (8.3 MB) and modest memory usage (≤1.4 GB) further reinforce its suitability for embedded systems. In contrast to transformer or CNN-based models that suffer from large footprints and thermal throttling, SIFT-SNN’s sparse spike-based architecture enables low-overhead execution with minimal power and thermal stress.
All stages (from frame acquisition to inference) were validated under stream-compatible, batch-independent execution, ensuring live video compatibility without buffering-induced delays. Runtime profiling using NVIDIA Nsight and PyTorch’s torch.profiler confirmed stable performance, with inference loads staying below 70% on Jetson-class platforms. This leaves ample processing headroom for additional tasks such as telemetry and alerting.
Although the Jetson Nano benchmarks are simulated, they align with published baselines and were cross-validated using throttled tests on high-end hardware. Future work will involve deployment on physical edge platforms (e.g., Jetson Orin Nano, Intel Loihi) to measure energy draw and operational reliability in live environments.
In summary, the SIFT-SNN framework fulfils stringent real-time and power constraints, offering a deployable, interpretable, and scalable solution for intelligent infrastructure monitoring.

5.6. Limitations and Future Directions

While the SIFT-SNN framework demonstrates strong performance and deployment readiness, several aspects warrant consideration to enhance generalisation, scalability, and long-term adaptability, such as:
  • Domain adaptation for diverse infrastructure:
The current model is optimised for the Auckland Harbour Bridge’s Movable Concrete Barrier (MCB), with spike encoding and feature extraction tailored to its geometry and operational context. While this ensures high task-specific performance, adaptation to other infrastructure types (such as tunnels, rails, or anchor systems), would likely require retraining or adjustments to the encoding pipeline. Addressing cross-domain generalisation remains an active area for future research in neuromorphic vision.
2.
Synthetic data dependency:
To compensate for the rarity of unsafe pin conditions, the ‘Out’ class incorporates high-fidelity synthetic samples that domain experts have validated. Although effective in training, synthetic data may not fully capture environmental nuances, particularly under novel lighting or weather conditions. Expanding real-world datasets will further strengthen robustness and reduce domain gaps.
3.
Annotation and calibration requirements:
The current system depends on expert-labelled data and manual calibration of key parameters (e.g., ROI, exposure, spike window). While manageable at the prototype scale, the labour-intensive tasks would pose challenges for large-scale deployment. Incorporating semi-automated annotation tools and self-calibrating mechanisms would enhance efficiency and reproducibility.
4.
Rare class representation:
Edge cases, such as full misalignment or ambiguous transitions remain underrepresented in training. Although focal loss and cross-validation techniques help balance the learning process, performance on such classes remains lower than that for common conditions. Future data acquisition should prioritise for minority scenarios to improve system resilience and individual class performance.
5.
Limited on-device adaptation:
The current training pipeline operates offline on GPU hardware, with no support for on-chip learning. While this is consistent with most neuromorphic systems, enabling edge-level fine-tuning or drift adaptation (e.g., via continual or few-shot learning) would be beneficial for long-term autonomous deployment.
6.
Simulated edge performance:
Jetson Nano performance was estimated using a controlled simulation. While cross-validation with throttled hardware mitigates discrepancies, full empirical evaluation on physical edge devices (e.g., Jetson Orin Nano, Intel Loihi) remains essential for deployment certification in regulated environments.

5.7. Contribution to Neuromorphic and Safety AI Research

This study advances the practical deployment of neuromorphic computing in safety-critical infrastructure by introducing a hybrid framework that combines interpretable, handcrafted features with biologically inspired spiking inference. By addressing challenges of transparency, latency, and energy efficiency, our work contributes to a new class of real-world-ready neuromorphic systems, as follows:
  • Spike-based visual perception at scale:
We show that combining SIFT descriptors with latency-coded Spiking Neural Networks (SNNs) yields competitive classification performance in a complex real-world setting. The use of time-to-first-spike encoding and LIF dynamics enables sparse yet salient visual inference, extending the applicability of spike-based computing beyond synthetic benchmarks.
2.
Interpretable AI for critical infrastructure:
The use of SIFT retains spatial interpretability, enabling traceable decisions through keypoint visualisation and spike dynamics. Unlike black-box CNNs, this architecture supports transparency, a crucial requirement for infrastructure diagnostics where model explainability can impact operational decisions and regulatory compliance.
3.
Operational neuromorphic deployment:
Our work enables neuromorphic vision to transition from lab environments to embedded systems. The system’s 9.5 ms inference on GPU and 28.4 ms latency under Jetson Nano simulation demonstrate real-time viability for field deployment which is particularly important for near-future AI computing expectations as well as in energy-constrained scenarios, such as mobile bridge inspection platforms.
4.
Architectural innovation through hybrid design:
Rather than adopting end-to-end deep models, we integrate classical computer vision with spiking neural networks and inference. The presented hybrid approach offers a compelling trade-off: interpretable and robust spatial encoding from SIFT, which is combined with temporally aware, low-power classification from SNNs. As computational approach, our design reflects a biologically-inspired AI strategy aligned with both cognitive principles and hardware implementation realities.

5.8. Future Directions

Building on the success of the SIFT-SNN framework, several promising directions can further extend its applicability, robustness, and scalability for real-world deployment in safety-critical infrastructure.
  • Multimodal sensor fusion:
Integrating complementary inputs, such as LiDAR, thermal imaging, or IMUs, could enhance detection in poor visibility or complex environments. Fusing multimodal sensor fusion with visual spike-based inference would strengthen anomaly localisation and resilience.
2.
3D tracking and multi-camera integration:
Incorporating SLAM and synchronised multi-camera views can enable spatiotemporal tracking of structural changes. This would allow not only detection but precise localisation of faults across complex infrastructure segments.
3.
Neuromorphic hardware deployment:
Porting the system to physical platforms, such as Intel Loihi or Jetson Orin, will enable empirical validation of power, latency, and thermal constraints. Benchmarking on real hardware is critical for regulatory compliance and edge deployment.
4.
Cross-domain transfer and adaptation:
Investigating techniques could facilitate deployment with minimal retraining in new environments including diverse infrastructure (e.g., tunnels, railways), and exploration of AI applications such as transfer learning and adaptive spike encoding strategies.
5.
Explainability and human-in-the-loop alerts:
Future versions could include real-time visual explanations or keypoint heatmaps to assist operators in interpreting alerts, thereby improving trust and situational awareness in semi-autonomous inspection systems.
6.
Continual learning and adaptation:
Incorporating mechanisms for online learning would allow the system to adapt over time to new anomaly patterns or environmental drift, supporting long-term deployment without complete retraining cycles.
This section has discussed the novel SIFT-SNN framework as a way of striking a practical balance between accuracy, interpretability, and edge efficiency. Through neuromorphic computation, temporal smoothing, and compact architecture, it addressed the key challenges in infrastructure monitoring. The outlined future directions aim to broaden the utility of this approach across modalities, hardware, and deployment scenarios, paving the way for scalable, explainable, and adaptive neuromorphic AI in safety-critical domains.

6. Conclusion

This study presents a hybrid neuromorphic framework that combines handcrafted SIFT features with biologically inspired spiking neural networks (SNNs) for real-time anomaly detection in traffic flow control infrastructure. The model was validated against the Auckland Harbour Bridge’s movable barrier system under realistic deployment conditions, including vibration, occlusion, and low-light conditions.
The key contributions are summarised as follows:
  • Developed a deployable hybrid system combining handcrafted vision features with neuromorphic inference.
  • Demonstrated low-latency, energy-efficient performance suitable for embedded infrastructure monitoring.
  • Advanced the interpretability and transparency of safety AI through spike-based processing and explainable features.
  • Provided a reproducible evaluation framework spanning accuracy, temporal coherence, and edge readiness.
The proposed system achieved strong classification accuracy (up to 92.5% macro F1 with temporal fusion) while maintaining real-time, low-power inference on edge devices. Its architecture, featuring latency-coded spikes and LIF dynamics, demonstrates high efficiency and interpretability, with simulated Jetson Nano performance confirming sub-30 ms latency and minimal resource overhead.
Temporal aggregation strategies further enhanced stability and robustness, enabling consistent classification across ambiguous or transitional frames. Comparative evaluations against conventional CNNs reinforced the advantages of spike-based computation in edge-deployable, explainable AI systems. The integration of temporal feature aggregation into the SIFT-SNN pipeline played a key role in improving stability and spatiotemporal context awareness, thereby strengthening real-world deployability.
The findings collectively address the research questions posed at the outset. Temporal aggregation improved anomaly detection beyond the static SIFT-SNN baseline, while maintaining real-time performance across edge platforms, confirming the framework’s deployability. The multi-class formulation further enabled more nuanced safety diagnostics aligned with graded failure modes observed in the field.
Despite current limitations in domain transferability and class imbalance, the system establishes a viable template for deploying neuromorphic vision in safety-critical infrastructure.
Future work will extend the generalisability and transferability of the artefacts presented in the new case studies, as well as in new human-computer interaction paradigms for various near-future advancements in computer vision, sensors, and deployment infrastructures.

Author Contributions

Conceptualisation, M.R. and B.B.; Methodology, M.R. and B.B.; Software, M.R.; Validation, M.R.; Formal analysis, M.R. and B.B.; Investigation, M.R.; Data curation, M.R.; Writing and original draft preparation, M.R.; Writing, review and editing, M.R. and B.B.; Visualisation, M.R.; Pilot data collection protocol, B.B.; Project administration, B.B.; Industry involvement, B.B.; Supervision, B.B. and M.D.

Funding

This research was funded by the Auckland University of Technology Doctoral Fees and Stipend Scholarship.

Data Availability Statement

Data supporting the findings of this study are available from the corresponding author upon reasonable request. Some data may not be publicly shared due to security or privacy constraints related to access to infrastructure.

Acknowledgments

The authors would like to thank Gary Bonser, Martin Olive, David Ranby, and Angela Potae from Waka Kotahi New Zealand Transport Agency (NZTA) and Auckland System Management (ASM) for their continued support with field access, transport logistics, operational safety, and domain expertise that enabled data collection on the Auckland Harbour Bridge.
We are also grateful to the Auckland University of Technology (AUT) for PhD funding support and for providing access to high-performance computing infrastructure and audiovisual equipment critical for fieldwork and model training.
The research benefited from publicly available libraries and tools developed by open-source and academic communities, including OpenCV, PyTorch, snnTorch, and related Python ecosystems. Simulation of embedded deployment was enabled through NVIDIA Nsight Systems and related profiling utilities, which were instrumental in estimating Jetson Nano performance characteristics.
During manuscript preparation, the authors used OpenAI's ChatGPT (GPT-4, July 2025) for assistance with technical writing and language refinement. All diagrams and figures were designed and finalised independently using Canva and draw.io. The authors have reviewed and validated all outputs and take full responsibility for the content and conclusions presented.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Abbreviation Full Form
AUT Auckland University of Technology
BTM Barrier Transfer Machine
CNN Convolutional Neural Network
CLAHE Contrast Limited Adaptive Histogram Equalization
EMA Exponential Moving Average
F1 F1 Score (harmonic mean of precision and recall)
FPS Frames Per Second
GRU Gated Recurrent Unit
IoT Internet of Things
LIF Leaky Integrate-and-Fire
MCB Movable Concrete Barrier
MLP Multi-Layer Perceptron
NZTA New Zealand Transport Agency
ROI Region of Interest
SIFT Scale-Invariant Feature Transform
SNN Spiking Neural Network
TTFS Time-To-First-Spike

References

  1. Centers for Disease Control and Prevention (CDC). Global road safety. Available online: https://www.cdc.gov/transportation-safety/global (accessed on 15 June).
  2. Global Road Safety Facility. Global Road Safety Facility annual report 2023. Available online: https://documents1.worldbank.org/curated/en/099634305072412983/pdf/IDU14bccf9fb16f8914d9f1a3861ebb97d26441e.pdf (accessed on 15 June).
  3. World Bank. Roads, total network (km) - World development indicators. Available online: https://databank.worldbank.org/metadataglossary/world-development-indicators/series/IS.ROD.TOTL.KM (accessed on 15 June).
  4. World Health Organisation. Global status report on road safety: Time for action. Available online: https://www.who.int/publications/i/item/9789241563840 (accessed on 15 June).
  5. Chen, S.; Kuhn, M.; Prettner, K.; Bloom, D.E. The global macroeconomic burden of road injuries: Estimates and projections for 166 countries. The Lancet Planetary Health 2019, 3, e390–e398. [Google Scholar] [CrossRef]
  6. United Nations Economic Commission for Europe (UNECE). Statistics of road traffic accidents in Europe and North America – 2023 edition. Available online: https://unece.org/transport/publications/2023-statistics-road-traffic-accidents (accessed on 29 June).
  7. Reuters. Death toll in Gujarat bridge collapse rises to 132. Available online: https://www.reuters.com/world/india/death-toll-gujarat-bridge-collapse-rises-132-2022-10-31/ (accessed on 15 June 2025).
  8. News, B. Genoa bridge collapse: What we know. Available online: https://www.bbc.com/news/world-europe-45193452 (accessed on 15 June).
  9. Administration, F.H. Bridge Inspection Program. Available online: https://www.fhwa.dot.gov/bridge/inspection/ (accessed on 15 June).
  10. Chen, K.; Zhou, X.; Bao, Z.; Skibniewski, M.J.; Fang, W. Artificial intelligence in infrastructure construction: A critical review. Frontiers of Engineering Management 2024, 1–15. [Google Scholar] [CrossRef]
  11. Rathee, M.; Bačić, B.; Doborjeh, M. Automated road defect and anomaly detection for traffic safety: A systematic review. Sensors 2023, 23, 5656. [Google Scholar] [CrossRef] [PubMed]
  12. New Zealand Transport Agency Waka Kotahi. Auckland Harbour Bridge factsheet. Available online: https://www.nzta.govt.nz/assets/site-resources/content/about/docs/auckland-harbour-bridge-factsheet.pdf (accessed on 10 April).
  13. New Zealand Transport Agency Waka Kotahi. How to move a concrete motorway barrier. Available online: https://www.nzta.govt.nz/media-releases/how-to-move-a-concrete-motorway-barrier/ (accessed on 10 April).
  14. Cottrell, B.H. Evaluation of a movable concrete barrier system. Available online: https://rosap.ntl.bts.gov/view/dot/19352 (accessed on 2 January).
  15. Li, M.; Wang, X.; Wang, H.; Yang, M. LTGS-Net: Local Temporal and Global Spatial Network for Weakly Supervised Video Anomaly Detection. Sensors 2025, 25, 4884. [Google Scholar] [CrossRef] [PubMed]
  16. Sooksatra, S.; Watcharapinchai, S. A comprehensive review on temporal-action proposal generation. Journal of Imaging 2022, 8, 207. [Google Scholar] [CrossRef]
  17. Shahsavarani, S.; Lopez, F.; Ibarra-Castanedo, C.; Maldague, X.P. Advanced image stitching method for dual-sensor inspection. Sensors 2024, 24, 3778. [Google Scholar] [CrossRef]
  18. Zhu, X.; Liu, Z.; Zhang, X.; Sui, T.; Li, M. A very fast image stitching algorithm for PET bottle caps. Journal of Imaging 2022, 8, 275. [Google Scholar] [CrossRef]
  19. Rathee, M.; Bačić, B.; Doborjeh, M. Hybrid SIFT-SNN for Efficient Anomaly Detection of Traffic Flow-Control Infrastructure. arXiv 2025, arXiv:2511.21337. [Google Scholar] [CrossRef]
  20. Tang, X.; Zhang, Z.; Qin, Y. On-road object detection and tracking based on radar and vision fusion: A review. IEEE Intelligent Transportation Systems Magazine 2021, 14, 103–128. [Google Scholar] [CrossRef]
  21. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road crack detection using deep convolutional neural network. In Proceedings of the 2016 IEEE international conference on image processing (ICIP), 2016; pp. 3708–3712. [Google Scholar]
  22. Maeda, H.; Sekimoto, Y.; Seto, T.; Kashiyama, T.; Omata, H. Road damage detection and classification using deep neural networks with smartphone images. Computer-Aided Civil and Infrastructure Engineering 2018, 33, 1127–1141. [Google Scholar] [CrossRef]
  23. Yeum, C.M.; Dyke, S.J. Vision-based automated crack detection for bridge inspection. Computer-Aided Civil and Infrastructure Engineering 2015, 30, 759–770. [Google Scholar] [CrossRef]
  24. Zhou, K.; Wang, Z.; Ni, Y.-Q.; Zhang, Y.; Tang, J. Unmanned aerial vehicle-based computer vision for structural vibration measurement and condition assessment: A concise survey. Journal of Infrastructure Intelligence and Resilience 2023, 2, 100031. [Google Scholar] [CrossRef]
  25. Eisenbach, M.; Stricker, R.; Seichter, D.; Amende, K.; Debes, K.; Sesselmann, M.; Ebersbach, D.; Stoeckert, U.; Gross, H.-M. How to get pavement distress detection ready for deep learning? A systematic approach. In Proceedings of the 2017 international joint conference on neural networks (IJCNN), 2017; pp. 2039–2047. [Google Scholar]
  26. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. BDD100K: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 13-19 June 2020, 2020; pp. 2633–2642. [Google Scholar]
  27. Amhaz, R.; Chambon, S.; Idier, J.; Baltazart, V. Automatic crack detection on two-dimensional pavement images: An algorithm based on minimal path selection. IEEE Transactions on Intelligent Transportation Systems 2016, 17, 2718–2729. [Google Scholar] [CrossRef]
  28. Zhu, J.; Chen, Y.; Heng, J.; Wu, M.; Zhang, Y.; Li, Y. Probabilistic corrosion-fatigue prognosis of rib-to-deck welded joints in coastal weathering steel bridges exposed to heavy traffics. International Journal of Fatigue 2024, 182, 108210. [Google Scholar] [CrossRef]
  29. Kong, X.; Liu, Z.; Liu, H.; Hu, J.; Deng, L. Recent advances on inspection, monitoring, and assessment of bridge cables. Automation in Construction 2024, 168, 105767. [Google Scholar] [CrossRef]
  30. Graves, W.; Aminfar, K.; Lattanzi, D. Full-scale highway bridge deformation tracking via photogrammetry and remote sensing. Remote Sensing 2022, 14, 2767. [Google Scholar] [CrossRef]
  31. Luo, K.; Kong, X.; Zhang, J.; Hu, J.; Li, J.; Tang, H. Computer vision-based bridge inspection and monitoring: A review. Sensors 2023, 23, 7863. [Google Scholar] [CrossRef]
  32. Kim, I.B.; Cho, J.S.; Zi, G.S.; Cho, B.S.; Lee, S.M.; Kim, H.U. Detection and identification of expansion joint gap of road bridges by Machine Learning using line-scan camera images. Applied System Innovation 2021, 4, 94. [Google Scholar] [CrossRef]
  33. Rathee, M.; Bačić, B.; Doborjeh, M. Hybrid machine learning for automated road safety inspection of Auckland Harbour Bridge. Electronics 2024. [Google Scholar] [CrossRef]
  34. Luo, H.; Cai, L.; Li, C. Rail surface defect detection based on an improved YOLOv5s. Applied Sciences 2023, 13, 7330. [Google Scholar] [CrossRef]
  35. Ferraris, C.; Amprimo, G.; Pettiti, G. Computer Vision and Image Processing in structural health monitoring: Overview of recent applications. Signals 2023, 4, 539–574. [Google Scholar] [CrossRef]
  36. Kralovec, C.; Schagerl, M. Review of structural health monitoring methods regarding a multi-sensor approach for damage assessment of metal and composite structures. Sensors (Basel) 2020, 20. [Google Scholar] [CrossRef]
  37. Bačić, B.; Rathee, M.; Pears, R. Automating inspection of Moveable Lane Barrier for Auckland Harbour Bridge traffic safety. In Proceedings of the Neural Information Processing, Cham, 2020//, 2020; pp. 150–161. [Google Scholar]
  38. Lee, H.; Oh, J. 3D displacement measurement of railway bridge according to cyclic loads of different types of railcars with sequential photogrammetry. Applied Sciences 2023, 13, 1359. [Google Scholar] [CrossRef]
  39. Bloemheuvel, S.; van den Hoogen, J.; Atzmueller, M. A computational framework for modeling complex sensor network data using graph signal processing and graph neural networks in structural health monitoring. Applied Network Science 2021, 6, 97. [Google Scholar] [CrossRef]
  40. Bhavanasi, G.; Neven, D.; Arteaga, M.; Ditzel, S.; Dehaeck, S.; Bey-Temsamani, A. Enhanced vision-based quality inspection: A multiview artificial intelligence framework for defect detection. Sensors 2025, 25, 1703. [Google Scholar] [CrossRef]
  41. Maass, W. Networks of spiking neurons: The third generation of neural network models. Neural Networks 1997, 10, 1659–1671. [Google Scholar] [CrossRef]
  42. Roy, K.; Jaiswal, A.; Panda, P. Towards spike-based machine intelligence with neuromorphic computing. Nature 2019, 575, 607–617. [Google Scholar] [CrossRef]
  43. Gerstner, W.; Kistler, W.M. Spiking neuron models: Single neurons, populations, plasticity; Cambridge university press, 2002. [Google Scholar]
  44. Pfeiffer, M.; Pfeil, T. Deep learning with spiking neurons: Opportunities and challenges. Frontiers in Neuroscience 2018, 12–2018. [Google Scholar] [CrossRef]
  45. Davies, M.; Srinivasa, N.; Lin, T.H.; Chinya, G.; Cao, Y.; Choday, S.H.; Dimou, G.; Joshi, P.; Imam, N.; Jain, S.; et al. Loihi: A neuromorphic manycore processor with on-chip Learning. IEEE Micro 2018, 38, 82–99. [Google Scholar] [CrossRef]
  46. Jiang, C.; Yang, L.; Zhang, Y. A spiking neural network with spike-timing-dependent plasticity for surface roughness analysis. IEEE Sensors Journal 2021, 22, 438–445. [Google Scholar] [CrossRef]
  47. Vemuru, K.V. Image edge detector with Gabor type filters using a Spiking Neural Network of biologically inspired neurons. Algorithms 2020, 13, 165. [Google Scholar] [CrossRef]
  48. Zhu, R.-J.; Wang, Z.; Gilpin, L.; Eshraghian, J. Autonomous driving with spiking neural networks. Advances in Neural Information Processing Systems 2024, 37, 136782–136804. [Google Scholar] [CrossRef]
  49. Duong, H.-T.; Le, V.-T.; Hoang, V.T. Deep learning-based anomaly detection in video surveillance: A survey. Sensors 2023, 23, 5024. [Google Scholar] [CrossRef] [PubMed]
  50. Iqbal, A.; Amin, R.; Alsubaei, F.S.; Alzahrani, A. Anomaly detection in multivariate time series data using deep ensemble models. PLoS One 2024, 19, e0303890. [Google Scholar] [CrossRef] [PubMed]
  51. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the Proceedings of the IEEE international conference on computer vision, 2015; pp. 4489–4497. [Google Scholar]
  52. Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. Slowfast networks for video recognition. In Proceedings of the Proceedings of the IEEE/CVF international conference on computer vision, 2019; pp. 6202–6211. [Google Scholar]
  53. Fan, Y.; Mao, S.; Li, M.; Kang, J.; Li, B. LMFD: lightweight multi-feature descriptors for image stitching. Scientific Reports 2023, 13, 21162. [Google Scholar] [CrossRef]
  54. Wei, L.; Zhong, Z.; Lang, C.; Yi, Z. A survey on image and video stitching. Virtual Reality & Intelligent Hardware 2019, 1, 55–83. [Google Scholar] [CrossRef]
  55. Du, C.; Yuan, J.; Dong, J.; Li, L.; Chen, M.; Li, T. GPU based parallel optimization for real time panoramic video stitching. Pattern Recognition Letters 2020, 133, 62–69. [Google Scholar] [CrossRef]
  56. Lagorce, X.; Orchard, G.; Galluppi, F.; Shi, B.E.; Benosman, R.B. HOTS: A hierarchy of event-based time-surfaces for pattern recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 2017, 39, 1346–1359. [Google Scholar] [CrossRef]
  57. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. International Journal of Computer Vision 2004, 60, 91–110. [Google Scholar] [CrossRef]
  58. Zuiderveld, K.J. Contrast limited adaptive histogram equalization. Graphics gems 1994, 4, 474–485. [Google Scholar]
  59. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, present, and future of simultaneous localization and mapping: Toward the robust-perception age. IEEE Transactions on Robotics 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  60. Turker, K.; Sharf, I.; Trentini, M. Step negotiation with wheel traction: a strategy for a wheel-legged robot. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, 14-18 May 2012, 2012; pp. 1168–1174. [Google Scholar]
Figure 1. (a) Gantry distribution and control zones for the movable concrete barrier (MCB) system along the Auckland Harbour Bridge corridor, showing active BTM-managed segments and gantry identifiers (Courtesy of Google Maps). (b) Schematic representation of the Barrier Transfer Machine (BTM) repositioning MCB units, with an inset showing metal pin linkage between adjacent barrier blocks essential for structural continuity (Source: Cottrell [14]).
Figure 1. (a) Gantry distribution and control zones for the movable concrete barrier (MCB) system along the Auckland Harbour Bridge corridor, showing active BTM-managed segments and gantry identifiers (Courtesy of Google Maps). (b) Schematic representation of the Barrier Transfer Machine (BTM) repositioning MCB units, with an inset showing metal pin linkage between adjacent barrier blocks essential for structural continuity (Source: Cottrell [14]).
Preprints 187765 g001
Figure 2. Temporal windowing improves label consistency across frame sequences. While frame-level predictions misclassify blurred or occluded frames as ‘Ambiguous,’ the surrounding temporal context enables correction to ‘Safe,’ demonstrating improved stability for anomaly detection.
Figure 2. Temporal windowing improves label consistency across frame sequences. While frame-level predictions misclassify blurred or occluded frames as ‘Ambiguous,’ the surrounding temporal context enables correction to ‘Safe,’ demonstrating improved stability for anomaly detection.
Preprints 187765 g002
Figure 3. SIFT preprocessing comparison for Pin_OK vs. Pin_OUT states: (1) Grayscale ROI, (2) intensity histogram, (3) CLAHE-enhanced ROI, (4) all SIFT keypoints, and (5) top 100 keypoints. The difference in keypoint density and distribution reveals distinct structural texture variations between the classes, supporting discriminative feature learning.
Figure 3. SIFT preprocessing comparison for Pin_OK vs. Pin_OUT states: (1) Grayscale ROI, (2) intensity histogram, (3) CLAHE-enhanced ROI, (4) all SIFT keypoints, and (5) top 100 keypoints. The difference in keypoint density and distribution reveals distinct structural texture variations between the classes, supporting discriminative feature learning.
Preprints 187765 g003
Figure 4. Panorama stitching failure analysis across three stages: (a) SIFT + homography-based stitching results in registration failure and partial image dropout, highlighting the limitations of monocular geometric stitching under high-motion field conditions. (b) SIFT keypoint matches between Frame 1 and Frame 2, with 685 high-confidence correspondences indicating potential alignment feasibility; (c) translation-based side-by-side stitching reveals misalignment and parallax artifacts despite visual continuity.
Figure 4. Panorama stitching failure analysis across three stages: (a) SIFT + homography-based stitching results in registration failure and partial image dropout, highlighting the limitations of monocular geometric stitching under high-motion field conditions. (b) SIFT keypoint matches between Frame 1 and Frame 2, with 685 high-confidence correspondences indicating potential alignment feasibility; (c) translation-based side-by-side stitching reveals misalignment and parallax artifacts despite visual continuity.
Preprints 187765 g004
Figure 5. Visualisation of class-wise spike encoding patterns for the four annotated pin alignment categories (‘OK’, ‘Partial Misalignment’, ‘Out’ (as Full Misalignment), and ‘Ambiguous’). Each row presents a representative input frame (left), the corresponding spike raster plot across 100 neurons over 10 discrete timesteps (middle), and the aggregated temporal spike rate plot (right). The TTFS (Time-to-First-Spike) encoding captures both the spatial salience and timing of activations derived from SIFT descriptors. Notably, the ‘OK’ class exhibits early, sparse, and temporally clustered firing. The ‘Partial’ and ‘Out’ categories show increasingly dense and prolonged spike activity, corresponding to structural irregularities. In contrast, ‘Ambiguous’ class frames result in disordered or suppressed spike patterns due to occlusions or motion blur. This visualisation highlights the discriminative potential and biological plausibility of the proposed spike encoding scheme.
Figure 5. Visualisation of class-wise spike encoding patterns for the four annotated pin alignment categories (‘OK’, ‘Partial Misalignment’, ‘Out’ (as Full Misalignment), and ‘Ambiguous’). Each row presents a representative input frame (left), the corresponding spike raster plot across 100 neurons over 10 discrete timesteps (middle), and the aggregated temporal spike rate plot (right). The TTFS (Time-to-First-Spike) encoding captures both the spatial salience and timing of activations derived from SIFT descriptors. Notably, the ‘OK’ class exhibits early, sparse, and temporally clustered firing. The ‘Partial’ and ‘Out’ categories show increasingly dense and prolonged spike activity, corresponding to structural irregularities. In contrast, ‘Ambiguous’ class frames result in disordered or suppressed spike patterns due to occlusions or motion blur. This visualisation highlights the discriminative potential and biological plausibility of the proposed spike encoding scheme.
Preprints 187765 g005
Figure 6. The Architecture of the proposed two-layer Leaky Integrate-and-Fire (LIF) Spiking Neural Network is used for anomaly classification.
Figure 6. The Architecture of the proposed two-layer Leaky Integrate-and-Fire (LIF) Spiking Neural Network is used for anomaly classification.
Preprints 187765 g006
Figure 8. End-to-end architecture of the proposed SIFT-SNN system, comprising preprocessing, spike encoding, SNN training, and evaluation feedback loops. Distinct stages are colour-coded: preprocessing (blue/purple), encoding (orange), training (yellow), and evaluation (green/red).
Figure 8. End-to-end architecture of the proposed SIFT-SNN system, comprising preprocessing, spike encoding, SNN training, and evaluation feedback loops. Distinct stages are colour-coded: preprocessing (blue/purple), encoding (orange), training (yellow), and evaluation (green/red).
Preprints 187765 g008
Table 1. Functional requirements for safe inspection of MCB systems on the Auckland Harbour Bridge versus the limitations of current manual and automated vision-based methods.
Table 1. Functional requirements for safe inspection of MCB systems on the Auckland Harbour Bridge versus the limitations of current manual and automated vision-based methods.
Requirement Description Limitations of Existing Methods
Detection of Mechanical Anomalies Identification of pin displacement (partial/full) Existing vision systems focus on surface defects, overlooking embedded mechanical elements
Spatiotemporal Awareness Integration of visual data across frames to capture transient anomalies Frame-based CNNs ignore temporal evolution and motion context
Real-time Monitoring Continuous surveillance during traffic flow. Manual inspection only during off-peak hours; lacks temporal continuity.
Deployability on Edge Devices Feasible latency, power, and compute profile for embedded or roadside units. CNNs/DL models are often too heavy, require GPUs or cloud connectivity.
Classification Granularity Ability to differentiate degrees of failure (safe, partial, full, ambiguous). Most existing models output binary ‘defect/no-defect’ results.
Visual Context Preservation (e.g., stitching) The capability to aggregate adjacent frames into a coherent spatial view during motion. Panorama stitching is unreliable under motion blur, change in viewing perspective, parallax, and structural occlusion.
Table 2. Comparative summary of key capabilities in related work vs. proposed system.
Table 2. Comparative summary of key capabilities in related work vs. proposed system.
Capability Traditional CV Methods CNN-Based Models 3D CNN / Optical Flow Prior SNN Approaches Proposed SIFT-SNN
Surface-Level Defect Detection ✓✓ ✓✓✓ ✓✓✓
Structural/Modular Component Analysis ✓✓✓
Temporal Modelling ✗ / ✓ (RNN variants) ✓✓✓ ✗ / ✓ ✓✓✓
Multi-Class Anomaly Interpretation ✓✓✓
Real-Time Edge Deployability ✓✓ ✓✓✓
Interpretable Feature Extraction ✓✓✓ (SIFT + SNN fusion)
Legend: ✓ = Basic Support, ✓✓ = Moderate Capability, ✓✓✓ = Advanced/Full Support, ✗ = Not Supported).
Table 3. Summary of data acquisition sessions, detailing the equipment used, camera configurations, environmental conditions, operational challenges, and progressive integration of advanced technologies across three field deployments.
Table 3. Summary of data acquisition sessions, detailing the equipment used, camera configurations, environmental conditions, operational challenges, and progressive integration of advanced technologies across three field deployments.
Session No. Equipment Camera Specs Environmental Conditions Challenges Tech. Integration
1st Deployment GoPro 5, GoPro 8, Samsung A7, iPad 6 720p @ 240 fps, narrow FOV, auto shutter/ISO, flat colour profile Overcast with light rain Mounting instability, device overheating Baseline setup: groundwork for multi-angle capture
2nd Deployment GoPro 5 & 8 (mounted front and rear on BTM) Same specs as above Sunny, clear visibility Battery drainage, thermal management Consistent multi-perspective pin tracking
3rd Deployment GoPro 9, iPhone 13 Pro with LiDAR + GPS GoPro 9: 1080p @ 240 fps; iPhone: LiDAR & 3D capture Dark, rainy morning; low light Waterproofing, power sustainment, and GPS stabilisation Integrated 3D spatial capture, GPS tagging, 5G for remote sync
Table 4. Comparison of classification performance and real-time deployment metrics across GPU, CPU, and simulated Jetson Nano configurations. Jetson Nano results are based on FP16 TensorRT inference emulated on a downclocked RTX 4060 to reflect embedded deployment conditions.
Table 4. Comparison of classification performance and real-time deployment metrics across GPU, CPU, and simulated Jetson Nano configurations. Jetson Nano results are based on FP16 TensorRT inference emulated on a downclocked RTX 4060 to reflect embedded deployment conditions.
Metric GPU (RTX 4060) CPU (Intel i7) Jetson Nano (Simulated) Notes
Accuracy (%) 92.3 92.3 ~91.1 5-fold cross-validation (macro avg.)
F1 Score (Macro Avg.) 91.0 91.0 ~89.6 Minor performance drop due to low-power constraints
Precision / Recall High on Safe/Partial Lower on Ambiguous Similar pattern Temporal smoothing improved recall for Full & Ambiguous classes
Inference Latency (ms/frame) 9.5 26.1 ~48.3 FP16 mode; simulated via TensorRT and downclocked RTX 4060
Model Size (MB) 2.9 2.9 2.9 Suitable for embedded platforms
Memory Usage (MB) ~180 ~210 ~290 Based on peak usage during batch inference (batch size = 32)
Power Consumption ~45W ~65W ~5–10W Jetson TDP inferred from NVIDIA documentation (low-power mode)
Note: Jetson Nano figures were obtained through controlled FP16 inference using TensorRT on a desktop GPU with matched clock/power limits. Results align with published benchmarks for MobileNet and SNN-class models.
Table 5. Cross-validation performance (Accuracy, Precision, Recall, F1-score) for all four alignment classes.
Table 5. Cross-validation performance (Accuracy, Precision, Recall, F1-score) for all four alignment classes.
Class Accuracy (%) Precision (%) Recall (%) F1-Score (%)
Safe 94.1 93.5 94.8 94.1
Partial 90.2 88.7 89.9 89.3
Full Misalignment 92.6 91.1 90.5 90.8
Ambiguous 87.4 84.6 86.1 85.3
Macro Average 91.1 89.5 90.3 89.9
Table 6. Deployment benchmarks: Inference time, memory usage, and model size across RTX 4060 GPU, Intel i7 CPU, and Jetson Nano simulation.
Table 6. Deployment benchmarks: Inference time, memory usage, and model size across RTX 4060 GPU, Intel i7 CPU, and Jetson Nano simulation.
Metric RTX 4060 GPU Intel i7 CPU Jetson Nano (Simulated)
Average Inference Time 9.5 ms/frame 17.3 ms/frame 28.4 ms/frame
Memory Usage (Runtime) 1.1 GB 1.4 GB 840 MB
Model Size (on disk) 8.3 MB 8.3 MB 8.3 MB
Estimated Power Usage ~45 W ~25 W ~5–7 W
Table 7. Temporal fusion comparison: F1 scores across pin alignment classes for different aggregation strategies. Sequence-aware models, particularly GRU-based readout, outperform frame-wise classification by improving consistency on ambiguous and partially misaligned cases, yielding a macro F1 gain of up to 2.6%.
Table 7. Temporal fusion comparison: F1 scores across pin alignment classes for different aggregation strategies. Sequence-aware models, particularly GRU-based readout, outperform frame-wise classification by improving consistency on ambiguous and partially misaligned cases, yielding a macro F1 gain of up to 2.6%.
Method ‘Safe’ (%) ‘Partial’ (%) ‘Out’ (%) ‘Ambiguous’ (%) Macro F1 (%)
Frame-wise Only 94.1 89.3 90.8 85.3 89.9
Majority Voting 95.2 90.5 91.4 87.2 91.1
EMA Smoothing 95.6 91.2 91.8 88.1 91.7
GRU Readout (Test) 96.1 91.7 92.4 89.6 92.5
Table 8. Comparison of different spike encoding strategies evaluated on classification F1 score, average inference latency, and spike density. Latency coding (TTFS) achieved the best balance of accuracy and computational efficiency, while rate coding incurred higher spike activity. The analogue input bypassed spiking altogether but underperformed compared to both spike-based variants.
Table 8. Comparison of different spike encoding strategies evaluated on classification F1 score, average inference latency, and spike density. Latency coding (TTFS) achieved the best balance of accuracy and computational efficiency, while rate coding incurred higher spike activity. The analogue input bypassed spiking altogether but underperformed compared to both spike-based variants.
Encoding Scheme F1 Score (%) Latency (ms) Spike Density (%)
Latency (TTFS) 91.1 9.5 8.1
Rate Coding 89.3 11.8 21.6
Analogue Input (No Spike) 88.7 10.2 N/A
Table 9. Comparative evaluation of the proposed SIFT-SNN framework against traditional CNNs and modern transformer-based models across key criteria: macro F1 score, model size, inference latency, edge deployment suitability, and interpretability. The SIFT-SNN offers a favourable balance between performance and deployability for real-time infrastructure monitoring applications.
Table 9. Comparative evaluation of the proposed SIFT-SNN framework against traditional CNNs and modern transformer-based models across key criteria: macro F1 score, model size, inference latency, edge deployment suitability, and interpretability. The SIFT-SNN offers a favourable balance between performance and deployability for real-time infrastructure monitoring applications.
Model Macro F1 Score (%) Model Size (MB) Inference Latency (ms/frame) Power Suitability (Edge) Interpretability
SIFT-SNN 91.1 8.3 9.5 Yes High
ResNet-50 91.4 98 34.0 No Low
MobileNetV2 89.7 14 22.0 Limited Low
EfficientNet-B0 90.2 20 28.0 Limited Low
ConvNeXt-T 91.6 29 36.5 No Low
ViT-B/16 92.0 85 41.0 No Very Low
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated