Preprint
Article

This version is not peer-reviewed.

A Survey of Multimodal Data Fusion in Earth Observation-Remote Sensing

  † These authors contributed equally to this work.

Submitted:

09 October 2025

Posted:

10 October 2025

You are already at the latest version

Abstract
Multimodal data fusion has emerged as a pivotal technique in Earth Observation–Remote Sensing (EO-RS) analysis as it enables the integration of heterogeneous data to enhance geo-analysis of the living environment. This survey comprehensively explores multimodal data fusion in EO-RS, highlighting a fundamental and already existing classification framework for fusion techniques. This survey presents a unique classification approach to fusion methods based on their underlying analytical paradigm, which fundamentally distinguishes between emerging and established techniques. Special consideration is given to pre-processing strategies critical for preserving modal integrity and ensuring fusion fidelity. Through detailed case studies, we demonstrate how the fusion of disparate modalities such as optical, radar, LiDAR, and hyperspectral data improves model accuracy and scene characterization. We also discuss the challenges inherent in multimodal fusion, including spectral and spatial distortions, computational demands, and data validation constraints. This work provides an overview of the current landscape of EO-RS data fusion and serves as a guide to future research on more robust, domain- and context-aware, and scalable fusion frameworks.
Keywords: 
;  ;  ;  ;  

1. Introduction

Definitions of data fusion have long varied across application domains, data type, and even among individual scientists [1,2], challenging the possibility of a single universally accepted definition. The first attempt at a formal definition of data fusion was in 1987 by the Data Fusion Sub-panel of JDL, which defined it with a military undertone, yet from a comprehensive viewpoint as
a process dealing with the association, correlation, and combination of data and information from single and multiple sources to achieve refined position and identity estimates, and complete and timely assessments of situations and threats, and their significance. The process is characterized by continuous refinements of its estimates and assessments, and the evaluation of the need for additional sources, or modification of the process itself, to achieve improved result [3].
This definition has over the years evolved into concision as "the process of combining data to estimate entity states" [4] and while it has been influential in certain domains, various fields have developed their own definitions and frameworks, which may or may not directly build upon the JDL definition. As noted in [5] from the field of classical Metrology, data fusion is defined as the combination of different information sources for new knowledge about physical quantities, events, and situations to achieve higher precision. [6] and [7] restricted the definition to the fusion of sensory data, while [8] and [9] presented it from a generic and holistic perspective as multidisciplinary, and from multiple sensors, information sources, and decision-makers. Although these definitions may be attuned to their respective fields of application, the common underlying tenet is the unification of complementary information from multiple sources into an improved single-view perspective. Similarly, from the Special Interest Group of EARSeL in 1996 emerged a formal definition of data fusion in the domain of EO-RS [10]. Data fusion in this regard was adopted as "a formal framework in which are expressed means and tools for the alliance of data originating from different sources" [11]. This aimed to establish a structured conduct of data fusion in Earth observation remote sensing and has become a widely accepted definition among researchers and practitioners. This review focuses on data fusion in EO-RS, set against a backdrop of increasing preference for remote sensing data in fields even outside the earth sciences for various applications.
The modality of a sensor is informed by internal (technical) specifications such as the resolutions (spatial, spectral, temporal, and radiometric) and imaging mechanism, as well as external factors like observation angle and acquisition time [12]. Typical earth observation remote sensing modalities include Panchromatic, Multi-spectral, Hyper-spectral, Radar, LiDAR, thermal, and microwave imaging. Each of these modalities is designed and tailored for a specific range of the EM spectrum, view angle, and application(s); hence, none can singularly provide a true and accurate approximation of an observed scene [13]. Multimodal data fusion mitigates this constraint by combining the near-precision of various modalities in a complementary fashion for improved data representation.
The early 2010s marked the beginning of a dramatic increase in earth observation satellites, mainly due to improved technology, reduced launch cost, and the increasing demand for high-resolution and real-time data for various fields of application; which includes agriculture, urban planning, environmental monitoring, and disaster management [14,15]. This, in turn, translates to a vast repository of disparate datasets that require data fusion to fully exploit and leverage the complementary aspects of their respective modalities [16,17]. Consequently, several multimodal data fusion techniques have emerged over the past years and have in recent times, gained significant traction, particularly at the turn of the third decade of the 21st century. This is largely driven by the increasing influx of EO data and improved computational power.
To put these fusion techniques into perspective, several taxonomies have been developed since the 1980s. [6] classified fusion techniques into four groups based on the level at which the fusion is performed; they are signal-level, pixel-level, feature-level, and decision-level fusions. [18] incorporated the transforms of disparate spatiotemporal resolutions onto a common grid and developed a three-level taxonomy, which includes sub-feature level, feature level, and decision level fusion techniques. A most recent taxonomy was proposed by [19], which classifies existing techniques based on the underlying mathematical principle. They inferred that data fusion algorithms are fundamentally designed with either of the following: a governing equation that describes the calculation process and geometric features of a surface (Gaussian Process algorithms), a design to reduce noise by assigning weights to outputs (Weighted Least Square Algorithms), or a design to learn patterns and relationships in datasets (Machine learning Algorithms). Some notable fusion techniques include Principal Component Analysis, sparce encoding, Hue Intensity Saturation, gradient descent, Bayesian inference, Wavelet transform, Probabilistic modeling, Simple Block Average, Laplacian Pyramid Decomposition, Discrete Cosine Transform, and Machine Learning (which has significantly contributed to the surge in the development of multimodal fusion methods today) [20,21]. The variations and growing number of classifications can be attributed to the fact that even within individual domains such as EO-RS, there isn’t a universal theoretical basis/framework upon which a comprehensive classification scheme can be developed for multimodal data fusion techniques [2,22].
While existing taxonomies offer valuable perspectives, none consider the underlying analytical paradigm driving fusion methods. All methods, regardless of what defines them or what taxonomy they constitute, are fundamentally model- or data-driven. In addition to an overview of the state of fusion methods, we present a classification based on Machine learning and Traditional methods of fusion in EO-RS, which should provide an understanding and distinction between established techniques and emerging approaches. Traditional methods of data fusion are based on predefined mathematical or statistical models underlined by fixed assumptions [20] as opposed to machine learning models that learn and adapt to their respective input data. We then examine several case studies that illustrate the practical applications of these approaches. But first, we highlight the main groups of pre-processing techniques employed as precursors to data fusion. Although not within the scope of this review, it is important to highlight the major prior operations that are equally fundamental to the fidelity of multimodal data fusion. Regardless of how sophisticated a fusion algorithm is, an inaccurate and inadequate pre-processing (including failure to preserve modal integrity) can significantly compromise results and render it ineffective. This paper is organized as follows: Section 1 introduces the research and provides a concise background on pre-processing techniques essential for fusion accuracy. Section 2 presents a classification based on the analytical paradigm, categorizing techniques into Machine Learning and Traditional approaches. Section 3 explores case studies demonstrating practical applications. Section 4 presents the challenges and limitations of data fusion in EO-RS. Section 5 presents the Discussion and outlook.

1.1. Pre-Processing Techniques

Although recent advancements in satellite and other remote sensing technologies have improved the spatial, temporal, and spectral resolutions of scene imaging, EO data still lack perfect continuity in space and time. This makes the alignment of the resultant disparate observations into a cohesive unit imperative and one that requires efficient strategies to create parity and extrapolate pseudo-continuous scenes, making the data appear more seamless and useful for analysis. Hence, the need and importance of pre-processing. Pre-processing refers to all activities of image rectification and restoration; thus, it aims to create a nearly true representation of an observed scene. It is so crucial that its accuracy is a major factor in determining the quality of the fused image [23]. To the best of our knowledge, very few or none of the publications on EO remote sensing data fusion have highlighted and acknowledged the pivotal role of prior or pre-processing techniques. We present a highlight of the most common groups of these techniques. Before these state-of-the-art data fusion algorithms can be effectively implemented, the tasks of projecting data onto a common grid or geographic space, assimilating observations of the same timeline, matching feature points between multiple images, and atmospheric corrections, among many others, are crucial. These techniques are generally grouped under geometric and radiometric corrections/pre-processing [24].
  • Geometric Pre-processing: This refers to all operations aimed at correcting spatial distortions in images and inconsistencies between images due to sensor specifications/mechanics, platform instability or movement, terrain relief, curvature, and rotation of the earth, etc. [25]. Examples include Radial and Tangential distortions, Skew, along-track scan error, scale error, etc. These distortions are unique to each imaging system’s self-geometric characteristics and, hence, require tailored models or approaches for correction. Even among images from the same sensor, a one-model correction solution is not always guaranteed, as sensor properties may change over time due to age, temperature, mechanical, and orbital stress. The sources of geometric distortions in remote sensing imagery can be broadly categorized into two: Observer-related distortions, which stem from the imaging system or sensor, and Observed-related distortions, which arise from the environment or objects being observed (e.g., atmosphere, Earth’s surface) [26]. Observer-related distortions, being predominantly systematic, are relatively easy to correct due to their consistency and predictability. In contrast, errors related to the observed scene can be more challenging to rectify, as they depend on factors such as time, position, and atmospheric conditions at the moment of image acquisition that may not always be readily available [27]. Geometric corrections have become increasingly critical in modern remote sensing due to the shift toward computerized data processing and the growing complexities of multimodal data fusion. Commonly used geometric correction methods include image registration, orthorectification, resampling, and georeferencing.
  • Radiometric Pre-processing: For a given pixel, the value recorded is a representation of energy emitted or reflected from both an observed target on the surface and its atmospheric interactions. The ideal imaging system should be able to measure the reflected energy from a scene accurately and uniformly. However, atmospheric perturbations affect the propagation of incident and reflected light, while factors such as platform instability and sensor properties (e.g., viewing geometry, noise, and response drift) further compromise the recorded digital numbers, introducing radiometric inconsistencies that degrade signal accuracy. Radiometric pre-processing aims to correct these errors associated with scene illumination through the process of sensor calibration, atmospheric correction, noise reduction[28], etc, for an improved estimation of an observation. Transforming the TOA spectral radiance to earth surface reflectance is the most crucial step in radiometric correction[29] as without this, signals would primarily reflect unintended observation of atmospheric conditions rather than actual surface characteristics.
For clarity and brevity, this paper uses abbreviations throughout the main text. A complete list of abbreviations and their corresponding full terms is provided in the Abbreviations section.

2. Taxonomy of Multimodal Data Fusion Techniques

2.1. Fusion Level

We highlight an underlying concept of all fusion approaches. Across all disciplines, fusion problems are a priori defined by the questions "when", "where", and "how" to fuse [18,30]. This concept is fundamental, as it forms the basis and building blocks for a wide range of techniques and taxonomies of multimodal data fusion in EO-RS. On this basis, data fusion techniques can generally be grouped under the following themes:
  • Signal-level fusion: This aims to create an image with a better signal-to-noise ratio than the source images. Fusion at this level combines the raw electrical signals measured from different sensors to create an enhanced representation of the observed scene.
  • Pixel-level fusion: Pixel-level fusion is fusion at the lowest level (pixels) where bits of measured quantities are merged[31]. This approach combines images by applying mathematical operations, such as averaging, to create an enhanced image that would otherwise not be obtained from the individual source images.
  • sub-feature level fusion: sub-feature involves fusing different spatial and temporal scales into a common multidimensional grid[18]. Fusion is at a lower level and involves integrating features within the extracted features, such as specific spectral bands or texture patterns from hyperspectral imagery.
  • feature-level fusion: This is relatively a high-level fusion technique as opposed to sub-feature level fusion. This involves combining features such as edges, textures, or shapes directly extracted from multiple images into a new and improved representation of all the source images [32].
  • decision-level fusion: The decision or results of independent processing algorithms on the disparate modalities are fused using specific techniques at a higher level to produce a final decision[33,34]. This approach aims to improve the accuracy of the final output via a thorough and robust decision-making process.
  • Hybrid-level fusion: Combines multiple fusion levels to produce a final output. This technique leverages the complementary advantage of each fusion level to improve the final output.

2.2. Machine Learning Methods

2.2.1. Deep Learning

Deep Learning, a subset of Machine Learning, is computation through multilayer neural networks and processing [35]. As concise as it is, deep learning is in fact a complex network of multilayered artificial neurons that simulate the complex decision-making process of the human brain. This complex architecture allows the capture of intricate and subtle patterns and relationships present in data. Hence, deep learning can provide finer-grained and detailed characterisation of EO imaging, and has garnered increasing attention in the field of EO-RS [36]. Accurate estimation of land use and land cover in an image scene is crucial for precision environmental monitoring and biodiversity conservation [37], agricultural management and urban planning [38], climate change studies [39], and disaster management [40]. [41] in a study in Indonesia developed ForestNet; a semantic segmentation-based Convolutional Neural Network that fuses the visible bands of Landsat 8 imagery and several variables of disparate modalities including NCEP-CFSv2 and USGS-SRTM data, to identify direct drivers of primary forest loss. A technique that can detect multiple clusters, land use, and land cover within the same region to ascertain drivers of the change. ForestNet, unlike canonical classification, takes a semantic segmentation approach, which conducts classification at the pixel level to identify intra-classes of land use or cover classes within a single image. [42] proposed a hybrid CNN which fuses a low resolution 3-meter PlanetScope imagery with a 3cm DJI Phantom 4 Pro UAV imagery to measure the areal extent of mangrove sites in Baja California. Using a modified efficientnet-b0 feature extractor (added a Perceptron), the network was able to combine the coarse optical image with the finer RGB UAV image, which was then fed into a fully connected layer for the task of classification (which could not be otherwise achieved with an efficientnet alone). Active cloud and aerosol profilers lack the wide swath their passive optical counterparts possess; hence, both are often combined for a detailed and yet extensive coverage of atmospheric profiles. A commonly fused data-pair is CALIOP and VIIRS, for which [43] developed an unsupervised DMSDA model to learn feature representations between the two. This approach works by mapping VIIRS into the feature space of the CALIOP cloud profile for a finer cloud characterization within the VIIRS image. This complements VIIRS with improved cloud profiling by providing detailed vertical cloud structure information. [44] developed a fully convolutional Siamese network for multimodal image registration, a generalized approach at the time, where numerous studies focused on specialized networks for registration of unimodal images. Each branch of the network has seven convolutional layers with filters ranging from 32 to 128. Specific to this design, batch normalization is applied only to the seventh layer. This is to preserve information encoded in negative values.

2.2.2. Classical Machine Learning

The backbone of today’s revolutionary sub-field of deep learning is Classical machine learning. It has laid the foundations through algorithms such as decision trees, SVM, and KNN, among other data-driven methods for deep learning today. Yet, Classical Machine Learning continues to excel in EO-RS data modeling. For example, the Random Forest machine learning algorithm, in spite of recent advances in neural networks, continues to achieve great results in land use and land cover classifications [45,46]. For structured and non-complex data patterns and relationships, CML yields excellent results with high efficiency and reduced strain on computing resources, and it remains a preferred algorithm in this field today [47]. [48] developed a gradient-boosting machine learning model, FI-GBM, to estimate the full coverage spatio-temporal distribution of hourly ground-level nitrogen oxide ( N O 2 ) over Tangshan in China. By combining in-situ data from ground stations with the spatial advantage of TROPOMI and LCSs, the proposed FI-GBM algorithm augments AQM N O 2 by assimilating TROPOMI and LCS data N O 2 to infer hourly ground N O 2 across time and space. A novel approach to integrate machine learning algorithms for the fusion of several multimodal datasets was conducted by [49] for the retrieval of VWC. A feature-level fusion of satellite and reanalysis data by employing five ensemble decision tree algorithms, including Random Forest, Extreme and Light Gradient Boosting Machines, Gradient Boosted Decision Trees, and Bagging Tree algorithms. Although the integrated approach in itself is nothing new, its application for VWC quantitative retrieval modeling is pioneering. [50,51] employed a random forest fusion machine learning strategy to downscale 25km course resolution AMSR-E soil moisture data using 1km MODIS data products (LST, NDVI, EVI, Albedo, LAI, and ET). The latter further extended the resulting AMSR-E product to create a high-resolution soil moisture drought index over Tangshan in China.
The inherent strengths and limitations of individual machine learning algorithms are some of the key drivers for data fusion in EO-RS. Integrating multiple modeling algorithms, particularly when dealing with highly complex datasets, creates a more robust workflow and significantly improves the accuracy of the results [52]. [53] proposed a decision-level machine learning fusion approach to extract urban buildings from Landsat 5, THEOS, and WorldView-3 satellite images. After the respective classifications of buildings using a SVM, KNN, and an ANN on computed spectral indices, the Majority Voting ensemble method was used to find the intersections of propositions from the various classifiers for the final output. Although decision-level fusion offers promising accuracy and robustness, it can be computationally intensive and time-consuming, especially for complex tasks. Therefore, it may not be the most efficient choice for simpler or less data-intensive applications.

2.3. Traditional Methods

2.3.1. Geostatistical Methods

Originally developed for mining problems, Geostatistics provides a set of techniques for estimating a given quantity by exploiting the spatial dependence between two or more statistically heterogeneous datasets. It leverages the existing spatial dependence between a scene (sensor data) and its true representation to estimate the relationship between. Some of the most commonly used methods include kriging-based methods, Bayesian methods, and geostatistical simulation methods. These methods provide the advantage of better spatial autocorrelation, computational efficiency, and better uncertainty quantification, among many others. [54] extended the Fixed-Rank Krieging geostatistical modeling technique for the fusion of AIRS and OCO-2 to estimate carbon dioxide in the PBL. In spite of using synthetic OCO data (due to OCO-2 launch failure coupled with insufficient ground truth data), the estimates returned were accurate representations of the global distribution of PBL C 0 2 except for a few places where synthetic data was sparse. Similarly, [55] combined AERONET, MISR and MODIS datasets for improved estimates of AOT using universal Kriging (which can take multisensor and spatial features as predictors). This was a novel study at the time because the majority of prior studies in this regard only fused passive sensor data, which are characteristically coarse. The estimated AOTs were significant improvements from their original versions, especially in areas with high correlation between the input data. A major challenge with traditional kriging in remote sensing is the intensive computation of inverting large datasets. With a time complexity of O ( N 3 ) , applying this technique to massive matrices is mostly infeasible. However, reducing the dimensions of such matrices promises a way around this conundrum. Given this, [56] proposed an approach of using SSDF that employs spatial random effects modeling (a variant of kriging for dimensionality reduction) to fuse data from AIRS and CrIMSS in the creation of enhanced near-surface air temperature data over the conterminous United States. This generated a refined dataset that exhibited consistent accuracy and errors with NOAA’s Integrated Surface Database.
Moving away from traditional kriging with variograms, [57] employed MPS to integrate ground points with coarse raster elevation data (SRTM and GMTED2010), producing higher a spatial resolution output. Their approach built upon the FILTERSIM MPS algorithm by [58], with modifications to better account for spatial structures in digital elevation model fusion. The modified MPS method demonstrated superior performance compared to traditional and MPS-based kriging across all tested metrics (ME, SDE, RMSE, MAE, and SSIM).

2.3.2. Other Traditional Methods

Traditional methods, also known as Standard methods as indicated earlier, are techniques developed over the years that rely on statistical and mathematical models. Geostatistics, although a traditional method in this sense, is an umbrella term encompassing various methods that focus on modeling the spatial relationships between variables to estimate a quantity. In this context, we refer to other traditional methods as simply non-geostatistical methods of EO-RS data fusion, yet underlined with not-so-flexible mathematical assumptions of a phenomenon or the relationship between quantities. In a passive-active sensor fusion, [59] employed a Wavelet-PCA fusion approach in a comparative study of the performance of pixel and object-based image classification over the alluvial Sanjing Plain in China. Prior performances of the wavelet-PCA fusion method [60] inspired its adoption for the integration of multispectral GF-1 and ALOS-PALSAR imagery for land cover classification. In a similar study, [61] evaluated the accuracy of the Random Forest and Support Vector Machine algorithms for land cover classification by using DWT to fuse HJ1B and ALOS-PALSAR coregistered images over arid and semi-arid landscapes. Although the focus of this work was on the performance of the classification algorithms, the wavelet transform fusion considerably mitigates information loss caused by modality-specific variations in the source images. Wavelet transform continues to demonstrate superior performance in remote sensing image analysis, particularly for classification tasks, by reducing both spectral degradations and spatial distortions in fused images across multiple scales [62,63]. Primarily, estimating AOD via the fusion of disparate remote sensing sources has been with passive-passive sensor fusion. This is due, among other things, to their widespread availability and large coverage, allowing for broader regional aerosol studies. However, this advantage is mired with the presence of cloud, haze, and illumination [64,65], which significantly reduces the spatial completeness of the derived or estimated AODs. In light of this shortcoming, [66] in retrieving AODs over mainland China combined retrievals from MODIS and CALIOP using BME. By combining the principles of maximum entropy and Bayesian inference, BME leverages the use of prior information such as physical laws, empirical models, and expert opinions in fusing the complementary attributes of both inputs for an improved result. The fine-grained CALIOP retrievals combined with MODIS’s wide scanning swath provided detailed AOD characterization and enhanced spatial continuity across large areas, demonstrating clear advantages over passive-sensor-only approaches that lack CALIOP’s active sensing capabilities. Intensity Hue Saturation (IHS) is a traditional fusion technique that is easy to implement but prone to spectral distortion. GTF is an approach proposed by [67] for fusing infrared and gray visible images. GTF avoids designing rules for pixel-level fusion while retaining their radiation and spatial information. To address these shortfalls and exploit the full advantage of both techniques, [68] developed the IHS-GTF hybrid approach for fusing TerraSAR and Sentinel-2A optical data in urban impervious surface extraction over Wuhan, China. The IHS-GTF method significantly outperformed individual techniques (IHS, GTF, and DWT) with R-squared values averaging 0.45 points higher.
A Comprehensive list of these fusion techniques detailing their fusion paradigms, fusion levels and modalites involved can be found in

3. Case Studies

In this Section, we explore the application of multimodal data fusion in EO-RS analysis for effective model performance and improved results. Areas to explore include Aerosol and Cloud detection, land cover change, and Shadow Detection and Image Restoration.

3.1. Aerosol and Cloud Detection

Optical sensors are characteristically designed with a wide scanning swath, making them suited for regional and near-global studies. However, they lack the quality of detailed vertical and horizontal differentiation of clouds and aerosols within an atmospheric column. Radar and LiDAR sensors, on the other hand, provide fine and detailed atmospheric columnar characterizations but lack a wide horizontal scanning range. This trade-off presents an opportunity to harmonize the complementary advantages of both sensors for detailed cloud and aerosol profiles on a wide scale. We present two use cases where multimodal data fusion significantly improves dust-aerosol and cloud machine learning detection algorithms.
Mineral dust aerosols play crucial roles in the bio-geochemical cycle of the Earth. They significantly influence cloud formation, radiative energy transfer, and nutrient supply. In addition, they are critical indicators of air quality and have important implications for human health. Earth observation remote sensing provides a means for monitoring the distribution and transport trajectory of dust aerosols on both local and global scales. Therefore, an efficient and effective dust detection algorithm becomes paramount for reliable retrievals. [69] developed a global dust detection algorithm; an FFNN with three hidden layers to detect dust pixels in cloud-free conditons from daytime VIIRS image. Given the limited scope of many prior studies which often focused on specific regions, restricted dust events, specific viewing angles, and use physical-based methods (eg. threshold and spectral signature-based methods) for model training and evaluation, the main objective of this study is to develop a global detection ML-based algorithm that is independent of study area or dust event. Data for the study is NASA’s SIP global collocated daytime VIIRS and CALIOP product. Although MODIS-CALIOP collocation is very common in this domain, the study opted for VIIRS for two reasons. First, is the extended operational lifetime of VIIRS on JPSS, and second, is the number of diverging VIIRS pixels off CALIOP nadir track. The collocated product from SIPS initially included only VIIRS-on-CALIOP (VIIRS-CALIOP intersections) track pixels. To enhance the robustness of the detection algorithm, four VIIRS pixels adjacent to the CALIOP track were added to the collocated product for improved generalization. To evaluate the algorithm’s performance, two dust aerosol products were employed. The first was the ADP generated using NOAA’s physics-based VIIRS aerosol detection model [70], which identifies dust based on spectral properties in the ultraviolet, blue, and thermal VIIRS bands. The second was the CALIOP aerosol product, which utilizes the V3 algorithm to distinguish between pure dust and polluted dust. Over the ocean, the low performance of FFNN can be attributed to the fact that the model finds it difficult to separate dust from other aerosols. However, it outperforms both ADP and CALIOP by a large margin over land where dust events originate and are predominant. The combined approach leverages the vertical discernment property of CALIOP and the extensive scanning swath of VIIRS for wide area retrievals. Table 1, Table 2 and Table 3 show the detection rate of each approach under clear skies (clouds masked using ADP cloud Mask) over land and the ocean.
The radiative forcings of clouds create feedback uncertainty, which makes it challenging to model cloud properties and actions for accurate climate modeling and projection. In a time where climate change threatens irreversible consequences [71], it is imperative to develop a more comprehensive and robust understanding of cloud dynamics and distribution, as both significantly affect the distribution of heat in the atmosphere [72,73,74]. Cloud profiles are critical for the accurate characterization of cloud albedo and radiative forcings [75], cloud-aerosol interactions [76], climate model validation and evaluation [77] and precipitation formation [78] emphasizing the integral role they play in Earth’s energy balance and climate system. [79] developed a deep CNN that estimates and maps cloud profiles from an active sensor unto a passive sensor product on the same grid. The model learns from labeled narrow nadir cloud pixels and extrapolates to off-nadir unlabeled cloud pixels on several kilometers of swath (approximately 2200km) for the estimation of full vertical cloud profile. Data for the study is the ATCS Dataset; which is a fusion (via alignment) of Polarization and Directionality of the Earth’s Reflectances data (POLDER level 1B) and 2B-CLDCLASS data from cloudSat’s CPR. Each POLDER pixel contains 16 co-registered multi-angle viewpoints from which the observation is made. The model returns a binary cloud/non-cloud binary output, regardless of the cloud profiles preserved in the ATCS, and was assessed using cloud estimations from 2B-CLDCLASS. The study allows for a series of ablation experiments focusing on sensor properties such as view angle, spectral band, and sensor polarimetry to evaluate their respective contributions in identifying cloud pixels. Cloud structure estimations improved with the ablation of two view angles while retaining all spectral bands. The ablation of at least a spectral band reduced model performance, indicating that all spectral bands informed the estimation of cloud structure. The approach achieved a dice score of 73.0% which by current segmentation standards may seem satisfactory at best. However, the absence of a combined human-algorithm annotation approach made accurate and precise pixel prediction difficult since the input image is multimodal, with each constituent having different sensitivities to cloud physics. For instance, the model struggled to identify optically thick clouds and lower layers of multi-layer clouds. Polarimetry had a negligible impact on model performance, responsible for a 0.4% reduction in dice score. This somewhat suggests the Stokes vector parameterization employed may not be ideal for describing the polarization state of reflected energy in this context. Nonetheless, the distribution of identified cloud pixels follows estimations in existing scientific literature and observed atmospheric patterns, with exact markedness of the vertical structure indicating cloud extinction from a height of approximately 14km upward. [80,81,82].

3.2. Land Cover Change

Land cover is the biophysical cover of the earth [83], and it is therefore a direct reflection of shifts in land use, environmental conditions, and human activities on spatio-temporal scales. These shifts, whether driven by natural processes or anthropogenic influences, have far-reaching consequences. Changes in land cover affect climate, the water cycle, biodiversity, urban heat, and food security, among other things. [84,85,86,87,88]. These highlight the importance of monitoring land cover as a key indicator of environmental change, and it reiterates the need for accurate land cover mapping for timely management and conservation efforts [37,89,90]. [36] sort to answer the question, can multimodal data fusion improve classification accuracy? More specifically, can the information gain from the fusion of disparate modalities be linked to improved classification accuracy? To ascertain this, a land cover type classification was conducted over Onondoga County in the central New York region using the Random Forest (RF) ensemble algorithm. Several Random Forest models were developed to assess the impact of different feature combinations on classification accuracy. Models were grouped under single-sensor, two-sensor, and three-sensor feature combinations. Data for the study is coregistered multi-temporal Landsat 5 Thematic Mapper (TM), ALOS-1/PALSAR, and LVIS LiDAR images. Except for the latter, optical and SAR images were sampled during leaf-on, leaf-off, and transition periods in between to evaluate the effect of spectral and scattering variations of some vegetation classes on the task. For each Landsat scene, surface reflectance (SR) for bands 1-5 and 7 was derived, from which the variance for each SR band and the NDVI using the near-infrared (band 5) and red bands (band 3) were computed. ALOS-1/PALSAR features extracted were categorized under intensity, interferometry, polarimetry, and texture, whereas relative height (rh) metrics for all flight lines and measured footprints were derived from LVIS Level 2 Geolocated Surface Elevation and Canopy Height Product (LGE). Table 4 shows the total features extracted from each sensor product.
All model groups showed varying levels of variability in classification accuracies. However, the most significant was observed among single-sensor feature models, with a standard deviation of 9.17, a mean of 61.1% and a statistical range of [74%, 48%], suggesting that this approach rarely makes consistent and reliable approximations of pixels to their true land cover classes. Classifications improved and became more consistent with multi-sensor feature models. Two and three-sensor feature combination models showed minimal variabilities of 4.9 and 3.6 standard deviations away from mean accuracies of 71% and 78%, respectively, indicating pixel assignments and discrimination of different land cover classes improved and remained consistent with the addition of information from multiple sensors. The best performance came from the fusion of LVIS and multitemporal Landsat and PALSAR features, which had an overall accuracy of 83%. It is also worth noting that Landsat-derived features consistently improved model accuracy, regardless of combining them with features from different modalities. This study pioneers the optical-SAR-LiDAR fusion approach in multimodal data fusion, and given the underpinning tenet of this practice, the established positive association between multimodal data fusion and model accuracy in this study does indeed suggest a relationship between the two. Table 5 shows the overall accuracy statistics for each model.

3.3. Shadow Detection and Restoration

Shadow is one of the most common errors in remote sensing data [91]. It occurs when there is an occlusion of direct incident light on a target object [92,93], thereby degrading the quality of imaging, which makes the identification and extraction of useful information very difficult, if not impossible. The financial cost of developing imaging systems, the economic losses incurred by degraded imagery, and the high potential for misinformation from remote sensing image analysis make the detection and restoration of faulty pixels a crucial and urgent task [94,95]. This effort is crucial for enhancing the precision and interpretability of images. The imaging of geologic outcrop formations is often affected by illumination constraints, such as shadows, due to their rugged edges and surfaces. Even HSI, which enables the detailed identification and mapping of minerals and rock types, is insufficient to fully compensate for the loss of information due to the presence of shadow pixels citep9874905. [96] combined SWIR-HSI (970-2500nm) with TLS data for the detection and restoration of shadowed pixels in vertical rock outcrop HSI imagery. HSI was fused with TLS point cloud data for shadow detection via geometric ray tracing. For restoration, TLC intensity information was directly integrated with HSI spectra and then assessed by matching regions in both sunlit and shadowed areas, which is unlike existing methods, which indirectly fuse intensity with the spectra of shadowed pixels. TLS point cloud data was acquired using a 1550nm laser wavelength Reiegl VZ-400 at three scan positions, each about 16m away from the rock face. The points were then coregistered, georeferenced to the WGS84 datum, and transformed to a local geodetic coordinate system in that order to enable shadow computations that utilize local solar zenith and azimuth angles. For occlusion analysis, a triangular mesh was created from the point data to identify locations of shadow pixels in the HSI and ascertain the visibility of each point to the HSI camera. TLS point intensity data were also converted to reflectance for the restoration effort. SWIR-HSI rock outcrop images were captured during morning (SWIR-shade), midday (SWIR-partial), and afternoon (SWIR-sun) hours to reflect different sunlit conditions of the day and as well as for the validation of the restoration technique on the SWIR-partial (which is the image for detection and restoration). All images were coregistered into the pixel space of the designated master image (SWIR-sun). The following corrections/pre-processing were conducted on the HSI image prior to fusion and subsequent analysis.
  • Dark current removal.
  • Removal of linear image artifact pixels.
  • Removal of strong water and carbon dioxide absorption wavelengths and extreme wavelengths in the 896-922nm and 2447-2504nm range of the camera spectrum.
  • Conversion of digital numbers (DN) to relative reflectance.
A camera model was developed to facilitate the fusion (registration) of HSI with TLS point-cloud data and intensity information via projection onto a common 2D image space or coordinate system. By first solving the interior (IO) and exterior orientation (EO) parameters of the camera, 3D point-cloud data points were projected from their respective object spaces (x,y,z) onto a 2D image space (x,y). All HSI images were re-projected onto image space following an initial projection into object space by tracing a ray from image pixel and determining its intersection with the outcrop mesh derived from the TLS point cloud. TLS intensity information on the other hand, was rasterized into 8-bit grayscale (active reflectance) images by projecting each point cloud point through the camera model using the IO and EO parameters of the master image. Shadow detection accuracy can be challenging due to the effect of different lightning conditions and their complex interactions with target materials, which sometimes create shadow composites and make edge and boundary detection more difficult [97]. In this study, instead of numerical accuracy metrics, a visual comparison was performed between true shadow pixels (well-defined with distinct boundaries) in the SwIR-shade and the ray tracing detections in the SWIR-partial. By this qualitative approach, an accurate pixel-level differentiation between shadowed and illuminated pixels was established. To achieve direct fusion of TLS information with HSI spectra for pixel restoration, a scale factor is applied to the entire HSI spectrum. This scale factor is calculated from the ratio of active sensor reflectance to HSI reflectance at the active sensor’s operating wavelength. To assess the accuracy of restored pixels, the spectral shape, spectral scale, and band correlation metrics were used to ascertain the preservation of the spectral profile, the effect of the restoration on the magnitude (intensity/brightness) of the signal across the spectra, and the numerical relationship between corresponding bands in the sunlit and shadowed matching material regions, respectively. The approach returned a band correlation of 0.6, about a 0.3 improvement from that of the unrestored SWIR-partial. The Spectral Scale measurements across pixels showed a nearly normal distribution centered at 1 (about 1.2 reduction from the unrestored SWIR-partial), with a variability of 27%. A spectral scale between 0.95 and 1.05 suggests matching materials across the spectrum with very minimal distortions. The statistics from both metrics validate the visual evaluation approach used for shadow detection. Spectral shape metrics for both the restored and original unrestored SWIR-partial images ranged between 0.97 and 1, an indication of quite a preservation of the spectra in both images; nonetheless, the application of the proposed direct method effected no change or improvement. A combination of both the direct and indirect spectra fusion approach did significantly improve the spectral shape of the image, suggesting that indirect shadow restoration, regardless of the computational constraint, does a better job at maintaining spectral profile across pixels. In relation to [98], the unimodal input data approach, coupled with the USRT model employed was able to improve the illumination of shadowed pixels while maintaining the spectral profile of the image. However, via careful visual inspection, it did not achieve the superior perceptual quality of the combined HSI-TLS data approach. This qualitative comparative assessment was chosen because of the differences in evaluation strategies between the two studies.

4. Challenges and Limitations of Data Fusion in Remote Sensing

Heterogeneity is the concept of distinctiveness and peculiarity between entities, which is informed by overt and/or covert qualities fundamental to their respective nature. It is, therefore, only natural, yet an inconvenience, to experience an initial friction or barrier in an attempt to harmonize such disparities. Remote sensing data fusion, although technically feasible and actively practiced, is not seamless by default and, hence, comes with certain challenges which can be attributed to but not limited to platform and sensor design, and externalities such as cost of geodata, limited validation data, computational infrastructure requirements, etc. Needless to say, these constraints can be broadly categorized into internal and External factors.

4.1. Internal Factors

4.1.1. Spectral Resolution

The spectral resolution of a sensor describes its property of detecting radiance from objects in specific wavelengths of the electromagnetic spectrum. It is the sampling rate and bandwidth at which a sensor collects information about a scene [99,100]. Monochromatic sensors detect radiance in a single wavelength band, while multispectral sensors capture data across several discrete spectral bands, typically ranging from 3 to 10 channels [101]. Hyperspectral sensors, as their name suggests, record scene information across hundreds of narrow, contiguous wavelength bands, enabling detailed material characterization and more comprehensive scene analysis through fine spectral resolution [102]. A major challenge in the fusion of mismatching spectral resolutions is spectral distortion [103], which describes alterations or deviations in the true spectral signature or profile of objects in the resulting image. Component Substitution methods (CS) such as IHS, PCA, and Brovey transform introduce spectral distortions when fusing imagery. These techniques operate on the premise that spatial details can be isolated and replaced between images [104]. However, spectral distortions arise from significant differences in spectral profile between the source images when sharp outlines from a single band image with high spatial resolution, e.g. panchromatic image, is injected into a multiband color image with low spatial differentiation and gradual color transitions between feature edges. Speckle noise is another spectral distortion in SAR-optical image fusion. This is caused by the random interference of many elementary reflectors within one resolution cell during the coherent processing of the backscatter signals in SAR imaging [105]. This speckle pattern or artifact can propagate into the fused SAR-optical fusion, bringing distortions into the spectral features of the optical component of the fused product. In spite of recent advances in fusion algorithms in the remote sensing domain, most correction methods like linear filtering, non-linear filtering, and partial differential equation filtering ignore the degradation induced by single bands images such as SAR [106], poorly correct degradations, or trade edge preservation for performance [107].

4.1.2. Spatial Resolution

The ground distance between two adjacent pixels, also known as Ground Sampling Interval, informs the pixel size or the spatial resolution of a terrestrial image [108]. Spatial resolution is therefore the scale or size of the smallest unit of an image or sensor capable of distinguishing objects [109]. One of the most common sources of spatial distortion in the fusion images with discordant spatial resolutions is pan-sharpening, a process that enhances the spatial details of an image using a panchromatic band (with a higher spatial differentiation of features). However, the effect of pan-sharpening is intertwined with the spatial and spectral resolution of the fused product, and as such, the discussion on spectral resolution above gives a glimmer of that. These spatial distortions can manifest in various ways, such as the block effect.
The Block effect is a condition that occurs when pixels of the same land cover class exhibit different reflectance values, making a naturally progressive and continuous phenomenon or class contrastive within by sharp discontinuities between pixels [110]. Block effect is a widely acknowledged problem in spatial unmixing-based methods in remote sensing, and it arises from the use of different local windows in the unmixing of different coarse pixels [111]. Spatial unmixing-based methods are employed when a high spatial resolution image is used to estimate what mix of surface features is likely present in each pixel of a spatially coarse image especially for applications in land cover classification, urban studies, and geological studies. Several studies have attempted to improve the performance of spatial unmixing methods [112,113,114] but have not achieved a significant resolution for blocky artifacts in images. This can be attributed to the following reasons: They are not specifically designed to account for that, they operate on inaccurate assumptions about the temporal dynamics of landcover change, and they ignore the compensation of model residuals [111,115]. The few that significantly minimize the occurrence of blocky artifacts do so at the expense of computational efficiency. Similarly, weight function-based methods, despite receiving considerable attention and extensive development in remote sensing literature [116,117,118,119], also remain susceptible to block artifacts. This vulnerability stems from the fact that they operate on pixel-level processing for the synthesis of fine and coarse-spatial resolution images for spatio-temporal analysis. They also fail to exploit the inherent and valuable shape information of features, which can help identify clusters and maintain the integrity of the natural shapes of features.

4.1.3. Temporal Resolution

Temporal resolution is a measure of the frequency of revisit of a satellite to a scene. More specifically, this is the revisit time of a sensor. The temporal constraint on high spatial resolution imaging, cloudy and shadowy conditions, scanning swath limitations and more specifically, the failure of the scan-line corrector of the enhanced thematic mapper+ on the legacy Landsat 7; suspended on January 19, 2024 [120,121,122], necessitate the fusion of images across different timelines, particularly for spatiotemporal analysis of the environment. Although platforms like Sentinel 2 and 3, and the development of CubeSats have significantly improved mitigating the conundrum of the high spatial and high temporal mutual exclusivity of images [123,124], multitemporal image fusion is still crucial for temporal gap-filling of historical satellite images [125,126,127]. This is, however, not without challenges, as multitemporal images are potentially affected by misregistration errors, varying climate and atmospheric conditions, spectral width and spectral response function, the BRDF effect from changing sensor observation and illumination angles, and phenological changes (which is a major challenge in ecological and urban change detection studies) [128,129]. To minimize these inconsistencies for a somewhat seamless integration, linear transformation models are employed to establish correspondence between these images [130,131,132,133,134], with which deficient and lost pixels can be approximated using derived coefficients [135].
A fundamental misconception underlining many methods of temporal and spatiotemporal fusion is the erroneous assumption of temporal and spatial invariability of surface features over time [114,126,129]. Consequently, these methods are challenged in heterogeneous landscapes and time series imagery with pronounced and abrupt changes in surface cover [128,136]. An example is the Fit-FC method, which blurs fine and sharp edges in highly multi-feature environments [137]. Hence, linear transformations cannot always accurately model the complex spatio-temporal dynamics of real-world scenarios. Non-linear and learning-based modeling approaches that are capable of modeling nuanced and complex relationships are therefore necessary to mitigate these constraints and produce more realistic and physically meaningful fusion results across heterogeneous and rapidly changing landscapes.

4.2. External Factors

4.2.1. Computational Infrastructure Requirements

According to the 2024 Earth Observation Satellite Systems report by Novaspace, EO satellites are expected to triple in number between 2024 and 2033, a 190% increase from the 1,864 in previous decades. This, coupled with the growing usage of UAV, ground instruments, and the continuous improvements in sensor resolutions, means a significant increase in the sheer size (file size and count) of data in an already substantial geospatial data repository [138]. Thus, EO-RS data is considered big data [139] and therefore requires very powerful and highly sophisticated hardware and algorithms for data processing and storage. This data barrage has necessitated the introduction of various processing and storage paradigms as well as the democratization of access to state-of-the-art computational resources [140] such as cloud-based computing (GEE, Google Collab, Microsoft Azure, AWS) and HPC. In spite of these and other existing solutions, the high dimensionality of geodata [141], the high bandwidth demands for data transfer [139,142,143], the intensive energy and data requirements along with specialized hardware demands of AI-driven algorithms, and the cost and geographical marginalization of existing infrastructure challenge the wide-spread adoption and effective utilization of these resources [144,145,146,147,148,149].

4.2.2. Cost of Geodata

The free and open data policy adopted by USGS and ESA for Landsat and Sentinel data products, respectively, has contributed significantly to the use of geospatial data in a wide range of environmental applications [150,151]. However, real-time and very high-precision environmental monitoring and analysis require very high-resolution (spatial, spectral, and temporal) data, which can be very expensive to acquire, especially in remote regions and the developing world for large-scale applications [152]. Although high spatial resolution aerial imagery such as that from the NAIP can be freely available, they are geographically restricted to the United States and severely lack the swath and spectral complements of space-borne imaging systems, which is crucial for large-scale detailed object-property characterization. Commercial data products such as those from PlanetScope and UAVs tend to have higher temporal and spatial resolutions [153], and are therefore often combined with open-access wide-swath products for high spatiotemporal analysis [154]. Furthermore, ground data is also crucial for model and data validation, data gap-filling, and scaling from point measurements to broader areas. However, the high cost of acquisition and maintenance, along with the challenge of achieving adequate spatial distribution of ground stations and equipment, present major barriers to accessing these niche datasets and their eventual integration with disparate modalities [155,156,157,158,159].

4.2.3. Limited Validation Data

Validation data in EO-RS takes several forms, including high-resolution raster, in situ ground measurements, model outputs, and reanalysis data. Validation, as defined by the CEOS WGCV, is the process of independently assessing the quality of derived data products [160], a crucial step in geoanalysis, as proper validation establishes rigor and confidence in remote sensing products and algorithms. Validation assesses the viability of multimodal data fusion and ensures reliable estimations of geophysical quantities through the quantification and propagation of uncertainties. Yet, this indispensable component is challenged by limited available data [161,162] despite mitigation efforts, especially for large-scale analysis. This is due to factors such as cost, terrain fragility and surface heterogeneity which limit data acquisition, untrustworthy reference dataset [163], and spatial scale mismatch between ground-based measurements and remote sensing products [164]. Some suggested and applied mitigation strategies include fusion of multiple validation data sources [165], establishing more in situ ground stations, improving ground stations to expand observation scale, improvements in upscaling transfer functions [163], and resampling [166]. Although these solutions offer potential pathways to address data scarcity, their successful implementation requires careful consideration of their respective inherent challenges. For instance, while resampling interpolates new pixel values, it assumes spatial continuity of features, which can potentially introduce errors that may affect accurate feature representation across heterogeneous landscapes.
Other challenges of multimodal data fusion include selection of appropriate algorithm [167,168], lack of standardized evaluation methods [169,170], misinterpretation of fused product [171], unfamiliarity with remote sensing data and methods especially among non-technical and untrained professionals, and on-demand coverage or real-time data [172].

5. Discussion

In this paper, we have undertaken a comprehensive exploration of data fusion: 1. Its ubiquity and how various fields and their respective applications have shaped its definitions 2). Fusion methods informed by the underlying analytical paradigms 3). Case Studies and 4). Challenges. We have also demonstrated how multimodal data fusion enables comprehensive object characterization by amplifying the information necessary for high-precision environmental analysis. As indicated earlier, improvements in sensor designs and specifications, coupled with the exponential growth in Earth observation data have catalyzed novel analytical techniques and driven significant improvements in existing infrastructure. These have widened the scope of data fusion beyond traditional modalities, allowing for more comprehensive and robust analysis with depth. However, these innovations, as groundbreaking and complementary as they are (especially with advances in fusion techniques, computational and storage infrastructure), either present new sets of challenges or compound existing ones, effectively rendering some solutions potentially inapplicable or highly marginalized to certain domains.
Besides computational inefficiency, many challenges in fusion algorithms stem from very little or no consideration of context and integration of domain knowledge in the learning or modeling process. Context-aware and domain-specific methods, although constrained to particular applications as their names suggest, prioritize precision within their respective sub-domains rather than making granular estimates across multiple fields. Understanding the nature, governing principles and physical laws, and the underlying dynamics within the focus of investigation considerably enhances modeling effectiveness and learning outcomes. Typical examples of contextual and domain-aware approaches are PIML and PINN, which maintain fidelity with physical laws by embedding them directly into data and the learning architecture. This provides a robust, efficient, and improved modeling accuracy in addressing various challenges across settings from mechanics to geophysics and even in the bio-chemical sciences [173,174]. And although working principles such as spatial continuity implemented through advanced fusion techniques like object-based unmixing methods and neural networks (e.g., CNNs) have improved capturing the spatial relationships between features, their accuracies are still constrained by feature heterogeneity and sharp discontinuities [175,176]. The emergence of vision transformers [177] with their self-attention mechanism effectively identify spatial nonstationarity between adjacent pixels in images, which is a limitation of CNNs in remote sensing image segmentation. This capability enables transformers to more effectively mitigate the mixed-pixel problem, leading to more accurate segmentation in heterogeneous landscapes.
While these technical advances in fusion algorithms have improved modeling accuracies, quantifying their comparative effectiveness remains a challenge due to the lack of a standard evaluation framework. Our analysis highlights this problem, particularly in the context of shadow detection and restoration, where metrics remain inconsistent across multiple studies. Shadow detection evaluation typically employs metrics such as precision, recall, F1-score, and balanced error rate, along with domain-specific measures like shadow detection rate and intersection over union. For shadow restoration, image quality metrics such as Peak Signal-to-Noise Ratio, Structural Similarity Index, and mean squared error, as well as spectral fidelity metrics of spectral shape and scale metrics for the restored regions are used. These variations make it more challenging to quantitatively compare the efficacy of fusion techniques and determine their appropriateness under specific conditions. Hence, the qualitative approach of visual inspection has assumed the de facto method when comparing accuracies between studies with different evaluation metrics. While it excels at perceptual evaluation, such as detecting unrealistic textures or contextually inappropriate spectral characteristics, visual inspection is highly subjective and influenced by context and observer perception. This evaluation inconsistency is so pervasive that it persists not only within shadow detection and restoration but also across multiple application domains, such as cloud detection. This fragmented evaluation landscape, as a result, not only hampers cross-study comparisons but also impedes the identification of optimal fusion strategies for specific use cases. This review strongly advocates for prioritizing future research in developing a unified benchmarking framework that establishes standardized evaluation protocols and metrics for assessing both the efficiency and efficacy of fusion algorithms across diverse application domains. Such a framework would enhance interdisciplinary research and collaboration while enabling robust comparative analysis.

Author Contributions

Conceptualization, supervision and funding acquisition, X.H.; writing—original draft preparation, review and editing, S.AN. All authors have read and agreed to the published version of the manuscript.

Funding

This research is part of the U.S. Geological Survey (USGS) National Land Imaging (NLI) Program under Grant No. G23AP00683. Additional support was provided by funded project supported by the School of Emerging Technologies (SET) seed funding, awarded by Towson University.

Acknowledgments

The authors gratefully acknowledge the researchers, particularly Dr. Pubu Ciren of NOAA/NESDIS and Prof. Maria Piles of ISP, University of Valencia, who took time from their busy schedule to provide helpful clarifications on their work cited in this manuscript.”

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this article..

Abbreviations

The following abbreviations are used in this manuscript:
JDL Joint Directors of Laboratories
EARSeL European Association of Remote Sensing Laboratories
E0-RS Earth Observation Remote Sensing
TOA Top of Atmosphere
NCEP CFSv2 National Centers for Environmental Prediction Climate Forecast System Version 2
RGB Red, Green Blue
UAV Unmmaned Aerial Vehicle
DMSDA Domain Multi-Sensor Domain Adaptation
LCS Low Cost Sensor
VWC Vegetation Water Content
ANN Artificial Neural Network
KNN K-Nearest Neighbor
SVM Support Vector Machine
OCO-2 Orbiting Carbon Observatory 2
AOD Aerosol Optical Depth
AOT Aerosol Optical Thickness
PBL Planetary Boundary Layer
AIRS Atmospheric Infrared Sounder
SSDF Statistical Data Fusion
CrIMSS Cross-Track Infrared Microwave Sounding Suite
NSAT Near-Surface Air Temperature
NOAA National Oceanic and Atmospheric Administration
MPS Multiple Point Geostatistical Simulation
BME Bayesian Maximum Entropy
PCA Principal Component Analysis
DWT Discrete Wavelet Transform
GTF Gradient Transfer Fusion
IHS Intensity Hue Saturation
ADP Aerosol Detection Product
JPSS Joint Polar Orbiting Satellie-Series
FFNN Feedforward Neural Network
HSI Hyperspectral Imaging
SWIR Shortwave Infrared
TLS Terrestrial laser Scanning
USRT Urban Solar Radiative Transfer
BRDF Bidirectional Reflectance Distribution Function
HPC High Performance Computing
GEE Google Earth Engine
AWS Amazon Web Service
NAIP National Agricultural Imagery Program
CEOS Committee on Earth Observing Satellites
WGCV Working Group on Calibration and Validation
PIML Physics Informed Machine Learning
PINN Physics Informed Neural Network
CNN Convolutional Neural Network
VIIRS Visible infrared Imaging Radiometer Suite
MODIS Moderate Resolution Imaging Spectroradiometer
TROPOMI Troposhperic Monitoring Instrument
AMSR-E Advanced Microwave Scaning Radiometer-Earth Observing System
LST Land Surface Temperature
NDVI Normalized Difference Vegetation Index
EVI Enhanced Vegetation Index
LAI Leaf Area Index
ET Evapotranspiration
LVIS Laser Vegetation Imaging Sensor
LGE LVIS Ground Elevation
USGS United States Geological Survey
ESA-B European Space Agency
SRTM Shuttle Radar Topography Mission
SAR Synthetic Aperture Radar
ALOS/PALSAR Advanced Land Observing Satellite/Phased Array L-band Synthetic Aperture Radar
LiDAR Light Detection and Ranging
CALIOP Cloud-Aerosol Lidar with Orthogonal POlarization
POLDER Polarization and Directionality of the Earth’s Reflectances
rh Relative Humidity
FI-GBM Fusion Imputation Gradient Boosting Machine
MLR Multimomial Logistic Regression
TTSM Transformer Temporal Spatial Model
MDL-RS Multimodal Deep Learning-Remote Sensing
AP-BME Active-Passive Bayesian Maximum Entropy
CCAM Clear air-Cloud-Aerosol-Mixed cloud and aerosol
DPOM Dust–Polluted dust–Other aerosol–Mixed aerosol
SF Surface Reflectance
CPR Cloud Profiling Radar
ATCS A-Train Cloud Segmentation
SIP Science Investigator-led Processing Systems
THEOS Thailand Earth Observation System
AQM Air Quality Monitoring
OTM Other Traditional Methods
ML Machine Learning

Appendix A

Table A1 summarizes the various fusion techniques techniques discussed and relevant to this study. It details their classification, the level at which the fusion is performed and the specific dat modalities utilized in each approach.
Table A1. Fusion Paradigmns: Techniques, modalities and fusion levels.
Table A1. Fusion Paradigmns: Techniques, modalities and fusion levels.
Fusion Paradigm Publication Technique Modalites/Sensor Fusion Level
Neural Network [41] Forestnet RGB, Infrared Bands (Landsat 8), Feature Level
CFSv2,
(Euclidean Distance, Elevation, slope, aspect, Peat)
Nerual Network [42] Hybrid CNN RGB, Near Infrared (PlanetScope), Featue Level
UAV-Optical
Neural Network [30] MDL-RS Hyperspectral (144 bands (364-1046 nm)), Feature Level
SAR,
Multispectral,
LiDAR
Neural Network [43] DAMA-WL RBG, Near-infrared (VIIRS), Feature Level
CALIOP
Neural Network [178] 2-Branch CNN HSI 144 Bands (IEEE Data fusion contest), Feature Level
LiDAR (IEEE Data fusion Contest)
Neural Network [179] FusAtNet HSI 144 Bands (IEEE Data fusion contest), Feature Level
LiDAR (IEEE Data Fusion contest)
Neural Network [180] TTSM Sentinel-2 Bands 8 and 4, Feature Level
S1 GRD (Sentinel-1 SAR)
Neural Network [69] FFNN V3 Operational Products (CALIIOP), Feature Level
Level-1b radiance + viewing/illumination geometries (VIIRS)
Neural Network [79] U-Net POLDER Level 1b Pixel Level
2B-CLDCLASS (RADAR)
Classical ML [48] FI-GBM Tropospheric NO2 (TROPOMI L2 version1.0.0-1.2.0), Featue Level
Ground Hourly NO2,
NDVI (MODIS),
Planet Boundary Layer (MERRA-2),
Land Use (ESA Second Phase Product),
Road Data (OpenStreetMap),
Population Density (GPWv4),
Hourly Meterological Data (CMDC)
Classical ML [181] MLR Hyperspectral (144 bands (364-1046 nm)) Feature Level
LiDAR (IEEE Data Fusion Contest)
Classical Machine Learning [182] Random Forest Sentinel 1 SAR Feature Level
Sentine 2 (B1, B9, B10)
Classical ML [183] Extra Tees (CCAM, DPOM) TOA Reflectance (FY-4A), Feature Level
Vertical Feature Mask (CALIPSO),
Surface Pressure, High and Low Vegetation indices, relative humidity, temperature, wind, vector and Land cover (ERA-5),
Ground PM2.5 and PM10, and visibility
Classical ML [49] Ensemble Methods (Random Forest, LightGBM, Bagging Tree, XGBoost, GB Decison TRee) CYGNSS L1B, Feature Level
GPM IMERGE Precipitation Data,
ECMWF SST data,
AMSRU Geophysical data,
MCD12C1 Land Cover data,
GSW data
Classical ML [53] KNN, SVM, ANN Landsat 5 (NDBI, NDVI, SAVI, MNDWI), Decision Level
HEOS or WorldView 3 (GEMI, NDVI, SAVI and MNDWI)
Classical ML [51] Random Forest X-Band AMSR-E JAXA Level3 soil moisture (AMSRE-E), Feature Level
LST-MYD11A2, Evapotranspiration-MOD16A2 (MODIS),
albedo-MCD43B3,
LAI-MCD15A2,
MYD13A2 (16-Day NDVI and EVI),
3B42 (TRMM daily rainfall data)
Geostatistical Methods [55] Universal Kriging AERONET (Level 2 AOT, Feature Level
Level 2 Aerosol product vF09_0017 (MISR),
Terra Collection 005 Level-2 (MODIS)
Geostatical Methods [56] SSDF AIRS Infrared Spectrometer Temperature soundings, Feature Level
SNPP-CrIMSS (Microwave-Infrared) Temperature Soundings
Geostatistical Methods [57] FILTERSIM SRTM (RADAR), Feature Level
GMTED2010 DEM,
Geostatistical Methods [54] Extend Fixed-Rank Krieging AIRS Level 2, Feature Level
Synthetic OCO2 data
OTM [59] Wavelet-PCA GF-1 (RGB and Near-infrared), Feature Level
PALSAR SAR,
Radarsat-2 SAR,
OTM [61] Discrete Wavelet Transform HJ-1B (RGB, Near-infrared), Pixel Level
ALOS/PALSAR SAR
OTM [66] AP-BME MOD04 L2 (Terra MODIS), Pixel Level
MYD04 L2 (Aqua MODIS)
OTM [68] IHS-GTF Sentinel -2A MSC LV1C (Bands 2-4) Pixel Level
SAR (TerraSAR)
OTM [96] Geometric Ray Tracing Shortwave Hyperspectral (970-2500mm) Pixel Level
Terrestrial laser Scanning data (Point and Intensity)

References

  1. Wald, L. A European proposal for terms of reference in data fusion. In Proceedings of the Commission VII Symposium" Resource and Environmental Monitoring", 1998, Vol. 32, pp. 651–654.
  2. Bokade, R.; Navato, A.; Ouyang, R.; Jin, X.; Chou, C.A.; Ostadabbas, S.; Mueller, A.V. A cross-disciplinary comparison of multimodal data fusion approaches and applications: Accelerating learning through trans-disciplinary information sharing. Expert Systems with Applications 2021, 165, 113885. [CrossRef]
  3. White, F.E. Data fusion lexicon, joint directors of laboratories, technical panel for C3, data fusion sub-panel. Naval Ocean Systems Center, San Diego 1987.
  4. Steinberg, A.N.; Bowman, C.L.; White, F.E.; Blasch, E.P.; Llinas, J. Evolution of the JDL model. Perspectives on Information Fusion 2023, 6, 36–40.
  5. León, F.P.; Ruser, H. Information fusion-Overview and taxonomy. Reports on Distributed Measurement Systems 2008, pp. Pages–1.
  6. Dong, J.; Zhuang, D.; Huang, Y.; Fu, J. Advances in multi-sensor data fusion: Algorithms and applications. Sensors 2009, 9, 7771–7784.
  7. Gustafsson, F. Statistical sensor fusion; Studentlitteratur, 2010.
  8. Zhang, J. Multi-source remote sensing data fusion: Status and trends. International Journal of Image and Data Fusion 2010, 1, 5–24. [CrossRef]
  9. Sidek, O.; Quadri, S. A review of data fusion models and systems. International Journal of Image and Data Fusion 2012, 3, 3–21.
  10. Ranchin, T.; Télédétection, G.; Antipolis, M.S. Data fusion in remote sensing: examples. Ecole des Mines de Paris [database online]. Sophia Antipolis, France, Available from http://www. researchgate. net/publication/47805585_Data_fusion_in_remote_sensing _examples/file/e0b49515ae591d7a04. pdf(last accessed June 2014) 2001.
  11. Ranchin, T. Data fusion in remote sensing: examples. Proceedings of the 4th annual conference on information fusion 2001.
  12. Li, J.; Hong, D.; Gao, L.; Yao, J.; Zheng, K.; Zhang, B.; Chanussot, J. Deep learning in multimodal remote sensing data fusion: A comprehensive review. International Journal of Applied Earth Observation and Geoinformation 2022, 112, 102926.
  13. Lahat, D.; Adali, T.; Jutten, C. Multimodal data fusion: an overview of methods, challenges, and prospects. Proceedings of the IEEE 2015, 103, 1449–1477.
  14. Clauson, J.; Cantrell, S.; Vrable, J.; Oeding, J.; Raniitkar, B.; Rusten, T.; Ramaseri, S.; Casey, K. Earth Observing Sensing Satellites Online Compendium: U.S Geological Survey digital data. Technical report, USGS, 2024.
  15. Ustin, S.L.; Middleton, E.M. Current and near-term advances in Earth observation for ecological applications. Ecological Processes 2021, 10, 1.
  16. Zhang, K.; Zhang, F.; Wan, W.; Yu, H.; Sun, J.; Del Ser, J.; Elyan, E.; Hussain, A. Panchromatic and multispectral image fusion for remote sensing and earth observation: Concepts, taxonomy, literature review, evaluation methodologies and challenges ahead. Information Fusion 2023, 93, 227–242. [CrossRef]
  17. Dowman, I.; Reuter, H.I. Global geospatial data from Earth observation: status and issues. International Journal of Digital Earth 2017, 10, 328–341. [CrossRef]
  18. Salcedo-Sanz, S.; Ghamisi, P.; Piles, M.; Werner, M.; Cuadra, L.; Moreno-Martínez, A.; Izquierdo-Verdiguier, E.; Muñoz-Marí, J.; Mosavi, A.; Camps-Valls, G. Machine learning information fusion in Earth observation: A comprehensive review of methods, applications and data sources. Information Fusion 2020, 63, 256–272.
  19. Zhang, Z.M.; Catalucci, S.; Thompson, A.; Leach, R.; Piano, S. Applications of data fusion in optical coordinate metrology: a review. The International Journal of Advanced Manufacturing Technology 2023, 124, 1341–1356.
  20. Kalamkar, S.; A., G.M. Multimodal image fusion: A systematic review. Decision Analytics Journal 2023, 9, 100327. [CrossRef]
  21. Ounoughi, C.; Yahia, S.B. Data fusion for ITS: A systematic literature review. Information Fusion 2023, 89, 267–291.
  22. Wu, Y.; Chang, E.Y.; Chang, K.C.C.; Smith, J.R. Optimal multimodal fusion for multimedia data analysis. In Proceedings of the Proceedings of the 12th annual ACM international conference on Multimedia, 2004, pp. 572–579.
  23. Mitchell, H.B., Spatial Alignment. In Image Fusion: Theories, Techniques and Applications; Springer Berlin Heidelberg: Berlin, Heidelberg, 2010; pp. 35–51. [CrossRef]
  24. Herold, M. Remote sensing science 2.0; Wageningen University: Netherlands, 2011.
  25. Amit, D.B.T.A.P.; Patel, V.P.P.M. Geometric Distortion and Correction Methods for Finding Key Points: A Survey. International Journal for Scientific Research & Development 2016.
  26. Toutin, T. Review Paper: Geometric processing of remote sensing images: models, algorithms and methods, 2004, 2004.
  27. Chapter 2 - Geometric processing and positioning techniques. In Advanced Remote Sensing (Second Edition), Second Edition ed.; Liang, S.; Wang, J., Eds.; Academic Press, 2020; pp. 59–105. [CrossRef]
  28. GeoLearn. Radiometric Correction, 2025. Accessed: 2025-02-05.
  29. Jenerowicz, A.; Wierzbicki, D.; Kedzierski, M. Radiometric Correction with Topography Influence of Multispectral Imagery Obtained from Unmanned Aerial Vehicles. Remote Sensing 2023, 15, 2059.
  30. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification. IEEE Transactions on Geoscience and Remote Sensing 2021, 59, 4340–4354.
  31. Pohl, C.; Van Genderen, J.L. Review article multisensor image fusion in remote sensing: concepts, methods and applications. International journal of remote sensing 1998, 19, 823–854.
  32. Al-Wassai, F.A.; Kalyankar, N.; Al-Zaky, A.A. Multisensor images fusion based on feature-level. arXiv preprint arXiv:1108.4098 2011.
  33. Zhang, J. Multi-source remote sensing data fusion: status and trends. International Journal of Image and Data Fusion 2010, 1, 5–24.
  34. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE transactions on pattern analysis and machine intelligence 2018, 41, 423–443.
  35. Sarker, I.H. Deep learning: a comprehensive overview on techniques, taxonomy, applications and research directions. SN computer science 2021, 2, 420.
  36. Jin, H.; Mountrakis, G. Fusion of optical, radar and waveform LiDAR observations for land cover classification. ISPRS Journal of Photogrammetry and Remote Sensing 2022, 187, 171–190.
  37. Abebe, G.; Getachew, D.; Ewunetu, A. Analysing land use/land cover changes and its dynamics using remote sensing and GIS in Gubalafito district, Northeastern Ethiopia. SN Applied Sciences 2022, 4, 30.
  38. Mehra, N.; Swain, J.B. Assessment of land use land cover change and its effects using artificial neural network-based cellular automation. Journal of Engineering and Applied Science 2024, 71, 70.
  39. Amini, S.; Saber, M.; Rabiei-Dastjerdi, H.; Homayouni, S. Urban land use and land cover change analysis using random forest classification of landsat time series. Remote Sensing 2022, 14, 2654.
  40. Tewabe, D.; Fentahun, T. Assessing land use and land cover change detection using remote sensing in the Lake Tana Basin, Northwest Ethiopia. Cogent Environmental Science 2020, 6, 1778998.
  41. Irvin, J.; Sheng, H.; Ramachandran, N.; Johnson-Yu, S.; Zhou, S.; Story, K.; Rustowicz, R.; Elsworth, C.; Austin, K.; Ng, A.Y. Forestnet: Classifying drivers of deforestation in indonesia using deep learning on satellite imagery. arXiv preprint arXiv:2011.05479 2020.
  42. Hicks, D.; Kastner, R.; Schurgers, C.; Hsu, A.; Aburto, O. Mangrove Ecosystem Detection using Mixed-Resolution Imagery with a Hybrid-Convolutional Neural Network 2020.
  43. Huang, X.; Ali, S.; Purushotham, S.; Wang, J.; Wang, C.; Zhang, Z. Deep multi-sensor domain adaptation on active and passive satellite remote sensing data. In Proceedings of the 1st KDD Workshop on Deep Learning for Spatiotemporal Data, Applications, and Systems (DeepSpatial 2020), 2020.
  44. Zhang, H.; Ni, W.; Yan, W.; Xiang, D.; Wu, J.; Yang, X.; Bian, H. Registration of multimodal remote sensing image based on deep fully convolutional neural network. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 2019, 12, 3028–3042.
  45. Talukdar, S.; Singha, P.; Mahato, S.; Pal, S.; Liou, Y.A.; Rahman, A. Land-use land-cover classification by machine learning classifiers for satellite observations—A review. Remote sensing 2020, 12, 1135.
  46. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS journal of photogrammetry and remote sensing 2016, 114, 24–31.
  47. Zhang, Y.; Zhang, H.; Lin, H. Improving the impervious surface estimation with combined use of optical and SAR remote sensing images. Remote Sensing of Environment 2014, 141, 155–167.
  48. Fu, J.; Tang, D.; Grieneisen, M.L.; Yang, F.; Yang, J.; Wu, G.; Wang, C.; Zhan, Y. A machine learning-based approach for fusing measurements from standard sites, low-cost sensors, and satellite retrievals: Application to NO2 pollution hotspot identification. Atmospheric Environment 2023, 302, 119756.
  49. Zhang, Y.; Bu, J.; Zuo, X.; Yu, K.; Wang, Q.; Huang, W. Vegetation Water Content Retrieval from Spaceborne GNSS-R and Multi-Source Remote Sensing Data Using Ensemble Machine Learning Methods. Remote Sensing 2024, 16, 2793.
  50. Im, J.; Park, S.; Rhee, J.; Baik, J.; Choi, M. Downscaling of AMSR-E soil moisture with MODIS products using machine learning approaches. Environmental Earth Sciences 2016, 75, 1–19.
  51. Park, S.; Im, J.; Park, S.; Rhee, J. Drought monitoring using high resolution soil moisture through multi-sensor satellite data fusion over the Korean peninsula. Agricultural and Forest Meteorology 2017, 237, 257–269.
  52. Sharma, M.; Kushwaha, P.; Kumari, P.; Kumari, P.; Yadav, R. Machine Learning Techniques in Data Fusion: A Review. In Proceedings of the International Conference on Communication and Intelligent Systems. Springer, 2022, pp. 391–405.
  53. Puttinaovarat, S.; Horkaew, P. Urban areas extraction from multi sensor data based on machine learning and data fusion. Pattern Recognition and Image Analysis 2017, 27, 326–337.
  54. Braverman, A.; Cressie, N.; Katzfuss, M.; Michalak, A.; Miller, C.; Nguyen, H.; Olsen, E.; Wang, R. Geostatistical data fusion for remote sensing applications. In Proceedings of the AGU Fall Meeting Abstracts C, 2011, Vol. 1014, p. 2009.
  55. Chatterjee, A.; Michalak, A.M.; Kahn, R.A.; Paradise, S.R.; Braverman, A.J.; Miller, C.E. A geostatistical data fusion technique for merging remote sensing and ground-based observations of aerosol optical thickness. Journal of Geophysical Research: Atmospheres 2010, 115.
  56. Kalmus, P.; Nguyen, H.; Roman, J.; Wang, T.; Yue, Q.; Wen, Y.; Hobbs, J.; Braverman, A. Data fusion of AIRS and CrIMSS near surface air temperature. Earth and Space Science 2022, 9, e2022EA002282.
  57. Tang, Y.; Zhang, J.; Jing, L.; Li, H. Digital elevation data fusion using multiple-point geostatistical simulation. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 2015, 8, 4922–4934.
  58. Zhang, T.; Switzer, P.; Journel, A. Filter-based classification of training image patterns for spatial simulation. Mathematical Geology 2006, 38, 63–80.
  59. Fu, B.; Wang, Y.; Campbell, A.; Li, Y.; Zhang, B.; Yin, S.; Xing, Z.; Jin, X. Comparison of object-based and pixel-based Random Forest algorithm for wetland vegetation mapping using high spatial resolution GF-1 and SAR data. Ecological indicators 2017, 73, 105–117.
  60. Otukei, J.R.; Blaschke, T.; Collins, M. Fusion of TerraSAR-x and Landsat ETM+ data for protected area mapping in Uganda. International Journal of Applied Earth Observation and Geoinformation 2015, 38, 99–104.
  61. Wang, X.Y.; Guo, Y.; He, J.; Du, L. Fusion of HJ1B and ALOS PALSAR data for land cover classification using machine learning methods. International journal of applied earth observation and geoinformation 2016, 52, 192–203.
  62. Senkerik, R. Remote Sensing Image Fusion Based on PCA and Wavelets. In Proceedings of the Intelligent Data Engineering and Analytics: Proceedings of the 10th International Conference on Frontiers in Intelligent Computing: Theory and Applications (FICTA 2022). Springer Nature, 2023, Vol. 327, p. 25.
  63. Amolins, K.; Zhang, Y.; Dare, P. Wavelet based image fusion techniques—An introduction, review and comparison. ISPRS Journal of photogrammetry and Remote Sensing 2007, 62, 249–263.
  64. Shao, Z.; Pan, Y.; Diao, C.; Cai, J. Cloud detection in remote sensing images based on multiscale features-convolutional neural network. IEEE Transactions on Geoscience and Remote Sensing 2019, 57, 4062–4076.
  65. Shao, Z.; Deng, J.; Wang, L.; Fan, Y.; Sumari, N.S.; Cheng, Q. Fuzzy autoencode based cloud detection for remote sensing imagery. Remote Sensing 2017, 9, 311.
  66. Xia, X.; Zhao, B.; Zhang, T.; Wang, L.; Gu, Y.; Liou, K.N.; Mao, F.; Liu, B.; Bo, Y.; Huang, Y.; et al. Satellite-derived aerosol optical depth fusion combining active and passive remote sensing based on Bayesian maximum entropy. IEEE Transactions on Geoscience and Remote Sensing 2021, 60, 1–13.
  67. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Information Fusion 2016, 31, 100–109. [CrossRef]
  68. Shao, Z.; Wu, W.; Guo, S. IHS-GTF: A fusion method for optical and synthetic aperture radar data. Remote Sensing 2020, 12, 2796.
  69. Lee, J.; Shi, Y.R.; Cai, C.; Ciren, P.; Wang, J.; Gangopadhyay, A.; Zhang, Z. Machine learning based algorithms for global dust aerosol detection from satellite images: inter-comparisons and evaluation. Remote Sensing 2021, 13, 456.
  70. Ciren, P.; Kondragunta, S. JPSS Enterprise Processing System Aerosol Detection Product. Technical report, NOAA-Center for Satellite Applications and Research, 2022.
  71. IPCC. Climate Change 2022: Impacts, Adaptation and Vulnerability; Summary for Policymakers, Cambridge University Press: Cambridge, UK and New York, USA, 2022; pp. 3–33.
  72. Wang, J.; Rossow, W.B. Effects of cloud vertical structure on atmospheric circulation in the GISS GCM. Journal of climate 1998, 11, 3010–3029.
  73. Ruppert Jr, J.H.; Wing, A.A.; Tang, X.; Duran, E.L. The critical role of cloud–infrared radiation feedback in tropical cyclone development. Proceedings of the National Academy of Sciences 2020, 117, 27884–27892.
  74. Bu, Y.P.; Fovell, R.G.; Corbosiero, K.L. Influence of cloud–radiative forcing on tropical cyclone structure. Journal of the Atmospheric Sciences 2014, 71, 1644–1662.
  75. Kato, S.; Sun-Mack, S.; Miller, W.F.; Rose, F.G.; Chen, Y.; Minnis, P.; Wielicki, B.A. Relationships among cloud occurrence frequency, overlap, and effective thickness derived from CALIPSO and CloudSat merged cloud vertical profiles. Journal of Geophysical Research: Atmospheres 2010, 115.
  76. Zhang, R.; Wang, Y.; Li, Z.; Wang, Z.; Dickerson, R.R.; Ren, X.; He, H.; Wang, F.; Gao, Y.; Chen, X.; et al. Vertical profiles of cloud condensation nuclei number concentration and its empirical estimate from aerosol optical properties over the North China Plain. Atmospheric Chemistry and Physics 2022, 22, 14879–14891. [CrossRef]
  77. Bertrand, L.; Kay, J.E.; Haynes, J.; de Boer, G. A global gridded dataset for cloud vertical structure from combined CloudSat and CALIPSO observations. Earth System Science Data 2024, 16, 1301–1316. [CrossRef]
  78. Sullivan, S.C.; Hoose, C. Clouds and Their Climatic Impact: Radiation, Circulation, and Precipitation; Vol. 281, John Wiley & Sons, 2023.
  79. Foley, S.R.; Knobelspiesse, K.D.; Sayer, A.M.; Gao, M.; Hays, J.; Hoffman, J. 3-D Cloud Masking Across a Broad Swath using Multi-angle Polarimetry and Deep Learning. EGUsphere 2024, 2024, 1–24. [CrossRef]
  80. Cheng, Y.; He, H.; Xue, Q.; Yang, J.; Zhong, W.; Zhu, X.; Peng, X. Remote Sensing Retrieval of Cloud Top Height Using Neural Networks and Data from Cloud-Aerosol Lidar with Orthogonal Polarization. Sensors 2024, 24, 541.
  81. Luo, H.; Quaas, J.; Han, Y. Examining cloud vertical structure and radiative effects from satellite retrievals and evaluation of CMIP6 scenarios. Atmospheric Chemistry and Physics 2023, 23, 8169–8186. [CrossRef]
  82. Jozef, G.C.; Cassano, J.J.; Dahlke, S.; Dice, M.; Cox, C.J.; de Boer, G. An overview of the vertical structure of the atmospheric boundary layer in the central Arctic during MOSAiC. Atmospheric Chemistry and Physics 2024, 24, 1429–1450.
  83. Di Gregorio, A. Land cover classification system: classification concepts and user manual: LCCS; Vol. 2, Food & Agriculture Org., 2005.
  84. Sleeter, B.; Loveland, T.; Domke, G.; Herold, N.; Wickham, J.; Wood, N. Chapter 5: Land cover and land use change. Impacts, risks, and adaptation in the United States: The fourth national climate assessment 2018, 2.
  85. Bununu, Y.A.; Bello, A.; Ahmed, A. Land cover, land use, climate change and food security. Sustainable Earth Reviews 2023, 6, 16.
  86. Frankenberger, J. Land Use.
  87. Frankenberger, J. Water resources of Boone County, Indiana. Watershed Connections 2010.
  88. Roy, P. Land Use and Land cover Change and Human dimensions. LinkedIn 2021.
  89. Saah, D.; Tenneson, K.; Matin, M.; Uddin, K.; Cutter, P.; Poortinga, A.; Nguyen, Q.H.; Patterson, M.; Johnson, G.; Markert, K.; et al. Land cover mapping in data scarce environments: challenges and opportunities. Frontiers in Environmental Science 2019, 7, 150.
  90. Jia, K.; Wei, X.; Gu, X.; Yao, Y.; Xie, X.; Li, B. Land cover classification using Landsat 8 operational land imager data in Beijing, China. Geocarto international 2014, 29, 941–951.
  91. Shahtahmassebi, A.; Yang, N.; Wang, K.; Moore, N.; Shen, Z. Review of shadow detection and de-shadowing methods in remote sensing. Chinese geographical science 2013, 23, 403–420.
  92. Dong, X.; Cao, J.; Zhao, W. A review of research on remote sensing images shadow detection and application to building extraction. European Journal of Remote Sensing 2024, 57, 2293163.
  93. Arévalo, V.; González, J.; Ambrosio, G. Shadow detection in colour high-resolution satellite images. International Journal of Remote Sensing 2008, 29, 1945–1963.
  94. Yang, L.; Ren, J. Remote sensing image restoration using estimated point spread function. In Proceedings of the 2010 International Conference on Information, Networking and Automation (ICINA), 2010, Vol. 1, pp. V1–48–V1–52. [CrossRef]
  95. Zhang, G.; Cerra, D.; Müller, R. Shadow detection and restoration for hyperspectral images based on nonlinear spectral unmixing. Remote Sensing 2020, 12, 3985.
  96. Hartzell, P.; Glennie, C.; Khan, S. Terrestrial hyperspectral image shadow restoration through lidar fusion. Remote Sensing 2017, 9, 421.
  97. Fernandez-Beltran, R.; Guzmán-Ponce, A.; Fernandez, R.; Kang, J.; García-Mateos, G. Shadow detection using a cross-attentional dual-decoder network with self-supervised image reconstruction features. Image and Vision Computing 2024, 143, 104922. [CrossRef]
  98. Li, T.; Hu, D.; Wang, Y.; Di, Y.; Liu, M. Correcting remote-sensed shaded image with urban surface radiative transfer model. International Journal of Applied Earth Observation and Geoinformation 2022, 106, 102654. [CrossRef]
  99. Pradham, P.; Younan, N.H.; King, R.L. Concepts of image fusion in remote sensing applications. Image fusion: algorithms and applications 2008, pp. 391–428.
  100. Kotawadekar, R. 9 - Satellite data: big data extraction and analysis. In Artificial Intelligence in Data Mining; Binu, D.; Rajakumar, B., Eds.; Academic Press, 2021; pp. 177–197. [CrossRef]
  101. Zhu, L.; Suomalainen, J.; Liu, J.; Hyyppä, J.; Kaartinen, H.; Haggren, H. A Review: Remote Sensing Sensors. In Multi-purposeful Application of Geospatial Data; Rustamov, R.B.; Hasanova, S.; Zeynalova, M.H., Eds.; IntechOpen: Rijeka, 2017; chapter 2. [CrossRef]
  102. Sulaiman, N.; Che’Ya, N.N.; Mohd Roslim, M.H.; Juraimi, A.S.; Mohd Noor, N.; Fazlil Ilahi, W.F. The Application of Hyperspectral Remote Sensing Imagery (HRSI) for Weed Detection Analysis in Rice Fields: A Review. Applied Sciences 2022, 12. [CrossRef]
  103. Kulkarni, S.C.; Rege, P.P. Pixel level fusion techniques for SAR and optical images: A review. Information Fusion 2020, 59, 13–29.
  104. Wang, Z.; Ma, Y.; Zhang, Y. Review of pixel-level remote sensing image fusion based on deep learning. Information Fusion 2023, 90, 36–58. [CrossRef]
  105. Liu, S.; Wu, G.; Zhang, X.; Zhang, K.; Wang, P.; Li, Y. SAR despeckling via classification-based nonlocal and local sparse representation. Neurocomputing 2017, 219, 174–185.
  106. Li, Y.; Wei, F.; Zhang, Y.; Chen, W.; Ma, J. HS2P: Hierarchical spectral and structure-preserving fusion network for multimodal remote sensing image cloud and shadow removal. Information Fusion 2023, 94, 215–228.
  107. Choi, H.; Jeong, J. Speckle Noise Reduction Technique for SAR Images Using Statistical Characteristics of Speckle Noise and Discrete Wavelet Transform. Remote Sensing 2019, 11. [CrossRef]
  108. Schowengerdt, R.A. Remote sensing: models and methods for image processing; elsevier, 2006.
  109. Zhang, J.; Li, J. Chapter 11 - Spacecraft. In Spatial Cognitive Engine Technology; Zhang, J.; Li, J., Eds.; Academic Press, 2023; pp. 129–162. [CrossRef]
  110. Guo, H.; Ye, D.; Xu, H.; Bruzzone, L. OBSUM: An object-based spatial unmixing model for spatiotemporal fusion of remote sensing images. Remote Sensing of Environment 2024, 304, 114046.
  111. Wang, Q.; Peng, K.; Tang, Y.; Tong, X.; Atkinson, P.M. Blocks-removed spatial unmixing for downscaling MODIS images. Remote Sensing of Environment 2021, 256, 112325. [CrossRef]
  112. Wu, M.; Niu, Z.; Wang, C.; Wu, C.; Wang, L. Use of MODIS and Landsat time series data to generate high-resolution temporal synthetic Landsat data using a spatial and temporal reflectance fusion model. Journal of Applied Remote Sensing 2012, 6, 063507–063507.
  113. Xu, Y.; Huang, B.; Xu, Y.; Cao, K.; Guo, C.; Meng, D. Spatial and temporal image fusion via regularized spatial unmixing. IEEE Geoscience and Remote Sensing Letters 2015, 12, 1362–1366.
  114. Zhu, X.; Helmer, E.H.; Gao, F.; Liu, D.; Chen, J.; Lefsky, M.A. A flexible spatiotemporal method for fusing satellite images with different resolutions. Remote Sensing of Environment 2016, 172, 165–177.
  115. Guo, H.; Ye, D.; Xu, H.; Bruzzone, L. OBSUM: An object-based spatial unmixing model for spatiotemporal fusion of remote sensing images. Remote Sensing of Environment 2024, 304, 114046. [CrossRef]
  116. Emelyanova, I.V.; McVicar, T.R.; Van Niel, T.G.; Li, L.T.; Van Dijk, A.I. Assessing the accuracy of blending Landsat–MODIS surface reflectances in two landscapes with contrasting spatial and temporal dynamics: A framework for algorithm selection. Remote Sensing of Environment 2013, 133, 193–209.
  117. Chen, X.; Liu, M.; Zhu, X.; Chen, J.; Zhong, Y.; Cao, X. " Blend-then-Index" or" Index-then-Blend": A theoretical analysis for generating high-resolution NDVI time series by STARFM. Photogrammetric Engineering & Remote Sensing 2018, 84, 65–73.
  118. Li, X.; Ling, F.; Foody, G.M.; Boyd, D.S.; Jiang, L.; Zhang, Y.; Zhou, P.; Wang, Y.; Chen, R.; Du, Y. Monitoring high spatiotemporal water dynamics by fusing MODIS, Landsat, water occurrence data and DEM. Remote Sensing of Environment 2021, 265, 112680.
  119. Zhu, X.; Cai, F.; Tian, J.; Williams, T.K.A. Spatiotemporal fusion of multisource remote sensing data: Literature survey, taxonomy, principles, applications, and future directions. Remote Sensing 2018, 10, 527.
  120. Shen, H.; Li, H.; Qian, Y.; Zhang, L.; Yuan, Q. An effective thin cloud removal procedure for visible remote sensing images. ISPRS Journal of Photogrammetry and Remote Sensing 2014, 96, 224–235.
  121. U.S. Geological Survey. Landsat 7. Online.
  122. Gevaert, C.M.; García-Haro, F.J. A comparison of STARFM and an unmixing-based algorithm for Landsat and MODIS data fusion. Remote sensing of Environment 2015, 156, 34–44.
  123. Cooley, S.W.; Smith, L.C.; Stepan, L.; Mascaro, J. Tracking dynamic northern surface water changes with high-frequency planet CubeSat imagery. Remote Sensing 2017, 9, 1306.
  124. Niroumand-Jadidi, M.; Bovolo, F.; Bruzzone, L.; Gege, P. Physics-based bathymetry and water quality retrieval using planetscope imagery: Impacts of 2020 COVID-19 lockdown and 2019 extreme flood in the Venice Lagoon. Remote Sensing 2020, 12, 2381.
  125. Caballero, G.; Pezzola, A.; Winschel, C.; Sanchez Angonova, P.; Casella, A.; Orden, L.; Salinero-Delgado, M.; Reyes-Muñoz, P.; Berger, K.; Delegido, J.; et al. Synergy of Sentinel-1 and Sentinel-2 Time Series for Cloud-Free Vegetation Water Content Mapping with Multi-Output Gaussian Processes. Remote Sensing 2023, 15. [CrossRef]
  126. Gao, F.; Masek, J.; Schwaller, M.; Hall, F. On the blending of the Landsat and MODIS surface reflectance: Predicting daily Landsat surface reflectance. IEEE Transactions on Geoscience and Remote sensing 2006, 44, 2207–2218.
  127. Piles, M.; Van der Schalie, R.; Gruber, A.; Muñoz-Marí, J.; Camps-Valls, G.; Mateo-Sanchis, A.; Dorigo, W.; de Jeu, R. Global estimation of soil moisture persistence with L and C-band microwave sensors. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium. IEEE, 2018, pp. 8259–8262.
  128. Wang, Q.; Tang, Y.; Ge, Y.; Xie, H.; Tong, X.; Atkinson, P.M. A comprehensive review of spatial-temporal-spectral information reconstruction techniques. Science of Remote Sensing 2023, 8, 100102.
  129. Shi, W.; Guo, D.; Zhang, H. A reliable and adaptive spatiotemporal data fusion method for blending multi-spatiotemporal-resolution satellite images. Remote Sensing of Environment 2022, 268, 112770. [CrossRef]
  130. Zhu, X.; Chen, J.; Gao, F.; Chen, X.; Masek, J.G. An enhanced spatial and temporal adaptive reflectance fusion model for complex heterogeneous regions. Remote Sensing of Environment 2010, 114, 2610–2623.
  131. Yang, J.; Yao, Y.; Wei, Y.; Zhang, Y.; Jia, K.; Zhang, X.; Shang, K.; Bei, X.; Guo, X. A Robust Method for Generating High-Spatiotemporal-Resolution Surface Reflectance by Fusing MODIS and Landsat Data. Remote Sensing 2020, 12. [CrossRef]
  132. Cao, Z.; Chen, S.; Gao, F.; Li, X. Improving phenological monitoring of winter wheat by considering sensor spectral response in spatiotemporal image fusion. Physics and Chemistry of the Earth, Parts A/B/C 2020, 116, 102859. [CrossRef]
  133. Zhao, Y.; Huang, B.; Song, H. A robust adaptive spatial and temporal image fusion model for complex land surface changes. Remote sensing of environment 2018, 208, 42–62.
  134. Wang, J.; Huang, B. A Rigorously-Weighted Spatiotemporal Fusion Model with Uncertainty Analysis. Remote Sensing 2017, 9. [CrossRef]
  135. Wu, M.C.; Ho, J.Y.; Lee, K.T.; Lin, G.F.; Liu, W.C. Development of real-time rainfall and runoff forecasting techniques. Taiwan Water Conservancy 2015, 63, 14 – 25. Cited by: 0.
  136. Wang, Q.; Atkinson, P.M. Spatio-temporal fusion for daily Sentinel-2 images. Remote Sensing of Environment 2018, 204, 31–42.
  137. Liu, M.; Ke, Y.; Yin, Q.; Chen, X.; Im, J. Comparison of five spatio-temporal satellite image fusion models over landscapes with various spatial heterogeneity and temporal variation. Remote Sensing 2019, 11, 2612.
  138. Chi, M.; Plaza, A.; Benediktsson, J.A.; Sun, Z.; Shen, J.; Zhu, Y. Big Data for Remote Sensing: Challenges and Opportunities. Proceedings of the IEEE 2016, 104, 2207–2219. [CrossRef]
  139. Ma, Y.; Wu, H.; Wang, L.; Huang, B.; Ranjan, R.; Zomaya, A.; Jie, W. Remote sensing big data computing: Challenges and opportunities. Future Generation Computer Systems 2015, 51, 47–60. Special Section: A Note on New Trends in Data-Aware Scheduling and Resource Provisioning in Modern HPC Systems, . [CrossRef]
  140. Dritsas, E.; Trigka, M. Remote Sensing and Geospatial Analysis in the Big Data Era: A Survey. Remote Sensing 2025, 17. [CrossRef]
  141. Vaddi, R.; Phaneendra Kumar, B.; Manoharan, P.; Agilandeeswari, L.; Sangeetha, V. Strategies for dimensionality reduction in hyperspectral remote sensing: A comprehensive overview. The Egyptian Journal of Remote Sensing and Space Sciences 2024, 27, 82–92. [CrossRef]
  142. Pan, S.; Li, Y.; Xu, Z.; Chong, Y. Distributed storage algorithm for geospatial image data based on data access patterns. PLoS One 2015, 10, e0133029.
  143. Kritikos, K.; Rousakis, Y.; Kotzinos, D. A cloud-based, geospatial linked data management system. Transactions on Large-Scale Data-and Knowledge-Centered Systems XX: Special Issue on Advanced Techniques for Big Data Management 2015, pp. 59–89.
  144. Gao, F.; Yue, P.; Cao, Z.; Zhao, S.; Shangguan, B.; Jiang, L.; Hu, L.; Fang, Z.; Liang, Z. A multi-source spatio-temporal data cube for large-scale geospatial analysis. International Journal of Geographical Information Science 2022, 36, 1853–1884.
  145. Nistor, A. Development of Geospatial Technologies by Using Artificial Intelligence. FAIMA Business & Management Journal 2024, 12, 61–68.
  146. Yin, J.; Qi, C.; Chen, Q.; Qu, J. Spatial-spectral network for hyperspectral image classification: A 3-D CNN and Bi-LSTM framework. Remote Sensing 2021, 13, 2353.
  147. Yin, L.; Xie, J. Multi-temporal-spatial-scale temporal convolution network for short-term load forecasting of power systems. Applied Energy 2021, 283, 116328.
  148. Han, H.; Liu, Z.; Li, J.; Zeng, Z. Challenges in remote sensing based climate and crop monitoring: navigating the complexities using AI. Journal of cloud computing 2024, 13, 1–14.
  149. Huq, M.E.; Dughairi, A.; Rahman, M. Remote Sensing Big Data: Challenges, Opportunities, Management and Application Introduction. Journal of Arabic and Human Sciences 2020, 13, 1–15.
  150. Czapiewski, S. Szumi nska, D. An Overview of Remote Sensing Data Applications in Peatland Research Based on Works from the Period 2010–2021. Land 2022, 11, 24, 2021.
  151. Zhu, Z.; Wulder, M.A.; Roy, D.P.; Woodcock, C.E.; Hansen, M.C.; Radeloff, V.C.; Healey, S.P.; Schaaf, C.; Hostert, P.; Strobl, P.; et al. Benefits of the free and open Landsat data policy. Remote Sensing of Environment 2019, 224, 382–385. [CrossRef]
  152. Bartel, P.; Cameron, G. Addressing the Challenges of Using Earth Observation Data for SDG Attainment: Evidence from the Global South and West Africa Region. Technical report, Thematic Research Network on Data and Statistics (TRENDS), Partners for Review, 2022.
  153. Niroumand-Jadidi, M.; Bovolo, F. Water quality retrieval and algal bloom detection using high-resolution CubeSat imagery. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences 2021, 3, 191–195.
  154. Hively, W.D.; Shermeyer, J.; Lamb, B.T.; Daughtry, C.T.; Quemada, M.; Keppler, J. Mapping crop residue by combining Landsat and WorldView-3 satellite imagery. Remote Sensing 2019, 11, 1857.
  155. Selmy, S.A.; Kucher, D.E.; Yang, Y.; García-Navarro, F.J. Geospatial Data: Acquisition, Applications, and Challenges. In Exploring Remote Sensing; Abdalla, R.M., Ed.; IntechOpen: Rijeka, 2024. [CrossRef]
  156. Leeonis, A.N.; Ahmed, M.F.; Mokhtar, M.B.; Lim, C.K.; Halder, B. Challenges of Using a Geographic Information System (GIS) in Managing Flash Floods in Shah Alam, Malaysia. Sustainability 2024, 16. [CrossRef]
  157. Zhang, S.; Lippitt, C.D.; Bogus, S.M.; Neville, P.R. Characterizing pavement surface distress conditions with hyper-spatial resolution natural color aerial photography. Remote Sensing 2016, 8, 392.
  158. Scheller, J.H.; Mastepanov, M.; Christensen, T.R. Toward UAV-based methane emission mapping of Arctic terrestrial ecosystems. Science of The Total Environment 2022, 819, 153161. [CrossRef]
  159. Harris, A.; Baird, A. Microtopographic drivers of vegetation patterning in blanket peatlands recovering from erosion. Ecosystems 2019, 22, 1035–1054.
  160. WGCV-CEOS. Working Group on Calibration and Validation. Accessed: 2025-03-17.
  161. Rastogi, A.; Antala, M.; Prikaziuk, E.; Yang, P.; van der Tol, C.; Juszczak, R. Exploring the Potential of SCOPE Model for Detection of Leaf Area Index and Sun-Induced Fluorescence of Peatland Canopy. Remote Sensing 2022, 14, 4010.
  162. Jiang, H. Indirect validation of ocean remote sensing data via numerical model: An example of wave heights from altimeter. Remote Sensing 2020, 12, 2627.
  163. Wu, X.; Xiao, Q.; Wen, J.; You, D.; Hueni, A. Advances in quantitative remote sensing product validation: Overview and current status. Earth-Science Reviews 2019, 196, 102875.
  164. Jin, R.; Li, X.; Ma, M.; Ge, Y.; Che, T.; Xiao, Q.; Zhao, K.; Xin, X. Remote sensing products validation activity and observation network in China. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), 2016, pp. 7623–7626. [CrossRef]
  165. Delaney, B.; Tansey, K.; Whelan, M. Satellite Remote Sensing Techniques and Limitations for Identifying Bare Soil. Remote Sensing 2025, 17, 630.
  166. Arroyo-Mora, J.; Kalacska, M.; Soffer, R.; Ifimov, G.; Leblanc, G.; Schaaf, E.; Lucanus, O. Evaluation of phenospectral dynamics with Sentinel-2A using a bottom-up approach in a northern ombrotrophic peatland. Remote sensing of environment 2018, 216, 544–560.
  167. Mhangara, P.; Mapurisa, W.; Mudau, N. Comparison of Image Fusion Techniques Using Satellite Pour l’Observation de la Terre (SPOT) 6 Satellite Imagery. Applied Sciences 2020, 10. [CrossRef]
  168. Wang, Q.; Shi, W.; Li, Z.; Atkinson, P.M. Fusion of Sentinel-2 images. Remote Sensing of Environment 2016, 187, 241–252. [CrossRef]
  169. Yilmaz, V.; Serifoglu Yilmaz, C.; Güngör, O.; Shan, J. A genetic algorithm solution to the gram-schmidt image fusion. International Journal of Remote Sensing 2020, 41, 1458–1485.
  170. Pohl, C. Challenges of remote sensing image fusion to optimize earth observation data exploitation. European Scientific Journal 2013.
  171. Allu, A.R.; Mesapam, S. Impact of remote sensing data fusion on agriculture applications: A review. European Journal of Agronomy 2025, 164, 127478.
  172. Kumar, S.; Imen, S.; Sridharan, V.K.; Gupta, A.; McDonald, W.; Ramirez-Avila, J.J.; Abdul-Aziz, O.I.; Talchabhadel, R.; Gao, H.; Quinn, N.W.; et al. Perceived barriers and advances in integrating earth observations with water resources modeling. Remote Sensing Applications: Society and Environment 2024, 33, 101119.
  173. Toscano, J.D.; Oommen, V.; Varghese, A.J.; Zou, Z.; Ahmadi Daryakenari, N.; Wu, C.; Karniadakis, G.E. From pinns to pikans: Recent advances in physics-informed machine learning. Machine Learning for Computational Science and Engineering 2025, 1, 1–43.
  174. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics 2019, 378, 686–707.
  175. Zhang, C.; Wang, Q.; Atkinson, P.M. Unsupervised object-based spectral unmixing for subpixel mapping. Remote Sensing of Environment 2025, 318, 114514.
  176. Liu, L.; Santos, J.E.; Prodanović, M.; Pyrcz, M.J. Mitigation of spatial nonstationarity with vision transformers. Computers & Geosciences 2023, 178, 105412. [CrossRef]
  177. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 2020.
  178. Feng, Q.; Zhu, D.; Yang, J.; Li, B. Multisource hyperspectral and LiDAR data fusion for urban land-use mapping based on a modified two-branch convolutional neural network. ISPRS International Journal of Geo-Information 2019, 8, 28.
  179. Mohla, S.; Pande, S.; Banerjee, B.; Chaudhuri, S. Fusatnet: Dual attention based spectrospatial multimodal fusion network for hyperspectral and lidar classification. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020, pp. 92–93.
  180. Li, J.; Li, C.; Xu, W.; Feng, H.; Zhao, F.; Long, H.; Meng, Y.; Chen, W.; Yang, H.; Yang, G. Fusion of optical and SAR images based on deep learning to reconstruct vegetation NDVI time series in cloud-prone regions. International Journal of Applied Earth Observation and Geoinformation 2022, 112, 102818.
  181. Khodadadzadeh, M.; Li, J.; Prasad, S.; Plaza, A. Fusion of hyperspectral and LiDAR remote sensing data using multiple feature learning. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 2015, 8, 2971–2983.
  182. Heckel, K.; Urban, M.; Schratz, P.; Mahecha, M.D.; Schmullius, C. Predicting forest cover in distinct ecosystems: The potential of multi-source Sentinel-1 and-2 data fusion. Remote Sensing 2020, 12, 302.
  183. Chen, B.; Ye, Q.; Zhou, X.; Song, Z.; Ren, Y. Aerosol classification under non-clear sky conditions based on geostationary satellite FY-4A and machine learning models. Atmospheric Environment 2024, 339, 120891.
Table 1. Comparison of FFNN model performance with CALIOP and ADP dust detection along CALIOP nadir over land and ocean.
Table 1. Comparison of FFNN model performance with CALIOP and ADP dust detection along CALIOP nadir over land and ocean.
Aerosol Product Land (%) Ocean (%)
FFNN 90.7 38.1
CALIOP 33.6 8.3
ADP 39.3 24.0 1
Table 2. Evaluation of FFNN dust-aerosol detection on CALIOP-track over land and ocean (16 March and 12 August, 2014).
Table 2. Evaluation of FFNN dust-aerosol detection on CALIOP-track over land and ocean (16 March and 12 August, 2014).
Aerosol Product Day of Year Land (% Ocean (%) Land + Ocean (%)
FFNN 075 65.3 7.3 30.07
FFNN 224 53.73 10.51 29.12
CALIOP 075 60.27 8.3 28.7
CALIOP 224 53.92 19.76 34.47 1
Table 3. Evaluation of FFNN dust-aerosol detection off CALIOP-track over land and ocean (16 March and 12 August, 2014).
Table 3. Evaluation of FFNN dust-aerosol detection off CALIOP-track over land and ocean (16 March and 12 August, 2014).
Aerosol Product Day of Year Land (% Ocean (%) Land + Ocean (%)
FFNN 075 55.47 9.21 18
FFNN 224 51.34 12.32 25.56 1
Table 4. Features extracted from each Modality.
Table 4. Features extracted from each Modality.
Sensor (Image) Feature Feature Total
Landsat 5 TM * SR and variance on bands 1-5 and 7, 52
NDVI
ALOS-1/PALSAR Intensity, 132
Polarimetry,
Interferometry,
texture
LVIS (LGE) rh25, 4
rh50,
rh75,
rh100
Table 5. Overall Classification Accuracy Statistics for Each Model Group.
Table 5. Overall Classification Accuracy Statistics for Each Model Group.
Sensor(s) (Model Group) Mean (%) Standard deviation Range (%)
One Sensor 61.1 9.1 [74, 48]
Two-Sensor 71 4.9 [81, 64]
Three-Sensor 78 3.6 [83, 74] 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated