Preprint
Article

This version is not peer-reviewed.

A Recurrent Neural Network for Forecasting Fuel Moisture Content with Inputs from Numerical Weather Models

Submitted:

04 November 2025

Posted:

05 November 2025

You are already at the latest version

Abstract
This paper proposes a recurrent neural network (RNN) model of dead 10-h fuel moisture content (FMC) for real-time forecasting. Weather inputs to the RNN are forecasts from the High-Resolution Rapid Refresh (HRRR), a numerical weather model. Geographic predictors include longitude, latitude, and elevation. Forecast accuracy is estimated in a~study that utilizes a~spatiotemporal cross-validation scheme. The RNN is trained on HRRR forecasts and observed FMC from weather station sensors within the Rocky Mountain region in 2023, then used to forecast FMC at new locations for all of 2024. The forecasts are compared to observed data from FMC sensors that were not included in training. The accuracy of the RNN is compared to several common baseline methods, including a~physics-based ordinary differential equation, an XGBoost machine learning model, and an hourly climatology. The RNN shows substantial forecasting accuracy improvements over the baseline methods.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

Wildfires impact human health and safety in many ways. In addition to direct damage done by the blazes, wildfires emit toxic gases that harm air quality [1]. Wildfires also contribute to erosion, flooding, and cause harmful deposits into waterways, thus impairing water quality [2]. There has been a substantial increase in the scope and severity of wildfires in the Western United States since the early 1980s [3]. More regions are at risk for wildfires, and the total area burned has grown [4]. In Colorado, the three largest fires in the state’s history were all in 2020 [5]. The most destructive fire in Colorado’s history was the Marshall Fire in December of 2021, which spread rapidly in dry fine fuels [6].
Researchers study wildfires to make predictions about which areas are prone to wildfire danger and how active fires will behave. WRF-SFIRE is a wildfire modeling platform that features a coupled atmosphere and fire spread model [7]. This paper is part of a broader research effort to improve the accuracy of wildfire simulations in WRF-SFIRE. Wildfire models incorporate many environmental variables, such as the weather, topological characteristics, and characteristics of the fuels that feed wildfires. Fuels are typically in the form of living or dead plant material. Some important fuel characteristics include the physical dimensions of the fuels, the density and composition of fuel layers, and the moisture content of fuels.
Fuel moisture content (FMC) is a measure of water content in burnable fuels [e.g., [8]. Dry fuels burn more quickly, and wetter fuels burn more slowly or may even completely stifle the spread of wildfire. FMC is a key input to wildfire models and is used to identify dry areas prone to wildfire spread for the purposes of issuing warnings and directing mitigation efforts. FMC is defined as the ratio of the weight of a fuel to the dry weight of the fuel, or the total mass of burnable woody material when all of the moisture has been removed.
FMC ( % ) = Fuel Weight Dry Weight Dry Weight · 100 %
Traditionally, FMC was measured by manually collecting samples, weighing their initial wet weight, drying the samples out in ovens, measuring the resulting dry weight, and then calculating FMC with Equation 1. Manual sampling is still used today, particularly for living plant material, but this process is slow and labor intensive. Today, automated sensors provide real-time observations of FMC from across the continental United States (CONUS) [9].
Dead fuels are of particular interest to wildfire researchers, since dead plant material tends to be much drier and burn more readily than living plant material. The moisture content of dead plant material changes quickly relative to living plant material in response to changes in atmospheric conditions. Dead fuels can be a variety of sizes and compositions, but researchers have defined idealized fuel types to make modeling tractable. Dead fuels are divided into classes based on how quickly the material is expected to change in response to changes in atmospheric conditions. The fuel classes are defined by a threshold lag-time, or the expected number of hours needed for a fuel to change its moisture content by 1 e 1 ( 63 % ) in response to an atmospheric change [e.g., p. 216 [10]. These include 1-h, 10-h, 100-h, and 1,000-h fuels, where “h” is “hour” [8]. Remote Automated Weather Stations (RAWS) provide real-time data from standardized sensors for 10-h dead FMC [9]. RAWS are ground-based weather stations with scientific instruments for acquiring weather data, operated by the National Interagency Fire Center [9] and other public and private organizations [11]. Many RAWS have an FMC sensor, which consists of a 1/2-inch pine dowel with a probe that measures moisture content in the woody interior [12,13]. Weather stations with FMC sensors are maintained by many different private and public organizations, and the data is aggregated and provided by Synoptic Data [11]. We will use the term RAWS to refer to any weather station in the Synoptic provider network with available FMC data.
This study will focus on building models to forecast 10-h dead FMC. The term “FMC” in this paper will from now on refer to dead 10-h FMC. Figure 1 shows a sample time series of FMC observations from RAWS “CHAC2”, located in Southwest Colorado. The observed FMC displays a cyclical pattern with a period of 24 hours indicative of diurnal cycles.
Mathematical models of FMC are used to make predictions at unobserved locations in the future. Wildfire simulations can be initialized with forecasts from FMC models [e.g., [7]. The FMC forecasts can also be used to develop fire weather risk warnings. Forecasts of FMC have been used within fire danger rating systems and wildfire spread models for decades [e.g., [14]. The Canadian Forest Fire Danger Rating System estimated fine fuel moisture content using daily weather observations of rainfall and equilibrium moisture content, along with the previous day’s moisture estimate [15]. Physics-based methods attempt to directly represent the physical processes of energy and moisture transfer, and such models have been the dominant way to predict FMC in publicly available wildfire modeling projects. Nelson Jr. [16] developed a set of physical equations to model FMC as part of the National Fire Danger Rating System. The model represents a stick as a 1-dimensional system, where moisture and energy exchange act radially towards the center of the stick, with various boundary conditions accounting for exchange of moisture and energy between the stick surface and the environment. The model receives daily weather inputs of air temperature, RH, accumulated rainfall, and downward shortwave solar radiation. Van der Kamp et al. [17] model a stick as two homogeneous layers. The outer layer of the stick exchanges thermal energy and moisture with the environment, and the inner layer exchanges thermal energy and moisture with the outer layer. They added other physical considerations of the system, such as the orientation of the stick to the position of the sun, and added wind speed as input. Both physics-based models were calibrated to manual field observations of FMC. The FMC measurements were made on standardized pine dowels, the same as those used for real-time measurements of the FMC from RAWS [18].
Mandel et al. [7] model a fuel element as homogeneous and use an ordinary differential equation (ODE) implementing a version of the classical timelag model, in which the rate of change in FMC is proportional to its difference from the equilibrium moisture content. Vejmelka et al. [19] added data assimilation of RAWS to the model. Real-time measurements from RAWS were extended by a spatial regression to grid nodes, then assimilated into a simple ODE timelag model using the augmented Kalman Filter on each grid node. In this paper, we will refer to that model as the “ODE+KF” model. A version of the ODE+KF model is currently used within the WRF-SFIRE wildfire workflow [20] to provide initial FMC values and an equilibrium bias correction as gridded arrays. The SFIRE component then runs the ODE model for FMC with atmospheric inputs from the WRF component, but without further RAWS data assimlation. The mathematical details of the ODE+KF model will be presented in Section 2.5. The purpose of this project is to provide a standalone FMC spatio-temporal forecast, and to potentially replace and enhance the ODE+KF model in the WRF-SFIRE system in the future.
In recent years, machine learning (ML) models of FMC built entirely from data have gained in popularity. Fan and He [21] used an LSTM to model the FMC with the output of the Van der Kamp et al. [17] physics-based model as input. The model was tested on the original data used by Van der Kamp et al. [17], which consisted of two physical locations in British Columbia over five months in 2014 with atmospheric input from ground-based stations. They concluded that the LSTM model was more accurate than the purely physics-based model, as well as several ML baselines. Kang et al. [22] use an LSTM to predict the monthly average live FMC across California using daily high and low weather variables as input. Schreck et al. [23] tested various ML architectures using FMC observations from RAWS and weather inputs from the High-Resolution Rapid Refresh (HRRR) weather model and VIIRS satellite reflectance band data. They evaluated the precision of different ML models in a study that spanned several years of data across the continental United States (CONUS). The ML architectures they tested were “static”, in the sense that there is no explicit recurrence relationship between the FMC at past times and the FMC at future times. They concluded that “XGBoost” was the most accurate, and predictions from a version of this model are publicly available in a product offered by the National Center for Atmospheric Research (NCAR) [24]. This product provides historical reconstructions of FMC and near real-time predictions. An implementation of an XGBoost model will be used as a baseline method for comparison, with details described in Section 2.5. However, this is a custom implementation with a smaller set of predictors that is not the same as the NCAR product.
This paper builds on the work of Mandel et al. [25], who used a Recurrent Neural Network (RNN) to forecast FMC. In that study, the analytical solution of the ODE model was exactly reproduced within a single RNN cell by manually pretraining the weights of the network. Thus, the RNN was shown to be capable of directly encoding the solutions to a physics-based model of FMC.
The goal of this study is to develop an RNN model to forecast FMC at arbitrary locations in the United States from weather prediction data. The RNN is trained using atmospheric inputs from HRRR and FMC observations from RAWS. Its performance is evaluated against several baseline methods, including an XGBoost model, a climatology method, and the existing FMC model based on ordinary differential equations with assimilation of RAWS FMC data [19,26], currently used in the WRF-SFIRE workflow [20] (see Sect. Section 2.5.1). Ultimately, the RNN is intended to replace this model.

2. Materials and Methods

This paper focuses on the development of an RNN architecture and the estimation of the forecast accuracy at unobserved locations. As a baseline of comparison for the RNN, we use a physics-based model, a non-recurrent ML model, and a climatology method. In all these models, FMC is considered to be a single continuous value for the idealized 10-h fuel that corresponds to the sensors used by RAWS. We evaluate model accuracy by comparing the FMC forecast to observed values with a cross-validation procedure described in Section 2.6.

2.1. Data

2.1.1. FMC Observations from RAWS

RAWS provide real-time, hourly observations of dead 10-h FMC, which can be retrieved through the Synoptic Weather API via Python software [27]. All RAWS FMC data in the study region from 2023 to 2024 were visually examined to identify suspect data from broken or faulty sensors. Automated filters were used to identify instances where the FMC is outside of physically reasonable ranges of 1% to 90%. This is a conservative threshold compared to operational implementations of the Nelson model [16], which has a built-in maximum FMC level of 60% [p. 205, [18]. In addition, automated filters were used to identify long stretches of data where the FMC is constant, which is indicative of a broken sensor. Missing observations were interpolated linearly if the stretch of missing data is 3 hours or less. Any longer stretches of missing data were kept as missing with no further imputation.

2.1.2. Predictors of FMC

For the ML models used in this project, we select a set of predictors of FMC that include atmospheric weather data as well as geographic features of the physical RAWS locations. The set of predictors was chosen based on the typical inputs to various physics-based models of FMC. The final set of predictors is shown in Table 1. Individual RAWS report spatial coordinates including longitude, latitude, and elevation above sea level in meters. These three spatial characteristics are used as predictors in the RNN and XGBoost models.
The weather variables were collected from the HRRR, a product developed by NOAA [28], using the python software package Herbie [29]. The HRRR model provides real-time, hourly forecasts of many weather variables on a 3km resolution grid across CONUS. All weather variables used in this project were extracted from HRRR version 4, which was introduced in December 2020 [28]. Every 6 hours, the model provides forecasts 48 hours into the future. The HRRR model can therefore be used within a real-time forecasting tool for FMC. Schreck et al. [23] used the HRRR in their FMC modeling project, and the HRRR is used initialize simulations in the WRF-SFIRE project. We retrieve predictors at a given time from the 3rd HRRR forecast hour for the corresponding time 3 hours in the past. The 3rd forecast hour is used for two reasons. First, certain modeled variables, such as wind, tend to be very smooth spatially at the zero hour of the model, and these variables begin to develop a more realistic structure by the third hour. Second, rainfall is one of the key predictors of FMC and the zero hour of HRRR forecasts always has zero precipitation accumulated. We use rainfall in units of accumulated millimeters per hour, since the ODE+KF uses rainfall data with these units. Also, there are ground-based rainfall capture sensors that update hourly on some RAWS, which can be used as a quality control check for HRRR model output for rain. Calibration efforts within the WRF-SFIRE project settled on using the 3rd HRRR forecast hour, when available, instead of the zero hour model. The HRRR weather variables at a given time are interpolated linearly to the location of the RAWS and joined to the FMC data by location and time.
The air temperature and RH at 2m height from HRRR are used to calculate wetting and drying equilibrium moisture content ( E d and E w ), in units of percent FMC (or unitless depending on convention) [Eqs. 4 and 5, [14]. Downward shortwave solar radiative flux (DSWRF) at the surface is collected from HRRR, in units W m 2 . DSWRF is a type of solar radiation used by several physics-based models of FMC, including Nelson Jr. [16] and Van der Kamp et al. [17]. Some RAWS have pyranometers which measure DSWRF and can be used as a quality control check. The rain in terms of m m h r 1 is calculated from HRRR by subtracting the accumulated precipitation at the 3rd HRRR forecast hour with the accumulated precipitation from the 2nd forecast hour. Some RAWS have rain capture buckets that produce observations hourly accumulated rainfall, which be used as a quality control check for the hourly rain forecasts. Finally, wind speed at 10 meters is collected from HRRR, in units of m s 1 . Some RAWS also have wind sensors positioned at a sampling height of 20 feet, which can be used for validation of the HRRR wind speed data.
Lastly, we utilize two predictors derived from the time of the observations. The hour of the day (0-23 UTC) and the Julian day of the year (0-365 or 366 for leap years) are used as predictors in the ML models. The hour of the day is used to help the ML models capture diurnal cycles in FMC. FMC is the highest at night just after when the humidity is highest and temperature is lowest. The relative position of the sun is a physical driver of the process, so the hour of the day is used as a proxy for the sun position. The day of the year is used to help the ML models capture seasonal affects.
All predictor data was scaled before being input into the ML models. Scaling inputs can lead to improved performance for neural networks. Standard scaling was used, which consists of subtracting the sample mean and dividing by the sample standard deviation for each predictor variable. The mean and standard deviation used for scaling was calculated only on training sets and then applied to testing data. See Section 2.6 for more information on the spatiotemporal cross-validation method used in this paper. The response variable of FMC was left unscaled.

2.1.3. Data Collection Summary

We collected all available FMC data for 2023 through 2024 within the Rocky Mountain Geographic Area Coordination Center (GACC). The spatial bounding box for this region is from Latitude 37 N to Latitude 46 N, and from Longitude -111 W to Longitude -95 W. This region includes parts of Colorado, Wyoming, Nebraska, Kansas, Montana, and South Dakota. There were 161 total RAWS with valid FMC data in this period and spatial domain, and 151 total RAWS with valid data in the forecasting period of 2024. Figure 2 shows the location of the RAWS within the Rocky Mountain GACC with valid FMC data from 2023-2024. This region is particularly prone to wildfire danger and numerous federal, state, and local agencies are involved with wildfire management. The atmospheric weather predictors were chosen based on physical considerations and based on inputs to other models of FMC in the scientific literature. A complete list of all data variables used in this study is listed in Table 1.

2.2. Recurrent Neural Networks

Recurrent neural networks (RNNs) are a class of neural network architectures that are specific to processing sequence data, where predictors and outcomes are ordered in time. An RNN cell is a basic processing unit where the output at a given time is a function of the input and the output of the cell at previous times [30][p. 498-501]. The simple RNN cell takes as input the set of predictors at time t, denoted X t , as well as the output of the recurrent connection at the previous time step, which we refer to in this paper as the “recurrent state”. The right-hand side of Figure 3 shows the computational flow in a single recurrent unit. The input X t and the recurrent state h t 1 are combined in a linear combination with fixed weights W x , W h and b. This linear combination is modified by an activation function, denoted σ , to produce the cell output, which in the simple RNN cell is stored as the new hidden state.
h t = σ ( W x X t + W h h t 1 + b )
The weight vector W x is the same length as X t , which represents input. The weight W h and the bias b are scalar values. For the simple RNN cell, the output is the same as the new recurrent state.
Figure 4 shows the recurrent unit “unrolled” in time, where the same computational unit is applied recursively. Over time, the recurrent state is modified, but the same set of weights and biases are used at each time step.
Multiple recurrent cells are typically stacked into a “recurrent layer”. One or more recurrent layers are used in conjunction with other neural network layers, such as the standard fully-connected “dense” layer. We will refer to an “RNN” as any neural network architecture with one or more recurrent layers.
A popular type of recurrent model architecture is Long Short-Term Memory (LSTM) (Figure 5), which was originally developed to address certain issues with the simple RNN cell [31]. Like the simple RNN cell, the LSTM cell maintains a hidden state, which is the output of the cell at the previous time step. In addition to the hidden state, an LSTM maintains a “cell state”, denoted c t , which incorporates information about the output of the cell over more time steps in the past than just the previous time step. At time t, the hidden state h t 1 and the predictors X t are used to modify the cell state through a series of computational gates. The resulting cell state c t can be viewed as a compromise between the old information as represented by the cell state c t 1 and the new information as represented by h t 1 and X t . The final output of the cell at time t is a combination of h t 1 , X t , and c t . This is stored as the new hidden state at time t, or h t . LSTM cells are typically stacked into an LSTM layer that is combined with other neural network layers to form a more complicated model architecture. We will refer to a “LSTM” as any neural network that contains one or more LSTM layers. The computational graph in Figure 4 still applies to the LSTM cell, with the modification that the state passed recursively in time to the unit consists of both the hidden state and the cell state, and the output of the unit is the hidden state. For the LSTM cell, the combination of the hidden state and the cell state is the recurrent state.
In this paper, an RNN is trained to map time series of weather variables to time series of FMC, processing one time step at a time. The weather inputs at time t not only result in a predicted FMC at the same time t, but also modify the recurrent state, and thus affect the predicted FMC at future times t + 1 , t + 2 ,... Since numerical weather forecasts are used as inputs, it is assumed that the weather is known in the future.
Neural networks contain many parameters that are fitted using a training set by an attempt to minimize a loss function of the model when compared to the observed data. With RNNs, the loss can be computed over time, so the model can learn to minimize the prediction error over a forecast window. During training, a fixed number of time steps, called the sequence length, is used to calculate the loss.
We use the mean squared error (MSE) as a loss function, which is standard in machine learning for a continuous response. Suppose that an input sequence is of length T. The predicted FMC at time t is denoted y ^ t , and the observed FMC at time t is y t . The loss for a single sequence is
MSE = 1 T t = 1 T ( y t y ^ t ) 2 ,
Loss is calculated by comparing the predicted time series of FMC to the observed values (Figure 6).

2.3. RNN Training, Prediction, and Tuning

We trained a recurrent neural network (RNN) with gradient descent using the Adam optimizer. Inputs combine time-varying meteorological variables with static spatial features (longitude, latitude, elevation). The target variable is the 10-h dead fuel moisture content (FMC) measured at Remote Automated Weather Station (RAWS) sites.
Training data are arranged as a tensor of shape (batch size, sequence length, features). Here, features is the number of unique inputs. Following Zhang et al. [32][Sect. 9.7], we use truncated backpropagation through time (BPTT). A single sample consists of a tensor of shape (sequence length, features). At the start of processing a sample, the recurrent state is reset. At each time step of the sample, a one-dimensional tensor of features is input to the model, the recurrent state is updated, and a single output is generated. This process is repeated for the entire sample. The output of the network from an entire sample is a timeseries of FMC predictions of the same sequence length as the input sample, which requires setting the function argument return_sequences to true [33]. The model loss for the sample is the MSE, as seen in Equation 3, averaged over the entire sequence length. The gradient of the loss with respect to all model parameters is then calculated via BPTT. Gradients are calculated for each sample sequence in a batch and then averaged across samples [34]. Model parameters are updated after each batch. A single epoch of training is completed when all batches of samples have been processed. Early stopping halts training when validation performance no longer improves.
The static predictors are repeated at each time step. Batches include samples from multiple locations to promote generalization and learn the dependence on location. The model parameters are updated after each batch.
During training, the effective temporal depth of optimization is limited by the sequence length and batch size. Consequently, truncated BPTT is applied only within this finite window, ensuring computational stability [32][Sect. 9.7.1.2]. In contrast, during prediction, the model operates with an unconstrained sequence length and a batch size equal to the unique number of locations for prediction, receiving new weather inputs sequentially and updating its recurrent state indefinitely. The prediction process is therefore equivalent to unrolling the RNN for an arbitrary number of time steps without truncation, allowing the model to integrate meteorological history over potentially long periods. The same network parameters, trained within the finite truncated BPTT horizon, are reused for recursive forecasting by copying the learned weights from the training configuration to the stateful inference model [30][p. 512].
Neural networks have many fixed hyperparameters, such as the number of cells to use within a layer, the overall number of layers, the batch size, the learning rate, etc. It is important to keep the data used to tune hyperparameters separate from data used to estimate forecast accuracy. The metrics calculated on forecasts should represent the accuracy of the model when predicting entirely new data. The hyperparameter tuning used in this project will be described in Section 2.6.

2.4. Recurrent Neural Network for Forecasting FMC

In this paper, we use the following RNN model architecture. The input is a collection of features consisting of weather variables and geographic locations of RAWS. These inputs are fed into a single LSTM layer consisting of 64 cells with hyperbolic tangent activation function. The LSTM layer outputs a sequence of the same length as the input data, and the entire sequence is passed through the network until the output layer. The output of the LSTM layer is fed into two dense, fully connected layers with ReLU activation and 32 and 16 cells, respectively. The LSTM layer and the dense layers process each time step in the sequence independently. Finally, the output of the model is a single neuron with linear activation. Linear activation is chosen for the final output layer, since we are mapping the inputs to a continuous real number. A single neuron is used as an output layer since model predictions are generated by deploying the model point-wise at each location, and thus the FMC at a given time is considered to be a single scalar value. The final model output is a two-dimensional vector, where the first dimension is the number of unique locations and the second is the number of time steps (48 hours in our case). Other model architectures can be used to generate outputs that correspond to a spatial grid, but we chose this architecture since it is the easiest to implement and is the most flexible when predicting across various spatial domains. The left side of Figure 3 shows a diagram of the architecture of the RNN model used in this study.
The HRRR model is used to provide weather input to the RNN. HRRR provides hourly forecasts up to a maximum of 48 hours. The sequence length used to structure the data input (batch size, sequence length, features), described in Section 2.2, is set to 48. Each input to the RNN consists of a 48-hour sequence of observations, and every output of the model is a 48-hour sequence of FMC predictions at the same times as the input. The batch size is a tuned hyperparameters, and the number of features was determined through theoretical considerations described in Section 2.1.2.
In this study, we evaluate the forecast accuracy over all of year 2024 broken into 48 hour periods to align with the sequence length of training inputs. The initial hidden state and cell state of the LSTM layers are zero by default. Starting at 00:00 UTC on January 1, 2024, we input 48 hours of HRRR weather data and geographic input at a given set of test locations to generate the first set of forecasts. Then, for the next 48 hours, the initial hidden state and the cell state of the LSTM are reset. This process is repeated until all 2024 forecasts are generated. The initial recurrent state of the LSTM is reset for each 48-hour forecast period and no information is retained across distinct forecasting periods. Resetting the recurrent state in this way is done for the purposes of estimating the forecast accuracy of the model 48 hours into the future. Resetting the recurrent state is not required for this model architecture, and an operational version of the model could retain the recurrent state for arbitrarily long.
The hyperparameters for the RNN were selected with a restricted grid search, which evaluates a subset of all possible hyperparameter combinations. The combination of hyperparameters that leads to the most accurate predictions on the validation set is chosen. Here, some hyperparameters were set to their default values from the TensorFlow software project, which are generally accepted in the literature as reasonable defaults. Table 2 shows the hyperparameters used in the model that are not part of the core model architecture shown in the left-hand side of Figure 3. The hyperparameter tuning method is discussed in Section 2.6, and Appendix A has additional technical details.

2.5. Baseline Methods

2.5.1. ODE Model with Kalman Filter in WRF-SFIRE

The ODE+KF uses a first-order timelag differential equation with data assimilation [7,26]. This model uses drying and wetting equilibrium moisture content and hourly accumulated rainfall as input. Data assimilation of FMC observations from RAWS is done using the augmented Kalman filter [26]. During wildfire simulations, WRF-SFIRE receives the initial fuel state from the ODE+KF after data assimilation, and then the ODE model inside WRF-SFIRE runs in forecast mode without data assimilation to update the FMC throughout the duration of the simulation. Model calibration was performed using wildfire simulations from WRF-SFIRE, comparison to the fine fuel moisture model from the Canadian fire danger rating system [14], and comparison to RAWS in Colorado and the Pacific Northwest. Equation 4 describes the mathematical form for m ( t ) , the FMC at time t:
d d t m ( t ) = S m ( t ) T r 1 exp r 0 r ( t ) r s if r ( t ) > r 0 ( soaking in rain ) E d ( t ) m ( t ) T k if r ( t ) r 0 and m ( t ) > E d ( t ) E w ( t ) m ( t ) T k if r ( t ) r 0 and m ( t ) < E w ( t ) 0 if r ( t ) r 0 and E w ( t ) m ( t ) E d ( t )
If the amount of rainfall accumulated per hour r ( t ) is greater than the threshold value r 0 , it is assumed that the stick absorbs water directly, and it approaches a saturation moisture level S with a characteristic delay time depending on rain intensity, asymptotically T r for a very intense rain. The saturation rainfall intensity r s is when rain wetting is 1 e 1 63 % of the maximal rate. The calibration of the model resulted in a threshold rainfall intensity of r 0 = 0.05 m m h 1 , a saturation level S = 250 % , a delay time T r = 14 h , and a rain saturation intensity of r s = 8 m m h 1  [7]. These values will be referred to as hyperparameters for consistency with other models. If the rainfall r ( t ) is less than the threshold value r 0 , the stick changes moisture by equilibrating to the equilibrium moisture content of the surrounding environment. There are three distinct dynamics for the FMC in no-rain conditions. If the FMC m ( t ) is greater than the current drying equilibrium moisture content E d ( t ) , then drying dynamics occur and the FMC approaches E d ( t ) with a characteristic time-lag of T k determined by the fuel class ( T k = 10 for 10-h fuels). If m ( t ) is less than the current wetting equilibrium moisture content E w ( t ) , then wetting dynamics occur and the FMC approaches E w ( t ) with the same characteristic time-lag of T k . If m ( t ) is between E d ( t ) and E w ( t ) , there is no change in moisture content. We use the ODE+KF with the calibrated parameters described above to forecast the FMC at the locations not included for training the ML models and at times in the future of all ML training data. The ODE equations described above can be solved analytically, and a discretized version of the analytic solution is used to model FMC.
In the model by Vejmelka et al. [26], the state of the system was extended to include an equilibrium bias correction to allow the model to learn differences between the equilibrium moisture content values as determined by laboratory experiments versus what is suggested by the data. In this paper, a spinup period of 24 hours is used for the equilibrium bias correction to stabilize. The spinup period is the number of time steps where data assimilation via the Kalman Filter occurs. After 24 hours of spinup, the model is run in forecast mode for 48 hours. In forecast mode, the model receives weather inputs, but no input from observed FMC.
In Vejmelka et al. [26], a spatial regression procedure is used to interpolate FMC observations from RAWS to a computational grid, and the interpolated FMC values are used for data assimilation into the ODE+KF model running on each grid node.
Here, since we are interested in a comparison at the locations of the RAWS in the test set, we run ODE+KF at those RAWS locations, and assimilate the RAWS FMC data directly. Table 3 shows a summary of the hyperparameters used within the ODE+KF.

2.5.2. The XGBoost Static Model

The ML model used as a baseline is an implementation of the Extreme Gradient Boosting (XGBoost) technique [35], which has been used by FMC researchers in recent years [23]. The XGBoost uses ensembles of regression trees that are iteratively fit and re-weighted. XGBoost produces a nonlinear model which is a weighted average of regression trees, resulting in a piecewise-constant mapping of the predictors to the output. Our implementation of XGBoost is “static”, or nonrecurrent, meaning that the model is learning a map between the predictors and the observed FMC at a moment of time and incorporating no information about the FMC at previous times. Thus, the instantaneous relationship between the predictors and the response is learned, and observations of FMC are implicitly considered independent in time. Time-based predictors are utilized in our implementation of the model, such as the hour of the day and the day of the year, but the model is still nonrecurrent since the output at a given time step is not a function of the output at the previous time step. Hyperparameters for the XGBoost model were based on previous work by Hirschi [36], where various static ML models were tested over 2023 in the Rocky Mountain region. Table 4 shows a summary of the hyperparameters used within the XGBoost model.

2.5.3. Climatology Method

Lastly, we utilize a climatology method. In weather forecasting, climatology refers to simple statistical methods in which historical weather patterns are used to predict new values of a particular variable. We utilize a climatology method for FMC that is based on a method used by Schreck et al. [23]. Given a particular time and location, the hour of the day (0-23 UTC) and the Julian day of the year (0-365 or 366) are extracted. The FMC forecast for that hour and location is the historical average of the FMC at that exact location and the same hour of the day, for the days of the year within 15 days of the target time. In this study, we build a climatology by looking back 10 years for each forecast time and location. If there are historical observations from less than 6 unique years at a particular location for a particular time, the FMC forecast was set to missing. These configurations for the climatology method will be referred to as hyperparameters for consistency with the other methods. Table 5 shows a summary of the climatology hyperparameters used in this study.

2.6. Analysis Design

We seek to forecast FMC at arbitrary locations in a region where the training FMC data are and at future times. The model maps a time series of weather forecasts to a time series of spatial FMC forecasts based on FMC and weather data that were presented to it during training. In a sense, the model interpolates in space and extrapolates the weather-to-FMC mapping in time. We will compare the models using the root mean squared error (RMSE) at new locations and at times in the future of the times used for training. The RMSE for the FMC models is interpretable in units of percent. In all reported RMSE values in this paper, the square root is applied after any averaging operations; that is, the square root is always the final operation.
Cross-validation methods estimate the accuracy with which a model predicts unseen data [Sect. 7.10 [37]. In machine learning, samples are randomly drawn to form independent training, validation, and test sets [Sect. 7.2 [37]. The training set is used to fit the model, the validation set to tune hyperparameters and monitor overfitting, and the test set, kept separate, to estimate final predictive accuracy. In time-dependent problems, the test data should come from a period after the training data to reflect the real forecast conditions. This requirement conflicts with the common assumption that training and test samples are independent and identically distributed because temporal ordering introduces dependence. The evaluation of time-series models must therefore balance statistical independence with causal realism. In spatially-dependent problems, the test data needs to be from locations that were not used in training model parameters. If a model is used to predict at locations included in its training data, it has already learned aspects of the data structure at those locations, leading to overly optimistic accuracy estimates [23].
To estimate the forecast error of FMC models at unobserved locations, we apply a spatiotemporal cross-validation procedure. In the hyperparameter tuning step, FMC observations from 2022 are used to train a set of candidate model architectures. A random sample of 90% RAWS is used to generate the training locations and 10% to generate the testing locations. Then, the forecast accuracy of those models are compared at the testing locations over all of 2023. During the forecast period, HRRR inputs are used to generate predictions, but no FMC data from the testing time period or testing locations are used to inform predictions. The model architecture with the lowest RMSE when forecasting in 2023 was selected.
With the model architecture fixed, the model is trained from scratch on observations in 2023 with a random sample of testing locations and training locations. The trained model is used to forecast at locations not used in the training over all of 2024. The final accuracy metrics for the model are calculated by comparing the forecasts from 2024 to FMC observations at the testing locations. The training period always precedes the forecast period in time, and random sampling of locations accounts for spatial uncertainty. A schematic diagram of this cross-validation method is shown in Figure 7, with the training period of 2023 and the forecast period of 2024.
The ODE+KF and the climatology methods do not fit directly into the training and testing paradigm used in ML. The ODE+KF is run in forecast mode (i.e. without data assimilation) over all of 2024 at the test locations in 48-hour increments to mimic the initialization with a spinup period. The ODE+KF utilizes 24 hours of data prior to the start of the forecast as a spinup period (i.e. with data assimilation). Each iteration of 24-hour spinup plus 48-hour forecasting were independent. The climatology method produces forecasts by retrieving historical data for each time in 2024 and each testing location. Therefore, the climatology method does not utilize spatial holdout like the ML models, so the accuracy metrics for the climatology method are more optimistic than for the ML methods.
To account for the variability in the random spatial samples and randomness due to model initialization, the training in 2023 and forecasting in 2024 was repeated 500 times with different random seeds. So, 500 random samples of RAWS were used as training and test splits and 500 different initial weights were used for the ML models. These will be referred to as replications from now on. The replications are used to construct uncertainty bounds on the final accuracy metrics. There were 151 RAWS with valid data in the forecast period of 2024, and each replication had 16 RAWS in the test set, and each RAWS was included in the test set on average 53.0 times in distinct replications.
Within each replication, the forecast residuals are calculated by subtracting the predicted FMC from the observed FMC for each model and at each time. The residual shows whether the forecast was too high or too low and can be positive or negative. We then calculate the squared residual, which is always positive. Within a replication for a given model, squared residuals are calculated across a set of test set locations and a set of forecast times. The per-replication MSE is calculated by averaging the squared residuals over all test locations and test times for a given replication. Then, the overall RMSE is calculated by averaging all the per-replication RMSE values and taking the square root. Uncertainty bounds are calculated from the square root of one standard deviation of the per-replication MSE values. The overall model bias is calculated in the same was using the raw residuals rather than the squared residuals. This metric is used to analyze whether the forecasts are systematically too high or too low. The overall model RMSE and bias are calculated by averaging over all locations and times in the test set. Both the RMSE and bias are interpretable in units of percent FMC. Equation 6 shows the mathematical form of the RMSE estimate used to quantify forecast error across the entire study region and all times in 2024, using mathematical definitions in Table 6. Equation 8 shows the mathematical form of the bias estimate. Equation 7 shows how the standard deviation of the RMSE is calculated, and the standard deviation of the bias estimate is calculated in an analogous way.
We calculate a per-replication MSE by averaging the squared error over all times and locations,
Per - Replication MSE MSE r = 1 N · T i = 1 N t = 1 T ( y i , t y ^ r , i , t ) 2
The overall RMSE is calculated by averaging the per-replication RMSE over all replications,
Overall Mean RMSE MSE ¯ = 1 R r = 1 R MSE r
We estimate the uncertainty in the overall RMSE by calculating the square root of the standard deviation of the MSE across all replications,
Overall Std . of RMSE 1 R 1 r = 1 R MSE r MSE ¯ 2
We calculate the bias of the models by averaging the error across all times and replications, and the associated standard deviation of the bias is calculated in an analogous way to the standard deviation of the MSE,
Overall Bias Bias ¯ = 1 R · N · T r = 1 R i = 1 N t = 1 T ( y i , t y ^ r , i , t )
We further analyze the forecast error by location and by hour of the day. For the RMSE by hour of the day, model errors are averaged over location, replication number, and days of the year to produce a single RMSE estimate for each hour to analyze how the forecast error changes throughout the diurnal FMC cycle. Further, this analyzes how the accuracy changes as the forecast is run longer into the future. To analyze the spatial variability of the forecast error, the RMSE and bias are calculated by averaging over all times and replications in the test set, producing accuracy metrics for each RAWS. Again, the replications are used to construct uncertainty bounds. Within a single replication we get an estimate of the forecast accuracy for 16 of the total RAWS. After many replications, we end up with estimated RMSE and bias, and uncertainty bounds, for every RAWS with data availability in 2024. The standard deviations of these RMSE estimates across replications are calculated in an analogous way to Equation 7.

3. Results

The accuracy of the RNN model and three baseline models was evaluated when predicting values at unobserved locations across all of 2024. The models were evaluated on the same randomly selected test locations across 500 random samples of locations. Table 7 shows the RMSE when averaged over all times, locations, and replications. The bounds on the estimated metrics are plus or minus one standard deviation of the given metric across all replications. The RNN achieved the lowest forecast error compared to the baseline methods as indicated by the RMSE. Further, the RMSE of the RNN is well below the uncertainty bounds of the other methods. Thus, there is strong statistical evidence that the RMSE of the RNN is lower than the other methods for the analysis period of 2024 in the Rocky Mountain region.
A bias of zero would indicate that a model is on average over-predicting FMC just as often as it is under-predicting. Based on the bias metrics and the associated uncertainty, the RNN show signs of being unbiased overall since zero lies within the uncertainty bounds. We examine the RNN errors more closely by stratifying the FMC values into low (0-10%), medium (10-20%), and high (+20%). Table 8 shows the mean bias and standard deviation for these thresholds, calculated over all 500 replications. Figure 8 shows the distribution of observed and predicted FMC.
For low moisture fuels, the RNN is slightly over-predicting, as indicated by a negative bias in Table 8. For high moisture fuels, the RNN is slightly under-predicting, as indicated by a positive bias. The distributions in Figure 8 reflect this, showing that there is less density of predicted FMC values in the low and high regions. In practice, these types of model biases could lead to underestimating the wildfire rate of spread in very dry fuel conditions, or overestimating the wildfire rate of spread in very wet fuel conditions.
Figure 9 shows the forecast error for individual RAWS. The RMSE is averaged over all times and replications to produce an estimated RMSE for each RAWS used in the forecasting period of 2024. On average, a RAWS was included in the test set in 53.0 distinct replications, so the metrics displayed in Figure 9 are averaged over roughly 53 different randomly selected training sets and initial RNN weights.
We evaluate the forecast error of the RNN and baseline models by the hour of the day (0-23 UTC). FMC varies throughout the day in a diurnal pattern, as shown in Figure 1, so it is important to understand how this affects model accuracy. Figure 10 shows the forecast RMSE averaged across all locations, replications, and days of the year. All models show a relative peak in forecast RMSE from around 14:00 UTC, which corresponds to either 7:00am MST or 8:00am MDT. This is the time of the day when FMC tends to be the highest, since it is slightly after the lowest typical temperatures and highest typical RH of the day, and dew is often forming directly on the surface of the fuels at these times. In this study, the first hour of forecasting was fixed to be midnight 00:00 UTC, so we wish to quantify how the model error changes as the forecast is extended in time. The ODE+KF shows signs of recursive error accumulation, where the modeling error grows as the forecast continues to be extended in time. The XGBoost and climatology methods are not recursive, so there is no accumulation of forecast error. The RNN calculates loss by comparing the predicted FMC to a 48-hour stretch of observed FMC data, so the network is learning to minimize the error across this entire forecasting window. Thus, we do not observe recursive error accumulation in the RNN.

4. Discussion

In this study, we compared an RNN model of dead 10-h fuel moisture content with weather inputs from HRRR to several baseline methods. The accuracy of the models was compared when forecasting 48 hours into the future and at unobserved locations for all of 2024 in the Rocky Mountain region. The RNN was the most accurate model of FMC compared to the baseline methods, namely a physics-based model, a non-recursive ML model, and a climatology method, as shown in Section 3.
Forecasts from the RNN could be used within wildfire simulations that need to be run for multiple days. Additionally, forecasts from the RNN could be used to inform wildfire risk indices that provide warnings for areas prone to rapid wildfire spread.
The RNN architecture used in this study was based on a restricted hyperparameter search scheme in the Rocky Mountain region. Spatial features were used as input to the RNN in this study, including elevation and latitude and longitude coordinates. In other geographic regions or on a national scale the choice of hyperparameters may be different. This will be evaluated in the future.
Future work will consider adding convolutional layers in combination with LSTM layers. Additionally, attention layers and transformers are alternatives to RNNs which have grown in popularity in recent years. We conducted minimal exploration with these methods and concluded that RNNs were preferable, but a full comparison of those model architectures should be considered in future work.
Future work will apply RNNs to other dead fuel classes, including 1-h and 100-h fuels. This modeling project only considered 10-h dead FMC since the vast majority of available data is for this fuel class. The RNN modeling approach in this paper could be adapted to model dead fuels of other sizes. Additionally, the RNN considered in this study will be tested using custom loss functions that attempt to account for the nonlinear relationship between FMC and wildfire rate of spread [36]. Future research will examine additional predictors of FMC within the RNN, including satellite data products such as VIIRS reflectance bands and CYGNSS soil moisture. Finally, the RNN will be integrated into the WRF-SFIRE modeling project so that the FMC model can be tested on coupled wildfire simulations.

Author Contributions

Jonathon Hirschi primary author, Jan Mandel initial RNN work and editing the manuscript, Angel FMC data collection and theoretical support, Kyle Hilburn general science support.

Funding

The research presented in this paper was partially funded by NASA grants 80NSSC23K1118, 80NSSC22K1405, 80NSSC23K1344, and 80NSSC25K7276. Computing resources were provided by the University of Colorado Denver supercomputing cluster Alderaan, partially funded by NSF Campus Cyberinfrastructure grant OAC-2019089.

Data Availability Statement

The code associated with this project was built in Python, and is publicly available at the GitHub repository https://github.com/openwfm/ml_fmda. This repository contains materials needed to reproduce the data set used in the analysis, the forecast accuracy metrics, talbes, and images. The data used in this project is derived from public domain resources, including the Amazon AWS data set of historical HRRR forecasts (https://registry.opendata.aws/noaa-hrrr-pds/) and the historical RAWS data from Synoptic, accessed through the software package SynopticPy (https://github.com/blaylockbk/SynopticPy). The data used in the project was collected, formatted, and stored using the University of Colorado Denver supercomputing cluster Alderaan. The exact data is available upon request due to privacy restrictions of the university computing system and the substantial storage space needed for this particular project.

Acknowledgments

A portion of this work used code generously provided by Brian Blaylock’s Herbie Python packages Herbie (Version 2024.8.0) (https://github.com/blaylockbk/SynopticPy) and SynopticPy (https://github.com/blaylockbk/SynopticPy)

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

    The following abbreviations are used in this manuscript:
FMC Fuel Moisture Content
ML Machine Learning
RNN Recurrent neural network
LSTM Long Short-term Memory
MSE Mean squared error

Appendix A. RNN Hyperparameter Tuning

Machine learning models have a set of fixed parameters that relate to model architecture or aspects of the fitting procedure, called “hyperparameters”. These hyperparameters are fixed in the sense that they are not updated during the normal training process. Instead, they are “tuned” by repeatedly running the model with various sets of hyperparameters and comparing the resulting predictive accuracy. Neural networks have many hyperparameters, which can greatly affect the training process and final model accuracy. There are many methods for hyperparameter tuning. Grid search is a rigorous method of hyperparameter tuning where all possible combinations hyperparameters are compared within their given ranges. For neural networks, a full grid search usually results in a massive space of possible hyperparameter combinations that is not computationally feasible to search exhaustively. We utilized a restricted grid search, where a subset of the full number of hyperparameters was systematically explored, and other hyperparameters were fixed to theoretically or empirically justified defaults.
Hyperparameter tuning was performed by training various model architectures on data from 2022 and evaluating forecast accuracy in 2023. Different model architectures and optimization parameters were evaluated on each of these forecast periods, and the overall model accuracy in terms of RMSE was averaged across all locations and times to compare models.
The model architecture with the lowest forecast RMSE in 2023 was selected. The hyperparameters were fixed from this point on. The model was trained from scratch in 2023 and used to forecast the entire year of 2024. Randomization of the RAWS used in training versus testing ensures spatial independence. The temporal ordering of training and testing ensures causal realism, and ensures that the FMC observations from 2024 used to calculate the accuracy metrics presented in Section 3 were not used in any way to conduct hyperparameter tuning.
We used a two-step hyperparameter tuning procedure. First, we tune the model architecture, including the type and number of hidden layers and the number of cell units per layer. The model architecture with the minimum average test RMSE from this step was used in the second stage of the hyperparameter tuning. In the second stage, hyperparameters related to optimization of trainable parameters were tuned, including batch size and learning rate. Using two stages in this way greatly reduces the number of possible model combinations and speeds up computation.
For tuning the model architecture, we use the following constraints. We only consider LSTM and dense hidden layers. The LSTM layer(s) always precede the dense layers. We consider one, two, or three recurrent LSTM layers, and we consider zero, one, two, or three dense layers. For each layer, we consider a grid of numbers of units: 8, 16, 32, 128. Further, we impose a “funnel” like structure where the number of units always decreases or stays the same as you move across the layers from input to output. So, a layer with 64 units could feed into a 64, 32, or 16 unit layer, but a 16 unit layer can only feed into another 16 unit layer. For the optimization related parameters, we did a grid search for learning rate (0.05, 0.01, 0.005, 0.001, 0.0005, 0.0001) and batch size (16, 32, 64, 128). Candidate number of units and batch size are chosen to be powers of two, which in some computational contexts can lead to more efficient memory usage since binary computers usually have computational limits associated with powers of two.
Other hyperparameters were fixed across all models. These include, but are not limited to, the following. The list of predictors was fixed to be equilibrium moisture content, hourly rainfall, elevation, wind speed, downward shortwave radiation, and the longitude and latitude coordinates of the RAWS. The sequence length, or number of time steps, was set to 48 because the goal is to predict 48 hours into the future based on the forecast availability in the HRRR model. The activation functions were fixed to commonly accepted defaults for the layers. These are hyperbolic tangent (tanh) for LSTM activation, sigmoid for LSTM recurrent activation, and rectified linear (ReLu) for dense layers. A dropout rate of 15% was used for dense layers and 20% for the recurrent layers. Dropout is a common regularization technique in which in every training batch processed by the network, a random sample of given % of the network weights is set to zero. The “Adam” optimizer was used in the training to minimize the loss function. Finally, we used early stopping with a patience of 5. This means that training is stopped if the accuracy on a validation set did not decrease for 5 epochs in a row. Using early stopping allows us to set the number of training epochs to a large number and avoid the need to tune this parameter. In practice, using 100 epochs with early stopping patience equal to 5 resulted in training where none of the iterations reached the maximum of 100, and training was typically halted after 5 to 10 epochs.

References

  1. Organization, W.H. Wildfires. https://www.who.int/health-topics/wildfires, 2024. Accessed: December 26, 2024.
  2. Water Resources Mission Area, U.S. Geological Survey. Water Quality After Wildfire. https://www.usgs.gov/mission-areas/water-resources/science/water-quality-after-wildfire, 2024. Accessed: December 26, 2024.
  3. Cartier, K. U.S. Fires Quadrupled in Size, Tripled in Frequency in 20 Years. Eos (Washington, D.C.) 2022, 103. https://eos.org/articles/u-s-fires-quadrupled-in-size-tripled-in-frequency-in-20-years.
  4. Gulev, I. Land-Climate Interactions. In IPCC Sixth Assessment Report; IPCC, 2021; chapter 2, pp. 133–168. https://www.ipcc.ch/site/assets/uploads/sites/4/2021/02/05_Chapter-2-V5.pdf, Accessed: July 3, 2024.
  5. CDFPC. Historical Wildfire Information. https://dfpc.colorado.gov/sections/wildfire-information-center/historical-wildfire-information, 2024. Accessed: December 26, 2024.
  6. Juliano, T.; Lareau, N.; Frediani, M.; Shamsaei, K.; Eghdami, M.; Kosiba, K.; Wurman, J.; DeCastro, A.; Kosovic, B.; Ebrahimian, H. Toward a Better Understanding of Wildfire Behavior in the Wildland-Urban Interface: A Case Study of the 2021 Marshall Fire. Geophysical Research Letters 2023, 50. [Google Scholar] [CrossRef]
  7. Mandel, J.; Amram, S.; Beezley, J.D.; Kelman, G.; Kochanski, A.K.; Kondratenko, V.Y.; Lynn, B.H.; Regev, B.; Vejmelka, M. Recent advances and applications of WRF-SFIRE. Natural Hazards and Earth System Science 2014, 14, 2829–2845. [Google Scholar] [CrossRef]
  8. NCEI. Dead Fuel Moisture. https://www.ncei.noaa.gov/access/monitoring/dyk/deadfuelmoisture, 2024. Accessed: May 31, 2024.
  9. NIFC. Remote Automatic Weather Stations (RAWS). https://www.nifc.gov/about-us/what-is-nifc/remote-automatic-weather-stations, 2024. Accessed: December 26, 2024.
  10. Viney, N.R. A review of fine fuel moisture modelling. International Journal of Wildland Fire 1991, 1, 215–234. [Google Scholar] [CrossRef]
  11. Synoptic Data Public API. Mesonet Station Networks and Providers. https://demos.synopticdata.com/providers/index.html, 2025. Accessed: October 15, 2025.
  12. Campbell Scientific, Inc.. CS506 Fuel Moisture Sensor Instruction Manual, 2015. https://s.campbellsci.com/documents/us/manuals/cs506.pdf, Accessed: July 3, 2024.
  13. FTS, I. Fuel Stick Sensor (FS-3) Technical Specifications. https://ftsinc.com/wp-content/uploads/2016/12/Fuel-Stick-Sensor-Technical-Specifications.pdf, 2016. Accessed: 2025-10-15.
  14. Van Wagner, C.E.; Pickett, T.L. Equations and FORTRAN Program for the Canadian Forest Fire Weather Index System. Canadian Forestry Service, Forestry Technical Report 33, 1985. https://ostrnrcan-dostrncan.canada.ca/handle/1845/228362.
  15. Van Wagner, C.E. Development and Structure of the Canadian Forest Fire Weather Index System. Canadian Forestry Service Forestry Technical Report 35, 1987. https://ostrnrcan-dostrncan.canada.ca/handle/1845/228434.
  16. Nelson Jr., R. M. Prediction of diurnal change in 10-h fuel stick moisture content. Canadian Journal of Forest Research 2000, 30, 1071–1087. [Google Scholar] [CrossRef]
  17. Van der Kamp, D.W.; Moore, R.D.; McKendry, I.G. A model for simulating the moisture content of standardized fuel sticks of various sizes. Agricultural and Forest Meteorology 2017, 236, 123–134. [Google Scholar] [CrossRef]
  18. Carlson, J.D.; Bradshaw, L.S.; Nelson, Ralph M., J.; Bensch, R.R.; Jabrzemski, R. Application of the Nelson model to four timelag fuel classes using Oklahoma field observations: Model evaluation and comparison with National Fire Danger Rating System algorithms. International Journal of Wildland Fire 2007, 16, 204–216. [CrossRef]
  19. Vejmelka, M.; Kochanski, A.K.; Mandel, J. Data assimilation of fuel moisture in WRF-SFIRE. In Proceedings of 4th Fire Behavior and Fuels Conference, 18-22 February 2013, Raleigh, NC and 1-4 July 2013, St. Petersburg, Russia; Wade, D.D.; Fox, R.L., Eds.; International Association of Wildland Fire: Missoula, MT, 2014; pp. 122–137. Compiled by M. L. Robinson. Available at http://www.iawfonline.org/4th_Fuels_Conference_Proceedings_USA-Russia.pdf, accessed July 2014.
  20. Mandel, J.; Vejmelka, M.; Kochanski, A.K.; Farguell, A.; Haley, J.D.; Mallia, D.V.; Hilburn, K. An Interactive Data-Driven HPC System for Forecasting Weather, Wildland Fire, and Smoke. In 2019 IEEE/ACM HPC for Urgent Decision Making (UrgentHPC), Supercomputing 2019, Denver, CO, USA; IEEE, 2019; pp. 35–44. [CrossRef]
  21. Fan, C.; He, B. A Physics-Guided Deep Learning Model for 10-h Dead Fuel Moisture Content Estimation. Forests 2021, 12. [Google Scholar] [CrossRef]
  22. Kang, Z.; Jiao, M.; Zhou, Z. Fuel Moisture Content Forecasting Using Long Short-Term Memory(LSTM) Model. In Proceedings of the IGARSS 2022 - 2022 IEEE International Geoscience and Remote Sensing Symposium; 2022; pp. 5672–5675. [Google Scholar] [CrossRef]
  23. Schreck, J.S.; Petzke, W.; Jiménez, P.A.; Brummet, T.; Knievel, J.C.; James, E.; Kosović, B.; Gagne, D.J. Machine Learning and VIIRS Satellite Retrievals for Skillful Fuel Moisture Content Monitoring in Wildfire Management. Remote Sensing 2023, 15. [Google Scholar] [CrossRef]
  24. Research Applications Laboratory, National Center for Atmospheric Research (NCAR). Fuel Moisture Content Retrievals. https://ral.ucar.edu/tool/fuel-moisture-content-retrievals, 2025. Accessed: October 31, 2025.
  25. Mandel, J.; Hirschi, J.; Kochanski, A.K.; Farguell, A.; Haley, J.; Mallia, D.V.; Shaddy, B.; Oberai, A.A.; Hilburn, K.A. Building a Fuel Moisture Model for the Coupled Fire-Atmosphere Model WRF-SFIRE from Data: From Kalman Filters to Recurrent Neural Networks. In SNA’23 Seminar on Numerical Analysis; Institute of Geonics of the Czech Academy of Sciences, 2023; pp. 52–55. https://www.ugn.cas.cz/event/2023/sna, also available as arXiv:2301.05427, https://doi.org/10.48550/arXiv.2301.05427. [CrossRef]
  26. Vejmelka, M.; Kochanski, A.K.; Mandel, J. Data assimilation of dead fuel moisture observations from remote automatic weather stations. International Journal of Wildland Fire 2016, 25, 558–568. [Google Scholar] [CrossRef]
  27. Blaylock, B. SynopticPy: Synoptic API for Python (Version 2024.12.0). https://github.com/blaylockbk/SynopticPy, 2024. Computer software.
  28. Dowell, D.C.; Alexander, C.R.; James, E.P.; Weygandt, S.S.; Benjamin, S.G.; Manikin, G.S.; Blake, B.T.; Brown, J.M.; Olson, J.B.; Hu, M.; et al. The High-Resolution Rapid Refresh (HRRR): An Hourly Updating Convection-Allowing Forecast Model. Part I: Motivation and System Description. Weather and Forecasting 2022, 37, 1371–1395. [Google Scholar] [CrossRef]
  29. Blaylock, B. Herbie: Retrieve Numerical Weather Prediction Model Data (Version 2024.8.0). https://github.com/blaylockbk/Herbie, 2024. Computer software.
  30. Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, second ed.; O’Reilly Media: Sebastopol, CA, 2019. https://research-ebsco-com.aurarialibrary.idm.oclc.org/linkprocessor/plink?id=6093bf50-e842-3a3a-a38b-be8566dfa5e1.
  31. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural computation 1997, 9, 1735–80. [Google Scholar] [CrossRef] [PubMed]
  32. Zhang, A.; Lipton, Z.C.; Li, M.; Smola, A.J. Dive into Deep Learning; Cambridge University Press, 2023. Internet Textbook: https://d2l.ai/index.html, (Accessed Oct. 31, 2025).
  33. Keras Team. Keras API: LSTM Layer. https://keras.io/api/layers/recurrent_layers/lstm/, 2025. Model training used return_sequences=True. Accessed: 2025-10-31.
  34. Keras Team. Keras API: Loss Functions. https://keras.io/api/losses/, 2025. "sum_over_batch_size" means the loss instance will return the average of the per-sample losses in the batch. Accessed: 2025-10-31.
  35. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, August 2016, KDD ’16, p. 785–794. [CrossRef]
  36. Hirschi, J. Custom Loss Functions in Fuel Moisture Modeling, 2025, [arXiv:stat.AP/2501.10401].
  37. Hastie, T.; Tibshirani, R.; Friedman, J. The elements of statistical learning: data mining, inference and prediction, 2 ed.; Springer, 2009. http://www-stat.stanford.edu/~tibs/ElemStatLearn/.
Figure 1. FMC observations over one week at station CHAC2.
Figure 1. FMC observations over one week at station CHAC2.
Preprints 183647 g001
Figure 2. Rocky Mountain GACC - 161 RAWS with valid data from 2023 through 2024. (Left) National view, (right) zoomed view
Figure 2. Rocky Mountain GACC - 161 RAWS with valid data from 2023 through 2024. (Left) National view, (right) zoomed view
Preprints 183647 g002
Figure 3. A single recurrent unit takes as input weather data and a recurrent state (left). An LSTM layer and subsequent dense layers forms the recurrent unit (right).
Figure 3. A single recurrent unit takes as input weather data and a recurrent state (left). An LSTM layer and subsequent dense layers forms the recurrent unit (right).
Preprints 183647 g003
Figure 4. The recurrent unit unrolled in time. The recurrent unit with fixed parameters is applied at each time step, and a recurrent state is maintained over time. The output is a mapping of weather inputs to the FMC at the same time.
Figure 4. The recurrent unit unrolled in time. The recurrent unit with fixed parameters is applied at each time step, and a recurrent state is maintained over time. The output is a mapping of weather inputs to the FMC at the same time.
Preprints 183647 g004
Figure 5. A single LSTM cell. (Adapted from p. 569, Géron [30].)
Figure 5. A single LSTM cell. (Adapted from p. 569, Géron [30].)
Preprints 183647 g005
Figure 6. Example of predicted versus observed timeseries of FMC.
Figure 6. Example of predicted versus observed timeseries of FMC.
Preprints 183647 g006
Figure 7. Spatiotemporal cross-validation diagram. Spatial locations are randomly sampled for train/test sets (left), and the training period precedes the testing period (right).
Figure 7. Spatiotemporal cross-validation diagram. Spatial locations are randomly sampled for train/test sets (left), and the training period precedes the testing period (right).
Preprints 183647 g007
Figure 8. Distributions of observed (left) and predicted (right) FMC.
Figure 8. Distributions of observed (left) and predicted (right) FMC.
Preprints 183647 g008
Figure 9. Forecast Error by RAWS, averaged over all times and replications.
Figure 9. Forecast Error by RAWS, averaged over all times and replications.
Preprints 183647 g009
Figure 10. Forecast error for RNN, ODE+KF, XGboost, and Climatology methods by hour of day, averaged over all locations, replications, and days of the year.
Figure 10. Forecast error for RNN, ODE+KF, XGboost, and Climatology methods by hour of day, averaged over all locations, replications, and days of the year.
Preprints 183647 g010
Table 1. Predictors Summary
Table 1. Predictors Summary
Variable Units Description Mean Low High
Drying Equilibrium % Derived from RH and temperature. 15.21 1.33 36.90
Wetting Equilibrium % Derived from RH and temperature. 13.84 0.61 34.65
Solar Radiation k W h / m 2 Downward shortwave radiative flux. 220.99 0.00 1130.00
Wind Speed m / s Wind speed at 10m. 4.70 0.13 32.50
Elevation meters Height above sea level. 1859.86 304.80 3521.96
Longitude degree X-Coordinate -104.78 -110.93 -95.85
Latitude degree Y-Coordinate 40.90 37.09 45.99
Rain mm/h Calculated from rain accumulated over the hour. 0.04 0.00 68.29
Hour of Day hours From 0 to 23, UTC time. 11.50 0.00 23.00
Day of Year days From 0 to 365 or 366 (leap year in 2024) 197.00 1.00 366.00
Table 2. RNN Hyperparameters
Table 2. RNN Hyperparameters
Hyperparameter Value Description
Learning Rate 0.001 Controls how fast weights update during training
Timesteps 48 Sequence size
Batch Size 64 Number of sequences to process before updating weights
Dropout 0.15 Rate that weights are randomly set at during training
Early Stopping Patience 5 Number of epochs with no improvement to validation error before stopping training
Scaling Standard Scaling inputs to mean zero and std. 1
Table 3. ODE+KF Hyperparameters
Table 3. ODE+KF Hyperparameters
Hyperparameter Value Description
Spinup 24 Number of hours to run model with data assimilation
Process Variance 1e-3 Uncertainty in model dynamics
Data Variance 1e-3 Uncertainty in measurement of input data
r 0 0.05 Threshold rainfall intensity ( m m h 1 )
r s 8.00 Saturation rainfall intensity ( m m h 1 )
T r 14.00 Characteristic decay time for wetting dynamics (h)
S 250 Saturation FMC level
T 10.00 Characteristic decay time for fuel class
Table 4. XGBoost Hyperparameters
Table 4. XGBoost Hyperparameters
Hyperparameter Value Description
N Estimators 100 Number of trees in ensemble
Max Tree Depth 3 Maximum tree depth.
Learning Rate 0.10 Controls how fast weights update during training
Min Child Weight 1 Minimum sum of Hessian of samples in a partition
Gamma 0.10 Minimum loss reduction required for a tree partition
Subsample 0.80 Percent of training data randomly selected in each boosting iteration
Colsample by Tree 0.90 Percent of predictors randomly selected in each boosting iteration
Table 5. Climatology Hyperparameters
Table 5. Climatology Hyperparameters
Hyperparameter Value Description
Years 10 Number of years to look for historical FMC at a given location
Days 15 Number of days plus or minus the target time to aggregate historical FMC data
Min. Years 6 Minimum unique number of years of observed FMC to generate a prediction
Table 6. Mathematical Definitions used for Accuracy Metrics
Table 6. Mathematical Definitions used for Accuracy Metrics
Term Description Data Summary
N Number of unique RAWS 151 RAWS with FMC data in 2024 in study region
T Number of forecast hours 8,784 hours in 2024 (a leap year)
R Number of replications 500
y i , t Observed FMC at location i and time1 t -
y r , i , t Predicted FMC at location i, time t, and replication2 r -
1 Observed FMC has no variance across replications. 2 Predicted FMC has variance across replications due to random selection of stations used in training set and random initializations of the ML model weights.
Table 7. Overall Forecast Error
Table 7. Overall Forecast Error
Model Bias (% FMC) RMSE (% FMC)
RNN -0.19 ± 0.33 3.02 ± 1.12
ODE+KF -2.41 ± 0.24 4.97 ± 1.52
XGBoost -0.4 ± 0.32 4.19 ± 1.31
Climatology -0.64 ± 0.4 5.0 ± 1.46
Table 8. RNN Bias by FMC Level.
Table 8. RNN Bias by FMC Level.
FMC Level Bias (% FMC) ± Std
Low (0-10) -1.19 0.28
Medium (10-20) 0.18 0.41
High (20+) 4.46 0.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated