Preprint
Article

This version is not peer-reviewed.

LFTD: Transformer-Enhanced Diffusion Model for RealisticFinancial Time-Series Data Generation

A peer-reviewed article of this preprint also exists.

Submitted:

16 January 2026

Posted:

16 January 2026

You are already at the latest version

Abstract
Firm-level financial statement data form multivariate annual time series with strong cross-variable dependencies and temporal dynamics, yet publicly available panels are often short and incomplete, limiting the generalization of predictive models. We present Latent Financial Time-Series Diffusion (LFTD), a structure-aware augmentation framework that synthesizes realistic firm-level financial time series in a compact latent space. LFTD first learns information-preserving representations with a dual encoder: an FT-Transformer that captures within-year interactions across financial variables and a Time Series Transformer (TST) that models long-horizon evolution across years. On this latent sequence, we train a Transformer-based denoising diffusion model whose reverse process is FiLM-conditioned on the diffusion step as well as year, firm identity, and firm age, enabling controllable generation aligned with firm- and time-specific context. A TST-based cross-decoder then reconstructs continuous and binary financial variables for each year. Experiments on Korean listed-firm data from 2011–2023 show that augmenting training sets with LFTD-generated samples consistently improves firm-value prediction for market-to-book and Tobin’s Q under both static (same-year) and dynamic (τ→ τ+1) forecasting settings, and outperforms conventional generative augmentation baselines and ablated variants. These results suggest that domain-conditioned latent diffusion is a practical route to reliable augmentation for firm-level financial time series.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

Financial data analysis has long relied on traditional regression-based statistical models. However, as the need to handle nonlinearity, high dimensionality, and heterogeneous data simultaneously has grown, the use of machine- and deep-learning-based methods has expanded rapidly. By capturing complex structures more flexibly, these approaches can improve predictive accuracy relative to traditional models and partially automate risk assessment. As a result, they are now applied to a wide range of tasks, including firm valuation, bankruptcy prediction, deal-structure optimization, and earnings forecasting. At the same time, because financial statements are numerical, tabular data with time-series properties, model performance depends heavily on the quantity and quality of the training data. In practice, however, publicly available financial data typically cover only a limited set of firms and years and are characterized by frequent missing values and discontinuities in coverage, making it difficult to construct consistent long-term time series. This scarcity and fragmentation of data can substantially degrade the generalization performance of machine- and deep-learning models, which must learn highly complex patterns.
A widely used approach to alleviating data scarcity is data augmentation, which refers to techniques that generate additional training data by transforming or imitating existing samples. In domains such as computer vision and speech recognition, where strong structural regularities are present, a variety of augmentation methods—including rotation, cropping, noise injection, masking, and interpolation—have been used effectively [1,2]. More recently, research has moved beyond simple transformations to leverage deep generative models such as GANs [3], VAEs [4], and diffusion models [5,6]. These methods learn the statistical properties and structural patterns of the original data and can generate samples that exhibit both realism and diversity; in image and speech applications, they produce synthetic samples that are “sufficiently plausible” and closely aligned with the true data distribution.
Financial statements, however, are a composite data type that combines a static tabular structure with a time-series structure, and the interdependence among variables is tightly constrained by accounting identities and the definitions of financial ratios. Financial statements exhibit tightly coupled relationships among variables and ratios, so naive augmentation may violate economically meaningful cross-variable relationships and temporal coherence. For financial data with an inherent time-series structure, it is essential to preserve both temporal dynamics and inter-variable relationships, so directly applying augmentation techniques developed for other domains can easily undermine financial consistency.
To overcome these limitations, we propose a conditional transformer–diffusion–based data augmentation model that simultaneously captures the tabular and time-series characteristics of financial data. Specifically, we first embed metadata such as year, and firm identifiers together with continuous and categorical financial variables and then use the FT-Transformer [9] to construct column embeddings that reflect inter-variable relationships. Because the FT-Transformer applies self-attention over column tokens, it does not directly model dependencies on temporal order or sequence length. We therefore employ a Time Series Transformer (TST) [10] to capture temporal dynamics and interactions among variables and to construct a conditioning representation sequence. We call our model Latent Financial Time-Series Diffusion (LFTD).
In the LFTD, the transformer module takes this conditioning representation sequence, along with firm identifiers, positional encodings, and year information, as conditioning inputs and maps them to a latent vector. The diffusion denoiser is trained to reconstruct the conditioning representation sequence from this latent representation. After training, given a user-specified firm identifier and sequence of years, we encode them through the same FT-Transformer and TST pipeline to obtain a conditioning representation sequence and then perform a diffusion-based reconstruction process conditioned on this sequence to generate new multivariate financial time series that simultaneously exhibit financial consistency and temporal coherence.
The main contributions of this study are as follows:
  • Domain-consistent augmentation: We present a structure-preserving augmentation framework for multivariate firm-level financial time series that explicitly targets cross-variable dependencies and targets cross-variable dependencies and reduces violations of economically meaningful relationships among financial indicators.
  • Dual-dependency representation: We introduce an encoding design that disentangles and captures temporal evolution across years and variable-wise interactions within each firm-year record, producing compact latents for generation.
  • Controllable conditional diffusion: We develop a conditional Transformer–Diffusion generator operating in latent space, conditioned on year and firm identity to enable controllable synthesis with temporal coherence.
The remainder of this paper is organized as follows. Section 2 reviews related work. Section 3 describes the proposed method in detail. Section 4 presents quantitative evaluations focusing on the statistical properties and predictive performance of the generated data. Section 5 concludes and discusses directions for future research.

2. Related Work

2.1. Financial Data Analysis with Machine Learning and Deep Learning

Financial statements are key instruments for assessing a firm’s financial soundness and operating performance, and they have traditionally been analyzed using statistical methods such as regression and discriminant analysis [11]. More recently, machine-learning models including Random Forest, XGBoost, and SVM have been introduced for tasks such as bankruptcy prediction and profitability analysis, enabling more effective modeling of complex nonlinear relationships among variables [12,13,14]. For example, Kang et al. (2018) applied XGBoost to forecast bankruptcy risk for Korean listed firms and reported a substantial improvement in AUROC relative to conventional regression models. Similarly, Zhou et al. (2020) showed that an SVM-based model for predicting bank loan defaults achieved higher accuracy than traditional discriminant analysis, thereby demonstrating the benefit of capturing nonlinear relationships. However, most of these studies rely on static data from a single point in time or, at most, year-by-year snapshots, and thus have limited ability to fully exploit temporal continuity. Recurrent neural networks such as LSTM and GRU [15,16] have seen some adoption, but they remain insufficient for learning long-term dependencies across years, are structurally difficult to parallelize, and are not well suited to handling tabular structures such as financial statements, which combine categorical and numerical variables. To overcome these limitations, Transformer-based models have recently attracted growing attention. TabTransformer [24] and FT-Transformer [9] exhibit strong representation-learning performance on tabular data with mixed categorical and numerical features. Huang et al. (2020), for instance, applied TabTransformer to a loan default prediction dataset and reported a 3.1% improvement in AUC compared with a Random Forest baseline. In the time-series domain, models such as TST (Time Series Transformer) [10] and TFT (Temporal Fusion Transformer) [25] have been proposed. Lim et al. (2021) showed that TFT improves long-term forecasting accuracy over LSTM for electricity demand prediction, while also providing interpretable variable importance. Nevertheless, prior work has mainly used Transformers for prediction or classification tasks, and only a few studies either combine learned Transformer representations with generative models or jointly account for the composite table–time-series structure characteristic of financial statements.

2.2. Tabular and Time-Series Data Augmentation

Data augmentation is a key technique for improving model generalization, and tabular and time-series data require distinct approaches because of their different structural characteristics. For tabular data, methods such as SMOTE [17], CTGAN [18], TableGAN [19], and VAE-based models [4] are widely used, with a primary focus on preserving the underlying statistical structure among variables. For example, CTGAN, proposed by Xu et al. (2019), demonstrated more stable distributional fidelity than conventional GANs on the UCI Adult income dataset. In contrast, time-series augmentation must simultaneously account for temporal dynamics and joint relationships among variables. Methods such as TimeGAN [20], T-CGAN [21], and TS-Augment [22] have been proposed for this purpose. Yoon et al. (2019) applied TimeGAN to financial time series and showed that the generated samples preserve autocorrelation structures similar to those of the original data. However, existing approaches are typically specialized for either tabular or time-series data alone and therefore face limitations when applied to composite data such as financial statements, which embody both structures. Time-series data possess a distinctive structure in which temporal dependencies and cross-variable correlations coexist, so naive augmentation strategies can easily introduce semantic distortions. In particular, both the magnitude of values and their evolution over time (i.e., ordering and trends) are important for interpretation, making it difficult to directly adopt augmentation schemes designed for static data. Um et al. (2017), analyzing sensor data from Parkinson’s disease patients, showed that traditional augmentation methods such as noise injection and temporal shifting can actually degrade predictive performance and interpretability [23]. Zerveas et al. (2021) likewise reported that simple resampling strategies may compromise clinical meaning in medical time-series data. These findings suggest that arbitrary transformations can similarly distort the economic meaning of financial data, underscoring the need for sophisticated generative augmentation methods that preserve both structural consistency and temporal dynamics.

2.3. Diffusion-Based Time-Series and Tabular Data Generation

Diffusion models have emerged as powerful generative methods across a wide range of domains—including images, text, and time series—because of their robustness in learning complex data distributions. A representative example is the Denoising Diffusion Probabilistic Model (DDPM), which generates samples by gradually corrupting data with noise and then learning the reverse denoising process [5]. Subsequent extensions, such as score-based and conditional diffusion models, have broadened their applicability to various tasks [6]. In the time-series domain, CSDI (Conditional Score-based Diffusion Model for Imputation) by Tashiro et al. (2021) [26] is a representative approach; on electricity consumption data, it achieved more than a 20% improvement in imputation performance over LSTM-based baselines. For tabular data, TabDDPM by Kotelnikov et al. (2022) [27] demonstrated lower Kolmogorov–Smirnov statistics than VAE- and CTGAN-based models on credit risk datasets, indicating improved statistical fidelity of the generated samples.
Nonetheless, most existing diffusion-based models are tailored to either static tabular data or pure time-series data, making them difficult to apply directly to composite datasets such as financial statements, where tabular and time-series structures are tightly coupled. To address this gap, the present study proposes a diffusion model that uses financial embeddings derived from FT-Transformer and TST as conditioning signals, with the aim of generating realistic, finely detailed financial time series that jointly satisfy statistical consistency and temporal coherence.

3. Methodology

In this paper, we proposed a conditional time-series generation model called Latent Financial Time-Series Diffusion (LFTD). The proposed model aims to generate realistic synthetic financial data from numerical time-series data in the form of financial statements. Because financial statement data exhibits a complex structure that combines a static tabular format of yearly observations and long-term temporal dynamics accumulated at the firm level, effectively learning such data requires latent representation learning that simultaneously captures both the intra-year inter-variable structure and the inter-year time-series structure. To this end, the proposed LFTD consists of three key modules organized in a two-stage training framework.
In the first stage, a time-series encoder that combines an FT-Transformer and a Time Series Transformer (TST) embeds firm-level financial time series into a latent representation space. Specifically, the FT-Transformer learns the structural relationships among the continuous and binary numerical variables contained in each year’s financial statements, summarizing the annual financial status into a high-dimensional vector. These yearly latent vectors are then fed into the TST, which models long-term financial dynamics and temporal dependencies at the firm level. This stage is pre-trained in an autoencoder manner, forming a stable latent financial time-series representation that reflects the structural characteristics of financial time series.
In the second stage, a conditional diffusion model with a Transformer-based denoiser is trained using the latent time-series representations obtained from the pre-trained encoder. The diffusion process is performed over the entire latent time series at the firm level, and the denoising in the reverse diffusion steps is conditioned not only on the diffusion time step but also on year, firm ID, and firm age. These conditioning signals are injected into each layer of the Transformer denoiser via FiLM (Feature-wise Linear Modulation), enabling condition-aware generation of latent time series.
Finally, the latent time series recovered through the reverse diffusion process is decoded back into the original financial variable space via a TST-based Cross-Decoder. This decoder performs cross-attention over the latent time-series representations using learnable query tokens, reconstructing the yearly financial variables into a temporally consistent time-series structure. Through this design, the proposed model can generate synthetic financial time-series data that satisfy firm-level and year-level conditions. Figure 1 provides an overview of the full architecture and training pipeline of the proposed LFTD model.

3.1. Dataset Description

The dataset used in this study is constructed from the financial statements of Korean listed firms and contains consecutive annual observations from 2011 to 2023, for a total of 13 years. Each firm is identified at the entity (stock) level, and for each year (YEAR), key financial indicators and accounting characteristics are recorded. For the final analysis, we restrict the sample to firms that are observed without missing values for the full 13-year period.
Table 1 summarizes the definitions of the variables included in the dataset. The variables can be broadly classified into continuous financial variables and binary indicators. The continuous variables include insider ownership (OWN), foreign ownership (FORN), firm size (SIZE), leverage ratio (LEV), current ratio (CUR), sales growth (GRW), return on assets (ROA), return on equity (ROE), operating cash flow ratio (CFO), property, plant, and equipment ratio (PPE), inventory and receivables ratio (INVREC), market-to-book ratio (MB), and Tobin’s Q (TQ). The binary variables indicate whether the external auditor belongs to a Big 4 accounting firm (BIG4) and whether the firm reports a net loss in the current year (LOSS).
In addition, the dataset includes three auxiliary variables that are not directly generated by the model but are used to define the panel structure and provide contextual information: the calendar year (YEAR) as a temporal index, a stock-level firm identifier (STOCK_ID), and firm age (AGE), which is treated as a firm-specific conditioning variable. All continuous main input variables are normalized prior to model training. This data design enables the model to capture both temporal patterns within firms and cross-sectional heterogeneity across firms.

3.2. Year-Wise Financial Feature Embedding

In this study, we consider firm-level financial time-series data, where each firm has consecutively observed financial statement information for a total of 13 years from 2011 to 2023. The financial data observed at each year τ are represented as a structured vector consisting of 13 continuous variables and 2 binary variables. In this section, we define the process of embedding these year-level financial inputs into a single latent vector.
We define the financial input firm n at year τ as follows: x n , τ = c n , τ , 1 , , c n , τ , 13 , b n , τ , 1 , b n , τ , 2 R 15 where c n , τ , i denotes the i-th continuous financial variable and b n , τ , i denotes the i-th binary financial variable. Each firm has a financial time series over 13 consecutive years, denoted as x n , 1 , x n , 2 , , x n , 13 .

3.2.1. Year-Wise Feature Tokenization

Because year-level financial data have a structured format in which inter-variable interactions are crucial, it is difficult for simple linear transformations or time-series models alone to adequately capture the intra-year structural relationships. To address this, we employ an FT-Transformer to embed each year’s financial input into a latent space [9].
The FT-Transformer is a Transformer-based model that treats each financial variable as an independent token and learns interactions among variables through a self-attention mechanism [9,24]. As shown in Eq. (1), the year- τ , firm n and financial input x n , τ is decomposed into variable-wise components, each of which is mapped into a vector space via an independent embedding function.
e n , τ , i = f i x n , τ , i , i = 1 , , 15
Here, f i ( ) is a learnable linear mapping corresponding to variable i . As a result, year τ is represented as the following sequence of variable tokens: E n , τ = e n , τ , 1 , , e n , τ , 15 R 15 × d m o d e l

3.2.2. FT-Transformer Encoding and Year-Level Aggregation

The variable-token sequence E n , τ is used as the input to the FT-Transformer encoder, and through self-attention it is transformed into contextualized representations that reflect the relative importance and correlations among variables within the year [9]. The encoder output tokens are then aggregated into a single year-level representation via mean pooling, as shown in Eq. (2).
h n , τ = 1 N i = 1 N e n , τ , i
Here, N denotes the number of variable tokens, and in this case N=15. Then, as shown in Eq. (3), the final year embedding is obtained via a linear projection.
z n , τ = W h n , τ + b , z n , τ R 128
In this way, each year’s financial input is transformed into a fixed-length latent vector that reflects the structural interactions among variables within the year. Figure 2 schematically illustrates the overall process by which year-level financial inputs are converted into year embeddings through variable tokenization, the FT-Transformer encoder, and aggregation. The resulting year embeddings z n , τ are then used as inputs for learning firm-level temporal structure and are fed into the Time-Series Transformer module described in the next section.

3.3. Representation Learning via Time Series Transformer

In this section, we describe how the year-level embeddings z n , τ R 128 obtained in Section 3.2 are organized as a sequence along the time (year) axis to learn a firm’s long-term financial dynamics and inter-temporal dependencies. For firm n , we denote the sequence of T=13 yearly embeddings as Z n = z n , 1 , z n , 2 , , z n , 13 R T × 128 . This sequence is then used as the input to the Time Series Transformer (TST) encoder. Through self-attention, the TST directly models interactions between any two time steps, enabling it to learn long-range dependencies such as trends, turning points, and lagged effects [10,35].

3.3.1. Temporal Smoothing via 1D Convolution

The year-embedding sequence Z is first passed through a one-dimensional convolution (1D convolution) layer applied along the year axis. This layer aggregates local context across adjacent years, helping to smooth short-term noise or abrupt fluctuations and to form a more stable time-series representation. The resulting output sequence Z ~ is expressed as Eq. (4).
Z ~ = C o n v 1 D Z

3.3.2. Sinusoidal Positional Encoding

Because a Transformer does not inherently encode sequential order, we add sine–cosine positional encoding to explicitly inject year ordering and relative distance information. Let the positional encoding matrix be P R 13 × 128 ; then, the input to the TST is constructed as shown in Eq. (5).
H 0 = Z ~ + P
In this case, each element of P follows the standard sinusoidal encoding scheme [35].

3.3.3. Transformer Encoder for Temporal Dependency Modeling

The TST consists of L Transformer encoder blocks, where each block is composed of multi-head self-attention and a position-wise feed-forward network (FFN), along with residual connections and layer normalization [10,35]. Let H l 1 denote the input to the l -th block and H l its output; then H l can be expressed as shown in Eq. (6).
H l = TSTBloc k l H l 1 , l = 1 , , L
Through self-attention, the representation at each time step can reference information from all other time steps as a weighted sum, where the weights reflect the importance of inter-temporal interactions [35]. As a result, the final output Z ' = H L R 13 × 128 is a refined latent time-series representation in which each year-level embedding is contextualized by the entire sequence.
This Z ' simultaneously captures intra-year variable interactions (via the FT-Transformer) and inter-year temporal dynamics (via the TST). It is subsequently used as a latent-space input for the reconstruction learning in Stage 1 and for the conditional diffusion model in Stage 2. Figure 3 schematically illustrates the process by which the year-level embeddings are transformed into Z ' through Conv1D, positional encoding, and the TST encoder.

3.4. Latent Space Pretraining via Autoencoder (Stage 1)

For a conditional diffusion model to be stably trained in the latent space, the latent time-series representations used as diffusion inputs must sufficiently preserve the original financial information. To this end, we construct the FT–TST encoder (Section 3.2 and Section 3.3) in an autoencoder framework and conduct Stage 1 pre-training [36]. The goal of Stage 1 is not generation, but rather to establish an information-preserving latent space that is suitable for subsequent diffusion training.

3.4.1. Decoder Architecture for Reconstruction

In Stage 1, we train a decoder to reconstruct the original financial variables (13 continuous and 2 binary) from the TST output latent time series Z ' R 13 × 128 . The decoder first maps each latent vector into a shared hidden space and then branches into separate output heads for continuous and binary variables. Let z n , τ ' denote the latent vector of firm n at year τ ; the shared transformation is given in Eq. (7).
u n , τ = ϕ W s h z n , τ ' + b s h
Here, ϕ ( ) denotes a nonlinear activation function (SiLU). The reconstructed continuous variables c ^ n , τ R 13 and the binary logits b ^ n , τ logit R 2 are then computed as in Eq. (8), respectively.
c ^ n , τ = W c u n , τ + b c , b ^ n , τ logit = W b u n , τ + b b

3.4.2. Reconstruction Objectives

The reconstruction loss for the continuous variables is computed using the mean squared error (MSE), as given in Eq. (9).
L c o n t = 1 D c k = 1 D c ( c ^ n , τ , k c n , τ , k ) 2
In this case, dimension D c is 13.
The binary variables are trained using the logit-based binary cross-entropy loss (BCE with logits). Equation (10) is given as follows.
L b i n = 1 D b j = 1 D b max b ^ n , τ ,   j logit , 0 b ^ n , τ ,   j logit b n , τ , j + l o g ( 1 + exp b ^ n , τ ,   j logit   )
Where D b is 2. The overall Stage 1 objective function is defined as a weighted sum of the two losses, as shown in Eq. (11).
L A E = L c o n t + λ b i n L b i n
Here, λ bin is a hyperparameter that controls the balance between the reconstruction losses for the continuous and binary variables.

3.4.3. Role of Stage 1 Pretraining

Through this reconstruction-based training, the encoder learns latent representations that not only compress the input but also preserve the structural relationships among financial variables and the underlying temporal dynamics[36]. Consequently, the encoder trained in Stage 1 enables the Stage 2 conditional diffusion model to perform stable denoising in the latent space, and it serves as a foundation for reducing information loss when reverse-diffusion samples are decoded back into the original variable space. Figure 4 schematically illustrates the autoencoder pre-training framework, where the decoder branches into continuous and binary output heads.

3.5. Conditional Latent Diffusion with Transformer-Denoiser

In Stage 2 of this study, the diffusion process is performed directly on the latent time-series representations produced by the FT–TST encoder. Specifically, we use the time-series latent representation Z n ' = z n , 1 ' , z n , 2 ' , , z n , T ' obtained in Section 3.3 as the reference state for the diffusion process and generate new latent time series through forward diffusion and reverse diffusion. By conducting diffusion in the latent space, we can model the high-dimensional and complex distribution of the original financial data in a more stable and efficient manner [5,6].

3.5.1. Forward Diffusion in Latent Space

In the forward diffusion stage, Gaussian noise is gradually injected into the latent time series Z n ' , producing noisy states Z s for diffusion steps s = 1,2 , , S . This process follows the DDPM framework, where the noise injection at each step is defined as in Eq. (12).
Z s = α ¯ s Z n ' + 1 α ¯ s ϵ , ϵ N 0 , I
Here, α ¯ s = i = 1 S α i is the cumulative attenuation coefficient, and α s follows a predefined noise schedule. As the diffusion step increases, Z s gradually approaches pure Gaussian noise [5].

3.5.2. Transformer-Denoiser Architecture

In the reverse diffusion stage, we use a Transformer-based denoiser to learn how to progressively recover the original latent structure from the noisy latent time series Z n , s . By adopting a self-attention–based Transformer architecture, the denoiser models global interactions across the entire sequence and effectively captures long-range temporal dependencies[10,35].
The noisy latent time series Z n . s is mapped to the denoiser’s model dimension via a linear projection, as shown in Eq. (13).
H n , s = W in Z n , s + b in
Afterward, a year-index–based positional encoding is added to reflect the yearly order, and the diffusion step s is transformed into a sine–cosine time embedding and injected into the input. The final denoiser input is constructed as H n , s + P E year + TE s In this study, we adopt conditional diffusion to improve controllability and realism in the generation process. Year information, firm identifier, and firm age are each represented via embeddings, and these condition embeddings are concatenated into a single condition vector. This condition vector is injected into each layer of the Transformer denoiser using FiLM (Feature-wise Linear Modulation) [37]. FiLM applies a linear modulation to the denoiser’s hidden representation h as in Eq. (14), where c denotes the combined conditioning information (year, firm ID, and firm age). Here, γ ( ) and β ( ) are learnable functions. Through this mechanism, conditioning signals directly influence the denoiser’s internal computations, enabling the model to learn latent time-series generation that reflects firm-specific and year-specific characteristics [26,37].
FiLM h c = γ c h + β c
The Transformer denoiser consists of multiple Transformer blocks, each comprising multi-head self-attention and a feed-forward network, along with residual connections and layer normalization [35]. With this architecture, the denoiser predicts noise at each diffusion step by considering the context of the entire latent time series.

3.5.3. Training Objective with SNR-Weighted Loss

During training, we design the loss function such that the denoiser accurately predicts the injected noise ϵ at each diffusion step. Given the denoiser’s noise prediction ϵ ^ θ Z n , s , s , c , we minimize the following objective [5].
In addition, we apply an SNR (signal-to-noise ratio)–based weighting scheme across diffusion steps to mitigate training instability in high-noise regimes [6,26]. This design encourages stable learning throughout the diffusion trajectory and helps preserve the long-term structure of the latent time series. Diffusion loss is defined in Equation (15).
L d i f f = E s , ϵ [ w s | | ϵ ϵ ^ θ Z n , s , s , c | | 2 ]
Here, w s denotes min S N R s ,   γ and S N R s is defined as α ¯ s / ( 1 α ¯ s ) .

3.5.4. Sampling via Reverse Diffusion

During generation, we perform reverse diffusion starting from pure Gaussian noise using the trained Transformer denoiser. At each step, the predicted noise is substituted into the reverse diffusion update, repeatedly applying Z n , s Z n , s 1 to ultimately recover a latent time series with the same structure as Z ' [5,6]. The generated latent time series is then transformed back into the original financial variable space through the decoder. Figure 5 schematically illustrates the conditional diffusion process in the latent time-series space and the overall architecture of the Transformer denoiser.

3.6. Cross-Decoder and Financial Time-Series Generation

Through the conditional latent diffusion process described in the previous section, the model obtains denoised latent time-series representations after reverse diffusion. In this section, we explain how the recovered latent time series is transformed into the original financial variable space, ultimately generating firm-level financial time-series data.
In our framework, we employ a TST-based Cross-Decoder to reconstruct year-level financial variables from the latent time series. The Cross-Decoder takes a sequence of learnable query tokens as input and uses a cross-attention mechanism that attends the latent time series obtained via reverse diffusion as the encoder memory [35]. This design allows the model to stably recover financial variables corresponding to each year while preserving the temporal structure embedded in the latent representations.

3.6.1. Cross-Attention Decoder Architecture

Specifically, the decoder takes as input a learnable query token sequence of length T = 13 , Q = q 1 , q 2 , , q 13 where each query token represents a single year. Within the Transformer decoder blocks, self-attention is first applied among the query tokens to learn inter-year dependencies [35]. Then, in the cross-attention stage, the query tokens attend to the latent time-series representations restored via reverse diffusion as the memory.
In this cross-attention mechanism, each query q τ is designed to reference information from all time steps of the latent time series as key–value pairs. This enables the decoder to consider not only the target year but also the global time-series context when reconstructing the financial state of a specific year. Unlike independent year-by-year reconstruction, this structure allows the model to recover financial variables while reflecting long-term financial trends and inter-temporal interactions.
Each Transformer decoder block consists of self-attention, cross-attention, and a feed-forward network, with residual connections and layer normalization applied between sublayers [35]. As the query tokens pass through multiple decoder blocks, they are progressively refined into year-level representations that integrate latent time-series information.

3.6.2. Output Projection and Variable Reconstruction

The output of the Cross-Decoder is a sequence of hidden vectors corresponding to year-level financial variables. This output is projected into the original financial variable space through a final linear mapping, producing predictions for both continuous and binary variables simultaneously. Specifically, the continuous variables are predicted as a 13-dimensional real-valued vector, while the binary variables are output in logit form.
During training, the Cross-Decoder takes the latent time series recovered via reverse diffusion as input and is optimized to reconstruct the original firm-year financial variables x n , τ . For the continuous variables, we use the mean squared error (MSE), and for the binary variables, we use the BCEWithLogits loss. Because this reconstruction loss is combined with the denoiser’s noise prediction loss to form the final Stage 2 objective, the Transformer-Denoiser and the Cross-Decoder are jointly trained in Stage 2 by minimizing the following objective [5,6]. Final loss of Stage 2 is defined as Equation (16).
L S t a g e 2 = L d i f f θ + λ r e c L r e c θ , ψ L r e c θ , ψ = E n 1 T τ = 1 T L c o n t + λ b i n L b i n
where L d i f f is the denoiser’s noise prediction loss defined in Eq. (15) (with diffusion step s), and λ r e c controls the trade-off between diffusion learning and financial-variable reconstruction.

3.6.3. Financial Time-Series Generation

During generation, we first sample a latent time series using the trained conditional diffusion model and then transform it into the financial variable space using the Cross-Decoder. For continuous variables, the decoder outputs are used directly. For binary variables, we apply a sigmoid function and discretize the results using a threshold of 0.5 to obtain the final values.
Overall, the proposed generation pipeline consists of two steps: conditional diffusion-based sampling in the latent time-series space and reconstruction of financial variables via a cross-attention decoder [5,6,35]. With this design, the proposed model can generate firm-level long-term financial time series while maintaining structural consistency across years, and the generated data can be leveraged for downstream tasks such as financial forecasting, risk analysis, and data augmentation. Figure 6 schematically illustrates the Cross-Decoder architecture, which reconstructs financial variables via cross-attention between learnable query tokens and the latent time series.

3.7. Post-Processing and Assignment of Year and Firm Identifiers

This section describes the post-processing procedure used to assign year and firm identifiers to the synthetic financial time series generated by the proposed pipeline. In this study, year and firm identifiers are not variables generated by the model itself; rather, they are metadata assigned a posteriori to organize the generated financial variables into an interpretable and usable firm-level panel data format.

3.7.1. Assignment of Year Information

Year information is provided as a conditional index for each time step (t=0, …, 12). In our dataset, we map t to calendar years 2011–2023 for reporting and downstream evaluation. The proposed generative model does not treat calendar year as a prediction target. Instead, year information is incorporated as a conditional signal representing temporal position. Specifically, the generation process in the latent time-series space is performed on sequences of a fixed length, where each timestep is distinguished by its relative position within the sequence (0, 1, …, 12). This positional information is explicitly encoded within the model via positional encoding and year embeddings, allowing the model to recognize the temporal ordering of each timestep in the sequence.
After the generation process is completed, actual calendar years are assigned to each timestep through a one-to-one post hoc mapping. In this study, each generated sequence consists of 13 timesteps, which are sequentially mapped to calendar years from 2011 to 2023. This assignment is performed independently of the model outputs and serves to attach real-world year labels while preserving the relative temporal structure learned by the model.

3.7.2. Assignment of Firm Identifiers

The firm identifier (stock_id) is not generated by the model. Instead, we use a firm ID only as a conditioning key to capture firm-level heterogeneity during latent diffusion sampling. Specifically, for each synthetic sequence, we sample a conditioning firm ID from the training split according to a predefined sampling rule, generate a 13-step latent trajectory conditioned on the sampled firm ID (and other metadata such as year index and firm age), and decode it into financial variables.
After generation, we discard the conditioning firm ID for indexing purposes and assign a new virtual firm ID to the generated sequence so that it is treated as an additional firm in the synthetic panel. The assigned virtual firm ID is used only for panel organization and is not used as an input condition to the generator. All yearly observations within a generated sequence share the same virtual firm ID, resulting in a firm–year panel indexed by (virtual_stock_id, year).

3.7.3. Final Synthetic Panel Construction

Through the above procedures, the generated data are ultimately organized into a firm-level long-term financial panel dataset. Each row corresponds to the financial observations of a specific firm in a specific year and includes both continuous financial variables and binary financial indicators. Because the synthetic data conform to the same structural format as the original dataset, they can be directly utilized—without additional preprocessing—for downstream applications such as financial forecasting, risk analysis, and data augmentation.
In summary, the proposed generation framework synthesizes financial time series using a conditional latent diffusion model, reconstructs financial variables via a Cross-Decoder, and assigns year and firm identifiers in a post-processing step, thereby producing interpretable and practically usable synthetic firm-level panel data.

4. Experiments

This section systematically evaluates the statistical validity and predictive utility of the synthetic financial data generated by the Latent Financial Time-Series Diffusion model, which simultaneously captures nonlinear inter-variable interactions and year-wise temporal dependencies. Our dataset is derived from annual accounting/tax-related financial disclosures, where variables for year τ are typically finalized and become available after the fiscal year ends. As a result, we separate two evaluation settings: a static setting that assesses contemporaneous cross-sectional relationships within the same year (useful for explaining firm-value differences at time τ), and a dynamic setting that evaluates a one-year-ahead forecasting protocol (ττ+1) that better matches information availability and practical prediction use cases.
In detail, we first assess how faithfully the proposed model preserves key statistical properties of the original data, including marginal distributions, correlation structures, and temporal continuity. The same evaluation protocol is applied to synthetic data generated by recent state-of-the-art generative models, enabling a quantitative comparison that highlights the relative advantages and distinctive property-preservation capabilities of the proposed approach [33].
Next, to examine the practical usefulness of the synthetic data, we compare the performance of predictive models trained solely on the original data with those trained using augmented datasets that incorporate the generated synthetic samples. In particular, this study goes beyond assessing gains attributable to a simple increase in sample size and instead disentangles the respective contributions of recovered inter-variable relationships and learned temporal patterns. To this end, component-wise ablation experiments are conducted, and predictive performance is reported under both static prediction settings (within the same year) and dynamic prediction settings (from the previous year to the subsequent year).
All experiments are conducted under strictly controlled conditions to ensure fair comparison: the same preprocessing pipeline, the same data generation ratio, identical training–validation splits, the same predictive models and hyperparameters, and multiple random seed repetitions are consistently applied throughout. This design allows us to clearly attribute observed performance improvements to the structural and temporal properties preserved in the synthetic data, rather than to differences in experimental configurations.

4.1. Static Prediction Model Design

In this experiment, a static prediction model is constructed to predict firm value (MB, TQ) for a given year using financial characteristics from the same year as input variables. This setup aims to examine the relationship between cross-sectional financial attributes at a specific point in time and contemporaneous firm-level differences, and to quantitatively analyze how factors such as financial structure, profitability, and growth jointly affect firms’ market valuation. This setting is consistent with the year-wise cross-sectional analysis perspective widely adopted in empirical financial research [33]. Note that, this static setting is intended to evaluate contemporaneous cross-sectional associations between firm characteristics and firm value within a given year, rather than a real-time forecasting scenario.
The input variables consist of a total of 13 financial features: OWN, FORN, SIZE, LEV, CUR, GRW, ROE, CFO, PPE, INVREC, ROA, BIG4, and LOSS. The output variables are MB and TQ, which represent firm market value and performance.
Model training is conducted using data from 2011 to 2022, while data from 2023 are reserved as the test set for evaluating predictive performance. Specifically, the model learns the relationship between financial characteristics and firm value over the 2011–2022 period during training, and in the evaluation stage, it predicts MB and TQ for the year 2023 based on the corresponding input features. This approach is analogous to the year-wise cross-sectional regression framework commonly employed in empirical finance studies.
The model is based on an XGBoost regressor [13], designed to capture nonlinear structural patterns in the data. XGBoost is a decision-tree-based ensemble boosting method that is well known for its strong predictive performance and effective control of overfitting, making it suitable for financial data prediction [12,13]. The XGBoost regression model is configured with 800 trees, a learning rate of 0.05, a maximum tree depth of 5, a subsampling ratio of 0.8, and a column sampling ratio of 0.8.
Model performance is evaluated using four metrics: Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), R² (coefficient of determination), and Mean Absolute Percentage Error (MAPE), as defined in Equations (17), (18), (19), and (20). The R² metric indicates the proportion of variance in the dependent variable explained by the predictive model, RMSE and MAE measure the absolute magnitude of prediction errors, and MAPE reflects the relative scale of the prediction error.
R M S E = 1 n i = 1 n y i y i ^ 2
M A E = 1 n i = 1 n y i y i ^
R 2 = 1 i = 1 n y i y i ^ 2 i = 1 n y i y ¯ 2
M A P E = 100 n i = 1 n y i y i ^ y i
The datasets consist of the original data and synthetic data generated by the proposed models: Denoiser, FT_Denoiser, FT_TST_Denoiser (LFTD), and TST_Denoiser. All models are evaluated using the same set of input variables and an identical XGBoost regression architecture with the same hyperparameter settings. Based on the model that achieves the best performance across the four-evaluation metrics, additional comparative experiments are conducted against commonly used data augmentation approaches, including CTGAN [18], DDPM [5], and VAE [4]–based models. This stepwise comparative design serves as a benchmark framework for analyzing how different data generation and refinement strategies affect the performance and reliability of predictive models, thereby enabling a systematic assessment of the impact of synthetic data quality on downstream prediction tasks.

4.2. Dynamic Prediction Model Design

In this experiment, a dynamic prediction model is designed to predict firm value (MB, TQ) in the following year using financial characteristics from the previous year as input variables. This approach aims to quantitatively analyze how firm-level financial information accumulates over time and how its temporal evolution affects future market valuation. This dynamic protocol reflects the practical constraint that annual accounting/tax-related variables for year τ become usable for prediction only after the reporting cycle, making ττ+1 a more realistic forecasting setup. In other words, to avoid information leakage, the generative model is trained using only the training period of each prediction setting, and synthetic samples are generated and used only within the corresponding training split.
The input variables consist of a total of 13 financial features: OWN, FORN, SIZE, LEV, CUR, GRW, ROE, CFO, PPE, INVREC, ROA, BIG4, and LOSS. The output variables are MB and TQ, which represent firms’ market value and performance.
For each firm, the data are sorted chronologically and restructured such that the input features from a given year τ are matched with the firm value in the subsequent year τ + 1 . Specifically, financial characteristics from 2021 are used to predict firm value in 2022, and those from 2022 are used to predict firm value in 2023. This design captures time-lagged effects and enables a realistic assessment of how financial performance influences market valuation in subsequent periods.
Model training is conducted using data from 2011 to 2021, while the 2022 data are used as inputs to predict firm value in 2023. This temporally ordered training–evaluation split prevents data leakage and ensures that the model is validated under conditions that closely resemble real-world forecasting scenarios.
The model is based on an XGBoost regressor [13], which is designed to learn nonlinear structural patterns in the data. The XGBoost regression model is configured with 800 trees, a learning rate of 0.05, a maximum depth of 5, a subsampling ratio of 0.8, and a column sampling ratio of 0.8.
Model performance is evaluated using four metrics: Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), R² (coefficient of determination), and Mean Absolute Percentage Error (MAPE), as defined in Equations (17)-(20).

4.3. Comparison and Analysis of Predictive Performance

This section provides a comprehensive analysis of the predictive performance of the proposed models and baseline methods under both static prediction (same-year prediction) and dynamic prediction (previous year → subsequent year) settings. Table 2 report the performance of the proposed denoising-based models in the static prediction setting, while Table 3 present the results in the dynamic prediction setting. In addition, Table 3 (static) and Table 5 (dynamic) compare the LFTD with conventional generative data augmentation methods, including DDPM, VAE, and CTGAN.

4.3.1. Static Prediction Performance Analysis

In the static prediction setting, LFTD achieves the best performance across all evaluation metrics.
For MB prediction (Table 2), LFTD records an RMSE of 0.281 and an R² of 0.948, substantially reducing prediction error and markedly improving explanatory power compared to the original data (RMSE 1.479, R² 0.203).
Similarly, for TQ prediction (Table 2), LFTD attains an RMSE of 0.169 and an R² of 0.955, demonstrating the highest predictive accuracy and explanatory power among all compared models.
While FT_Denoiser achieves the lowest absolute error for MB prediction (RMSE 0.270), its R² value (0.857) is lower than that of LFTD. This indicates that although FT_Denoiser effectively captures nonlinear inter-variable relationships, its inability to explicitly model temporal structure limits its overall explanatory power.
The Denoiser model also yields relatively low RMSE values (MB 0.323, TQ 0.189); however, its R² scores (0.908 and 0.916, respectively) remain consistently below those of LFTD.
In contrast, TST_Denoiser, which focuses solely on temporal structure, exhibits a substantial degradation in performance: for MB prediction, RMSE is 1.022 with an R² of 0.535, and for TQ prediction, RMSE is 0.591 with an R² of 0.524. These results suggest that while temporal continuity is preserved, the lack of sufficient modeling of nonlinear inter-variable interactions leads to significant information loss.

4.3.2. Dynamic Prediction Performance Analysis

A similar pattern is observed in the dynamic prediction setting (Table 3). LFTD consistently delivers the most stable and superior performance, achieving an RMSE of 0.302 and an R² of 0.941 for MB prediction, and an RMSE of 0.182 with an R² of 0.948 for TQ prediction. This demonstrates the model’s effectiveness in capturing time-lagged structures and cumulative effects when predicting firm value in the subsequent year based on prior-year financial information.
Both FT_Denoiser and Denoiser maintain relatively low RMSE values in the dynamic setting; however, their explanatory power remains lower than that of LFTD, with R² values of 0.842 and 0.901, respectively, for MB prediction.
The performance of TST_Denoiser further deteriorates in the dynamic setting, with MB prediction results of RMSE 1.112 and R² 0.450, and TQ prediction results of RMSE 0.597 and R² 0.514. These findings indicate that temporal information alone is insufficient to adequately explain inter-year variations in firm value.
Comparison with Conventional Generative Data Augmentation Methods
Using LFTD as the benchmark, comparative results with traditional data augmentation methods are presented in Table 2 (static) and Table 4(dynamic).
In the static prediction setting, DDPM exhibits moderate explanatory power, with R² values of 0.524 for MB and 0.679 for TQ; however, its RMSE values are considerably high (1.895 and 1.126, respectively).
VAE demonstrates relatively stable performance (MB: RMSE 0.744, R² 0.378; TQ: RMSE 0.462, R² 0.410), though its explanatory power remains limited.
CTGAN performs poorly even in the static setting, with low R² values (MB 0.231, TQ 0.305), indicating an inability to adequately learn the structural properties of financial data.
These limitations become more pronounced in the dynamic prediction setting. Both DDPM and CTGAN exhibit R² values that drop to zero or become negative for both MB and TQ, reflecting a severe loss of predictive stability. VAE likewise shows minimal explanatory power, with R² values remaining around 0.003. These results suggest that probabilistic generative models fail to preserve temporal dependencies and structural continuity across years.
In contrast, LFTD consistently achieves the highest R² values and maintains stable error levels across both static and dynamic settings. This performance underscores the advantage of a refinement- and reconstruction-centered approach, which jointly captures nonlinear structural relationships and cumulative temporal effects inherent in financial data. Overall, LFTD is shown to overcome the variance inflation and instability issues commonly encountered by conventional generative augmentation methods, while simultaneously improving prediction accuracy and generalization performance in firm value forecasting.
Table 5. MB, TQ Forecasting of Generative Augmentation Models using Previous-Year Features.
Table 5. MB, TQ Forecasting of Generative Augmentation Models using Previous-Year Features.
MB Prediction Performance Comparison TQ Prediction Performance Comparison
DataSet RMSE MAE R2 MAPE RMSE MAE R2 MAPE
Original 1.519 0.876 0.160 0.925 0.900 0.520 0.174 0.478
LFTD 0.302 0.112 0.941 0.109 0.182 0.067 0.948 0.060
DDPM 2.752 1.622 -0.003 1.685 1.991 1.279 -0.002 0.928
VAE 0.942 0.701 0.003 0.581 0.601 0.442 0.003 0.366
CTGAN 1.583 1.039 -0.002 0.831 1.185 0.785 -0.002 0.605

5. Conclusions

This paper has proposed LFTD, a generative framework for producing synthetic firm-level financial data that preserve both the tabular and time-series structures of financial statements. Rather than simply generating “plausible-looking” values, the framework is designed to respect accounting constraints, inter-variable dependencies, and temporal evolution. The proposed approach consists of three main stages. First, annual financial statements are embedded to obtain compact representations of firms’ financial conditions. Second, temporal patterns are learned to capture long-term dynamics at the firm level. Finally, a diffusion-based conditional generation process uses these representations to reconstruct and generate new financial time series that satisfy financial consistency and temporal coherence. In this way, the method reflects the hybrid tabular–time-series structure and strong accounting restrictions inherent in financial data. Experimental results show that the data generated by LFTD preserve statistically meaningful structures, consistently reproducing the qualitative relationships between financial indicators and firm value observed in the original data. Moreover, when these synthetic data are used as additional training samples for downstream prediction models, they improve model performance in both static and dynamic prediction settings. In particular, they enhance explanatory power and generalization while maintaining comparable error levels. These findings suggest that structure-aware, diffusion-based augmentation can serve as a practical tool for boosting model performance in financial data analysis.

6. Patents

This section is not mandatory but may be added if there are patents resulting from the work reported in this manuscript.

Supplementary Materials

Not applicable.

Author Contributions

Conceptualization, G.C.; methodology, G.C.; software, D.J.; validation, W.S.; formal analysis, H.N.; investigation, W.S; resources, H.K; data curation, H.N.; writing—original draft preparation, G.C.; writing—review and editing, H.K.; visualization, D.J.; supervision, H.N.; project administration, H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Foundational and Protective Field of Studies Support Project” at Changwon National University in 2024 and by the Ministry of Education of the Republic of Korea and the National Research Foundation of Korea (NRF) (grant number NRF-2025S1A5C3A01010737).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data and code that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

leave blank if none.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LFTD Latent Financial Time-Series Diffusion
TST Time Series Transformer
AE Autoencoder
DDPM Denoising Diffusion Probabilistic Model
VAE Variational Autoencoder
GAN Generative Adversarial Network
CTGAN Conditional Tabular GAN
TimeGAN Time-series Generative Adversarial Network
TFT Temporal Fusion Transformer
CSDI Conditional Score-based Diffusion model for Imputation
FiLM Feature-wise Linear Modulation
LSTM Long Short-Term Memory
GRU Gated Recurrent Unit
SVM Support Vector Machine
XGBoost Extreme Gradient Boosting
MSE Mean Squared Error
MAE Mean Absolute Error
MAPE Mean Absolute Percentage Error
R2 Coefficient of Determination
AUROC Area Under the Receiver Operating Characteristic curve
BCE Binary Cross Entropy

References

  1. Shorten, C.; Khoshgoftaar, T. M. A survey on image data augmentation for deep learning. Journal of Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  2. Ko, T.; Peddinti, V.; Povey, D.; Khudanpur, S. Audio augmentation for speech recognition. Proc. Interspeech, 2015; pp. 3586–3589. [Google Scholar]
  3. Goodfellow, I.; et al. Generative adversarial nets. Advances in Neural Information Processing Systems 2014, 2672–2680. [Google Scholar]
  4. Kingma, D. P.; Welling, M. Auto-encoding variational Bayes. arXiv 2014, arXiv:1312.6114. [Google Scholar]
  5. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems 2020, 33, 6840–6851. [Google Scholar]
  6. Song, Y.; et al. Score-based generative modeling through stochastic differential equations. International Conference on Learning Representations, 2021. [Google Scholar]
  7. Frid-Adar, M.; et al. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef]
  8. Kong, Z.; Ping, W.; Huang, J.; Zhao, K.; Catanzaro, B. DiffWave: A versatile diffusion model for audio synthesis. International Conference on Learning Representations, 2021. [Google Scholar]
  9. Gorishniy, Y.; Rubinstein, M.; Kleyko, D.; Yakubovskiy, A. FT-Transformer: A Transformer-based model for tabular data. arXiv 2021, arXiv:2106.01433. [Google Scholar]
  10. Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A transformer-based framework for multivariate time series representation learning. In Proceedings of the 27th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2021; pp. 2114–2124. [Google Scholar]
  11. Altman, E. I. Financial ratios, discriminant analysis and the prediction of corporate bankruptcy. The Journal of Finance 1968, 23, 589–609. [Google Scholar] [CrossRef]
  12. Breiman, L. Random forests. Machine Learning 2001, 45, 5–32. [Google Scholar] [CrossRef]
  13. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016; ACM; pp. 785–794. [Google Scholar]
  14. Cortes, C.; Vapnik, V. Support-vector networks. Machine Learning 1995, 20, 273–297. [Google Scholar] [CrossRef]
  15. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Computation 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  16. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder–decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  17. Chawla, N. V.; Bowyer, K. W.; Hall, L. O.; Kegelmeyer, W. P. SMOTE: Synthetic Minority Over-sampling Technique. Journal of Artificial Intelligence Research 2002, 16, 321–357. [Google Scholar] [CrossRef]
  18. Xu, L.; Skoularidou, M.; Cuesta-Infante, A.; Veeramachaneni, K. Modeling tabular data using Conditional GAN. Advances in Neural Information Processing Systems 2019, 7335–7345. [Google Scholar]
  19. Park, N.; Kim, S. Data synthesis based on generative adversarial networks. Proceedings of the VLDB Endowment 2018, 11, 1071–1083. [Google Scholar] [CrossRef]
  20. Yoon, J.; Jarrett, D.; van der Schaar, M. Time-series Generative Adversarial Networks. Advances in Neural Information Processing Systems 2019, 5509–5519. [Google Scholar]
  21. Esteban, C.; Hyland, S. L.; Rätsch, G. Real-valued (Medical) Time Series Generation with Recurrent Conditional GANs. arXiv 2017, arXiv:1706.02633. [Google Scholar] [CrossRef]
  22. Wen, Q.; Sun, L.; Yang, F.; Song, X.; Gao, J.; Wang, X.; Xu, H. Time Series Data Augmentation for Deep Learning: A Survey. arXiv 2020, arXiv:2002.12478. [Google Scholar]
  23. Um, T. T.; Pfister, F. M. J.; Pichler, D.; Endo, S.; Lang, M.; Hirche, S.; Fietzek, U. Data augmentation of wearable sensor data for Parkinson’s disease monitoring using convolutional neural networks. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, 2017; pp. 216–220. [Google Scholar]
  24. Huang, X.; Khetan, A.; Cvitkovic, M.; Karnin, Z. TabTransformer: Tabular data modeling using contextual embeddings. arXiv 2020, arXiv:2012.06678. [Google Scholar] [CrossRef]
  25. Lim, B.; Arık, S. Ö.; Loeff, N.; Pfister, T. Temporal fusion transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting 2021, 37, 1748–1764. [Google Scholar] [CrossRef]
  26. Tashiro, Y.; Song, J.; Song, Y.; Ermon, S.; Sohl-Dickstein, J. CSDI: Conditional score-based diffusion models for probabilistic time series imputation. Advances in Neural Information Processing Systems 2021, 34, 24804–24816. [Google Scholar]
  27. Kotelnikov, A.; Baranchuk, D.; Babenko, A. TabDDPM: Modelling tabular data with diffusion models. arXiv 2022, arXiv:2209.15421. [Google Scholar] [CrossRef]
  28. Claessens, S.; Djankov, S.; Lang, L. H. P. The separation of ownership and control in East Asian corporations. Journal of Financial Economics 2000, 58(1–2), 81–112. [Google Scholar] [CrossRef]
  29. Kim, W. S.; Wei, S.-J. NBER Working Paper No. 8967; Foreign investors and corporate governance in Korea. 2002.
  30. Eljelly, A. M. A. Liquidity–profitability trade-off: An empirical investigation in an emerging market. International Journal of Commerce and Management 2004, 14, 48–61. [Google Scholar] [CrossRef]
  31. Lee, C.-W.-J.; Li, L. Y.; Yue, H. Performance, growth and earnings management. Review of Accounting Studies 2006, 11(2–3), 305–334. [Google Scholar] [CrossRef]
  32. Dechow, P. M.; Sloan, R. G.; Sweeney, A. P. Detecting earnings management. The Accounting Review 1995, 70, 193–225. [Google Scholar]
  33. Fama, E. F.; French, K. R. The cross-section of expected stock returns. Journal of Finance 1992, 47, 427–465. [Google Scholar]
  34. Francis, J. R.; Wang, D. The joint effect of investor protection and Big 4 audits on earnings quality around the world. Contemporary Accounting Research 2008, 25, 157–191. [Google Scholar] [CrossRef]
  35. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Advances in Neural Information Processing Systems 2017, 30, 5998–6008. [Google Scholar]
  36. Hinton, G. E.; Salakhutdinov, R. R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
  37. Perez, E.; Strub, F.; de Vries, H.; Dumoulin, V.; Courville, A. FiLM: Visual reasoning with a general conditioning layer. In Proceedings of the AAAI Conference on Artificial Intelligence, 2018; 32. [Google Scholar]
Figure 1. Latent Financial Time-Series Diffusion.
Figure 1. Latent Financial Time-Series Diffusion.
Preprints 194679 g001
Figure 2. Feature-wise Tokenization and Aggregation in FT-Transformer.
Figure 2. Feature-wise Tokenization and Aggregation in FT-Transformer.
Preprints 194679 g002
Figure 3. Time Series Transformer for Contextualized Year Embeddings.
Figure 3. Time Series Transformer for Contextualized Year Embeddings.
Preprints 194679 g003
Figure 4. Stage 1 AE Architecture with Continuous/Binary Reconstruction Losses.
Figure 4. Stage 1 AE Architecture with Continuous/Binary Reconstruction Losses.
Preprints 194679 g004
Figure 5. FiLM-conditioned Transformer denoiser for latent diffusion.
Figure 5. FiLM-conditioned Transformer denoiser for latent diffusion.
Preprints 194679 g005
Figure 6. Cross-Decoder Architecture with Learnable Queries and Cross-Attention.
Figure 6. Cross-Decoder Architecture with Learnable Queries and Cross-Attention.
Preprints 194679 g006
Table 1. Definition of variables.
Table 1. Definition of variables.
VARIABLES DEFINITION OF VARIABLES
OWN The largest shareholder’s share ratio [28]
FORN Foreign investors’ share ratio [29]
SIZE Natural logarithmic value of total assets
LEV Debt-to-equity ratio
CUR Current ratio = Current assets / Current liabilities [[3[0[]
GRW Sales growth rate [[3[1[]
ROA Return on assets = Net income / Beginning total assets
ROE Return on equity = Net income / Beginning shareholders’ equity
CFO Operating cash flow ratio
PPE Proportion of tangible assets subject to depreciation [[3[2[]
INVREC Ratio of inventory and accounts receivable
MB MB = Market value of equity at year-end / Book value of equity at year-end [[3[3[]
TQ Tobin’s Q = (Total liabilities + Market value of equity) / Total assets at year-end
BIG4 A dummy variable indicating whether the auditor belongs to a Big 4 accounting firm [34]
LOSS A dummy variable equal to 1 if the firm reports a net loss in year τ (Net i n c o m e τ < 0), and 0 otherwise
YEAR Calendar year of observation, serving as a temporal index for the firm-level time series
STOCK_ID Unique firm identifier at the stock level, used to distinguish individual firms in the panel data
AGE Firm age, measured as the number of years since the firm’s establishment
Table 2. MB, TQ Prediction Performance of Proposed Denoising-Based Models.
Table 2. MB, TQ Prediction Performance of Proposed Denoising-Based Models.
MB Prediction Performance Comparison TQ Prediction Performance Comparison
DataSet RMSE MAE R2 MAPE RMSE MAE R2 MAPE
Original 1.479 0.844 0.203 0.875 0.867 0.497 0.233 0.459
LFTD 0.281 0.108 0.948 0.104 0.169 0.065 0.955 0.057
TST_Denoiser 1.022 0.551 0.535 0.439 0.591 0.324 0.524 0.266
FT_Denoiser 0.270 0.110 0.857 0.077 0.162 0.068 0.856 0.052
Denoiser 0.323 0.182 0.908 0.130 0.189 0.104 0.916 0.077
Table 3. MB, TQ Prediction Performance Comparison between Proposed and Conventional Augmentation Models.
Table 3. MB, TQ Prediction Performance Comparison between Proposed and Conventional Augmentation Models.
MB Prediction Performance Comparison TQ Prediction Performance Comparison
DataSet RMSE MAE R2 MAPE RMSE MAE R2 MAPE
Original 1.479 0.844 0.203 0.875 0.867 0.497 0.233 0.459
LFTD 0.281 0.108 0.948 0.104 0.169 0.065 0.955 0.057
DDPM 1.895 1.050 0.524 1.146 1.126 0.627 0.679 0.512
VAE 0.744 0.514 0.378 0.400 0.462 0.315 0.410 0.250
CTGAN 1.387 0.821 0.231 0.574 0.986 0.551 0.305 0.373
Table 4. MB, TQ Forecasting of Proposed Denoising-Based Models using Previous-Year Features.
Table 4. MB, TQ Forecasting of Proposed Denoising-Based Models using Previous-Year Features.
MB Prediction Performance Comparison TQ Prediction Performance Comparison
DataSet RMSE MAE R2 MAPE RMSE MAE R2 MAPE
Original 1.519 0.876 0.160 0.925 0.900 0.520 0.174 0.478
LFTD 0.302 0.112 0.941 0.109 0.182 0.067 0.948 0.060
TST_Denoiser 1.112 0.578 0.450 0.452 0.597 0.328 0.514 0.273
FT_Denoiser 0.284 0.112 0.842 0.080 0.169 0.069 0.845 0.053
Denoiser 0.337 0.185 0.901 0.136 0.195 0.106 0.911 0.0798
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated