Structural equation models (SEM) have their origins in psychology and sociology. Basically, it is a parametric framework, and in social sciences, it is often used for the determination of latent constructs. Also, in the field of economics, many authors have found it appealing.
The economy is a social science with a bold tendency to include models borrowed from other fields of science in its analysis. Furthermore, economists sometimes impose laws that are treated as if they were laws of nature. This ambition to boost economic theories to the level of the laws of nature has a broad set of implications on peoples everyday life.
On the other hand, many aspects of the economy are not measurable directly: the rule of law, the shadow economy, the transfer of technology, and many others. The shadow economy is part of the national economy that cannot be measured with the usual tools that national statistics use for the definition of macroeconomic aggregates.
In general, economic science recognizes two approaches to measuring unofficial economic activity. The first approach assumes the utilization of micro-data collected via surveys. This approach has reasonable limitations. Hence, respondents are not keen to report activities that are illegal. The second approach involves using existing macroeconomic aggregates to analyze the shadow segment of labor, tax evasion related quantities, and many other aggregates that official statistics cannot measure.
Structural equation models represent observed outcomes as linear functions of unobserved latent factors. At the same time, they model those latent factors as functions of observed exogenous drivers and structural interdependencies. Let us set this framework formally. The observed endogenous vector is generated through a loading structure acting on a vector of latent variables , subject to measurement disturbances . The latent variables themselves follow a structural system characterized by an autoregressive component governed by , exogenous influences captured by , and structural shocks . In frequentist formulations, both disturbance terms are typically assumed to follow multivariate normal distributions with parametric covariance structures.
If we want to be consistent in our intention to expose the whole procedure and to set it as a "glass box" by its essence, we should consider changing our general perspective on the issue. By contrast, Bayesian structural equation modeling treats as a random vector endowed with a prior distribution reflecting theoretical or empirical information. Inference is based on the posterior distribution obtained by combining the likelihood with the prior, typically approximated through Markov Chain Monte Carlo (MCMC) methods. Rather than relying on asymptotic normality, uncertainty is represented by the full posterior distribution, allowing identification constraints, along with scale assumptions and weakly identified directions to be reflected probabilistically rather than imposed deterministically.
In economic contexts, this choice becomes particularly critical due to several challenges that could be formulated as:
The standpoint of our analysis provides a glimpse into the issues related to the scale and, on the flip side, the exposure of the identification problems of the MIMIC shadow-economy model. The last one is best understood as a manifestation of the broader Bayesian versus frequentist inference dilemma in structural equation modeling. In the frequentist framework, identification relies on hard normalization constraints and asymptotic arguments. On the other hand, uncertainty about scale, calibration, and auxiliary assumptions is treated outside the likelihood. The straightforward question here is: why is something related to the certainty-uncertainty nexus is left out of the perspective "colorized with likelihood"? In contrast, a Bayesian formulation treats the latent scale, calibration parameters, and auxiliary information as random quantities. Not only that, Bayesian perspective gives us the explicit prior distributions as a direct control of those random quantities.
Therefore, the contribution of this paper is twofold. Substantively, it reframes the shadow-economy measurement problem as an inferential dilemma instead of a purely empirical one. On the methodological side, it shows that the Bayesian–frequentist distinction is not a matter of philosophical preference, but rather a bearer of concrete mathematical implications for identification, finite-sample behavior, and policy credibility in MIMIC-type models. By situating shadow-economy estimation within the broader theory of Bayesian versus frequentist SEM, the paper provides a unified framework that clarifies why Bayesian methods offer a principled and transparent resolution to the long-standing scale problem in MIMIC applications for the macroeconomic problems.
1.1. Literature Review
Bayesian inference has emerged in recent decades. Recently, Bayesian-freqeuntist confrontation (conditionally speaking), is exposed in the context of misspecification and something that is often referred to as decision-theoretic optimality. Therefore, many classical frequentist exercises can actually be executed as approximations of Bayes rules [
4]. When we have sample issues and the specifications are not proper, optimality properties become fragile.
On the other hand, Bayesian posteriors can remain consistent even when the model is not fully specified [
5]. Essentially, posteriors converge to pseudo-true parameters defined by Kullback–Leibler projections. This feature is particularly relevant when estimation is executed in a framework that assumes a system of equations in which exclusion restrictions are often only approximately valid. Hence, they are often backed by certain theoretical approaches that do not need to have an explicit mathematical interpretation.
In practice, full identification is difficult to achieve for models with latent constructs. Therefore, partial identification and weak identification are usually the only options to consider. In such a framework, it is plausible to treat identification probabilistically, as we do in the Bayesian framework, opposed to using hard constraints [
6]. There is no consequential observation of identification strength. It is directly reflected through posterior uncertainty. This enables MIMIC models to become more robust against calibration challenges.
The case of Bosnia and Herzegovina is particularly instructive here. Official statistics that themselves carry measurement uncertainty and a transition economy history that makes baseline assumptions questionable, truly make the calibration problem substantive, not only merely technical.
Frequentist ML breaks down in small samples, while Bayes remains robust [
7], which was later confirmed by systematic reviews [
8].
The classical frequentist SEM perspective assumes the following:
This is a vector of fixed but unknown parameters that defines the covariance structure
of the observed variables. Further, the actual estimation is executed through minimization exercise. This is done by minimizing a likelihood discrepancy function.Wishart distribution of the sample covariance matrix is the basis for this specific function [
9].
Let us take a step back and use classical SEM as a starting point. Bollen [
10] formalizes it as a general parametric system and makes it applicable to economic problems. This system is used to measure variables that cannot be measured directly. The quantity that is not directly measurable is represented by a set of directly measurable variables
. These observable variables have a joint distribution. This distribution is defined by a lower-dimensional vector of latent variables
, with
. This is straightforward from the relation for the measurable part:
where
is a loading matrix. This matrix links latent variable(s) to indicators, or variables that can be measured directly. Also,
is the covariance matrix, which can be either diagonal or block-diagonal.
The MIMIC model emerges as a special case in the SEM paradigm [
1]. It essentially nominally represents those applications of a formal mathematical procedure to problems related to the estimation of the informal economy. The shadow, informal, unofficial economy is introduced as a latent construct in the set of relations. That chunk of equations is given by the following relation:
In this set of equations,
is a vector of causal variables. Furthermore,
captures their impact on the shadow economy as a latent variable.
The MIMIC system of equations, therefore, can be represented by the following set of equations:
which together imply an observed data likelihood that depends on
only through the reduced-form covariance structure
In macroeconomic applications, the estimation of the shadow economy is recognized as a difficult task to tackle, but it is very important. Therefore, researchers cannot resist utilizing one controversial feature of SEM. The SEM framework can interpret the latent variable
as an index. Further, this index can be recognized as a time series that correlates with other macroeconomic aggregates, such as those related to fiscal or monetary policy, labor, and many others. Those are aggregates that are produced by official statistics. Let us articulate this formally. For
, the MIMIC model can be written as
where
denotes the observed causes and
stands for the observed indicators.
Now, let us write down the covariance structure of the observed data, or to be precise, its reduced form:
This depends on the latent scale only through the product:
. As a result, we can execute parameter transformation:
This can be done for any non-zero scalar
, the parameter transformation. Also, function here is left invariant. Therefore, even when the model is correctly specified and the sample size tends to infinity, the latent construct
is identified only up to an arbitrary multiplicative constant.
The reality is that in empirical studies, unofficial economic activity, or
, is not directly interpretable. Empirical exercises can only produce a relative index. This index only captures cross-sectional rankings or temporal dynamics. It is reasonable for authors to prefer the economically meaningful level measure. However, the problem is that it assumes the transformation of the estimated latent index
. This step assumes calibration information that comes out of the system. Furthermore, this calibration step lies outside the likelihood implied by (
6)–(
7) and is not identified by the MIMIC data-generating process itself. This aspect of producing an estimate of the shadow economy as a percentage of GDP is often neglected when certain interpretative features are taken for granted.
Formally, inherent scale indeterminacy is a problem in the popular application of SEM models in economics. This would not be a problem by itself if that indeterminacy were not intrinsic to the MIMIC likelihood structure. We can expose this issue in detail.
stands for the latent construct and
for the associated factor loadings.
is identified only up to an arbitrary scale factor. This is implied by (
9). Hence, the likelihood function
is invariant under the transformation
Therefore, indeterminacy results in the illusion of direct inference on economically interpretable levels of the latent variable. This is usually articulated as a percentage of GDP. On the flip side, although these depend only on monotone transformations of
, they do not affect inference on relative dynamics, temporal variation, or cross-sectional rankings.
Theorem 1 (Non-identifiability of Latent Levels in the MIMIC Model).
If we suppose that the MIMIC model is correctly specified and locally identified up to scale, as (6)–(7), then there exists no measurable function
such that identifies the level of the latent variable from the likelihood alone. For any estimator based on maximum likelihood, and for any scalar , the transformed estimator yields an observationally equivalent likelihood. Therefore our latent construct is not identified by the data-generating process.
We should stress that the non-identifiability stated in 1 is not a weakness that relates to poorly or inadequately specified models. It holds for correctly specified MIMIC systems under ideal conditions. However, researchers working with small samples face this issue acutely and more frequently than not. The reason is clear: typical time series spans rarely exceed a decent number of observations for post-transition periods, and on top of that, the span itself is effectively even shorter given the not so rare structural breaks.
Proof. The result follows directly from likelihood invariance. For any parameter vector
and scalar
, the transformation
gives the same implied covariance formulation
and, consequently, the same result, or to be more precise, the same likelihood value. Therefore, multiple distinct latent-level representations are observationally equivalent, implying that the level of
cannot be uniquely recovered from the likelihood. This non-identifiability persists asymptotically as
, since the invariance holds for the population likelihood. □
In that context, Breusch [
11] provides a formal critique of this practice, where he does not reject the MIMIC framework itself but emphasizes that MIMIC-based estimates should be interpreted as latent indices conditional on identifying assumptions, and that the uncertainty arising from calibration is typically ignored in reported inference. the formulation of this critique evolved throughout the years [
12], and provided additional arguments for Bayesian frameworks to better manage uncertainty [
13]. Parallel to that, there was a process of recognizing the issue related to the sensitivity of MIMIC results to the choice of indicators, which, from an overall perspective, aligns with our argument of model uncertainty [
14].
Recent results in the field of shadow economy estimation make an additional step by including an explicit currency demand equation. Hence, by augmenting the MIMIC system with a known relation from monetary economics, it is possible to interpret this as a methodological contribution that attempts to mitigate the identification problem highlighted in Theorem 1. Let
denote observed currency in circulation and consider the auxiliary regression
where
is a vector of control variables and
is the latent shadow economy index obtained from the MIMIC model. This attempt is based on macroeconomic theory, and the mathematical side of the story should be understood in the context that the inclusion of (
10) introduces an additional moment condition linking the latent variable to an observable monetary aggregate.
Further in that context, Dybka et al. [
15] proposes a structured hybrid identification strategy in which the MIMIC index is mapped into economically interpretable levels through a procedure termed
reverse standardization. Again, this has its mathematical interpretation as a potential solution. Let
denote the estimated latent index normalized under a conventional constraint (
). Reverse standardization constructs a level-consistent latent variable
where
is determined by matching the implied shadow economy level from (
10) to externally benchmarked values at selected reference points.
Parallel to these developments, Bayesian approaches to SEM have been advanced as an alternative. Lee [
16] formalized Bayesian SEM using Gibbs sampling. This results in the presentation of how models with latent constructs can be estimated through posterior simulation. In the same line is the work by Muthén and Asparouhov [
17], Asparouhov and Muthén [
18]. This author demonstrates that Bayesian priors can regularize ill-behaved likelihoods, including Heywood cases and near-boundary solutions.
Recent contributions on prior sensitivity and regularization in Bayesian SEM [
19,
20] emphasize that the principal advantage of Bayesian inference lies not in superior point, estimation but in the explicit probabilistic representation of identification assumptions.
In the context of time-series and macroeconomic analysis, Bayesian SEM has interesting features. Hence, when the calibration of the model is transparent through the set of priors, the model can be extended to accommodate non-stationarity and dynamic dependence structures through differencing orders and long-run relationships [
21].
Out of the many theoretical arguments that advocate for the use of Bayesian inference in the SEM framework, those related to simulation-based evidence on estimation in small samples, are essential for the main perspective of our analysis Therefore, papers that systematically investigate the relationship between sample size, model complexity, and the accuracy of Bayesian SEM estimators are an important cornerstone of our theoretical base [
22,
23]. Those results demonstrate that the Bayesian approach can induce stable parameter estimates and well-calibrated uncertainty, even in samples that would be considered critically small under conventional frequentist guidelines.
Hence, traditional SEM rules of thumb related to sample size are far more flexible in the Bayesian framework, where weakly informative priors are employed. There are empirical studies of a small-sample regime in which Bayesian SEM remains reliable while ML estimation is not. The Bayesian SEM offers a more flexible application of theory. This is especially obvious when small variance priors are used for handling parameter identification issues [
24].
To conclude, we build our analytical exercise around the anchor in the literature related to the methodological challenge that SE estimation is not only about the construction of latent indices, for which we provide some metrics for the variables that are not directly measurable. But, also the inferential treatment of scale, identification, and uncertainty under small-sample conditions. The frequentist, classical MIMIC framework relies on asymptotic arguments, hard normalizations, and post-estimation calibration. Therefore, these issues are treated as a black-box.