Preprint
Article

This version is not peer-reviewed.

Asymptotic Covariance of Factor Loadings and Uniquenesses in Classical Factor Analysis

Submitted:

15 April 2026

Posted:

21 April 2026

You are already at the latest version

Abstract
This paper derives closed-form expressions for the asymptotic covariance matrices of factor loading and uniqueness estimators obtained from several widely used factor extraction methods, including least squares, principal factor, iterative principal component, alpha factor, and image factor analysis. By treating factor solutions as implicitly defined estimators, the proposed framework characterizes the asymptotic behavior of factor loadings and uniquenesses as explicit functions of the asymptotic covariance matrix of the sample covariance or correlation matrix. This approach avoids reliance on likelihood-based information matrices, numerical differentiation, and resampling methods. Consequently, valid statistical inference is feasible under non-Gaussian sampling, serial dependence, and conditional heteroskedasticity, and can be implemented using heteroskedasticit- and autocorrelation-robust or other sandwich estimators of second moments. The framework naturally accommodates applications in which factor analysis is applied to residual covariance matrices arising from multivariate regressions, panel data models, or structural vector autoregressions (SVARs). Monte Carlo simulations demonstrate accurate finite-sample performance, and an empirical illustration shows how the proposed formulas can be implemented in practice. From an econometric perspective, the results are particularly relevant for settings in which factor structures serve as intermediate objects---such as dynamic factor models, factor-augmented regressions, and SVARs---allowing uncertainty in factor estimates to be coherently propagated into impulse response functions, forecast-error variance decompositions, and other nonlinear functionals used in structural inference.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

Factor-analytic representations play a central role in econometric modeling by providing a flexible framework for dimension reduction and structured covariance modeling in settings characterized by many correlated variables. In macroeconomics and finance, factor structures underpin dynamic factor models, structural vector autoregressions (SVARs), and variance-covariance decompositions of large systems, where a small number of latent shocks drive comovement among a high-dimensional set of observables. Beyond likelihood-based approaches, a broad class of non-maximum-likelihood (non-ML) factor extraction methods—such as principal factor, least-squares, alpha, and image factor analysis—remain widely used owing to their computational simplicity, robustness, and minimal distributional assumptions.
Let Y = ( X 1 , , X p ) denote a p-dimensional random vector with variance–covariance (or correlation) matrix Σ . Classical factor analysis postulates the decomposition
Σ = Λ Λ + Ψ ,
where Λ is a p × k loading matrix with k < p , and Ψ is a diagonal matrix of idiosyncratic variances (uniquenesses). In econometric applications, such a structure often emerges as a reduced-form representation of systems driven by clustered or common shocks. Prominent examples include structural VARs with latent structural factors, cross-sectional dependence in macroeconomic panels, and error-component models for residual covariance matrices. Consequently, estimation of Λ and Ψ is not merely of descriptive interest but constitutes a key input for structural inference, including impulse response analysis, forecast-error variance decompositions, and related functionals.
While point estimation of factor loadings is relatively straightforward across a wide range of extraction methods, formal statistical inference remains considerably more limited. In particular, closed-form standard error formulas for factor loadings and uniquenesses are readily available almost exclusively within the framework of maximum-likelihood (ML) factor analysis, where Fisher-information-based results follow directly from likelihood theory. By contrast, for non-ML extraction methods—despite their widespread use in applied econometrics and standard software packages—explicit asymptotic covariance formulas are largely unavailable or are typically obtained only indirectly through numerical differentiation or resampling procedures.
This gap has important implications for econometric practice. Non-ML factor estimators are often preferred in environments characterized by non-normality, heavy tails, or data heterogeneity, where likelihood-based assumptions are implausible. Observed variables may also be binary, categorical, ordinal, bounded, or integer-valued. Moreover, in time-series and panel settings, the sampling variability of the sample covariance or correlation matrix may itself exhibit serial dependence or heteroskedasticity, rendering distribution-free or sandwich-based inference particularly appealing. However, in the absence of closed-form expressions that link the asymptotic covariance of factor estimators to that of the sample covariance matrix, transparent inference becomes difficult to implement and the resulting estimates are challenging to interpret.
The purpose of this paper is to address this limitation by developing explicit, closed-form expressions for the asymptotic covariance matrices of factor loading and uniqueness estimators obtained from several widely used non-ML extraction methods. Specifically, we consider least-squares factor analysis, principal factor analysis, iterative principal component factor analysis, alpha factor analysis, and image factor analysis. Our approach expresses the asymptotic covariance of these estimators directly in terms of the asymptotic covariance of the sample covariance or correlation matrix, acov ( Σ ^ ) . As a result, the proposed formulas remain valid under both normal and non-normal sampling schemes and can be readily combined with existing asymptotic results for Σ ^ , including heteroskedasticity- and autocorrelation-robust estimators.
Methodologically, our analysis is based on multivariate implicit differentiation of the estimating equations that characterize each factor solution. By viewing the estimated loading matrix as a differentiable function of the sample covariance matrix, we employ the delta method to derive the asymptotic distribution of the estimators. This unified framework yields analytical expressions for the Jacobian matrices that map perturbations in Σ to changes in factor loadings and uniquenesses, resulting in tractable covariance formulas. Importantly, these results also provide the key inputs required for delta-method-based inference on rotated factor solutions, building directly on the classical contributions of Archer and Jennrich (1973), Jennrich (1973), and related studies.
From an econometric perspective, the proposed results offer key several advantages. First, they enable transparent computation of standard errors without reliance on numerical differentiation, simulation-based resampling, or likelihood-specific assumptions. Second, they apply naturally to settings in which factor analysis serves as an intermediate step—such as SVARs with latent structural shocks—thereby allowing uncertainty in factor estimates to propagate coherently into downstream quantities of interest. Third, because the formulas depend only on acov ( Σ ^ ) , they accommodate a broad class of data-generating processes, including non-Gaussian and high-frequency environments.
To illustrate the practical implementation of the proposed approach, we present an empirical application based on least-squares factor analysis and conduct a Monte Carlo simulation to assess the finite-sample performance of the derived standard errors for uniqueness estimators. The simulation results indicate that the empirical standard errors converge to their theoretical asymptotic counterparts, providing strong support for the analytical derivations.
The remainder of the paper is organized as follows. Section 2 reviews exploratory factor analysis. Section 3 derives the asymptotic covariance matrix for iterative principal component and principal factor analysis. Section 4 addresses least-squares factor analysis, while Section 5 examines alpha factor analysis and Section 6 studies image factor analysis. Section 7 derives closed-form expressions for the standard error of uniqueness estimators. Section 8 presents an empirical illustration and simulation study, and Section 9 concludes with a discussion of implications for econometric modeling and inference.

2. Structural Modeling by Exploratory Factor Analysis

For a vector of correlated random variables Y = ( X 1 , X 2 , , X p ) , common (or exploratory) factor analysis posits the existence of k mutually independent common factors that underlie the observed variables Y. Let these latent factors be jointly represented by the random vector F. The relationship between the observed variables and the latent factors is specified as
Y = c + Λ F + ϵ ,
where c is a constant vector and ϵ is a p-dimensional vector of mutually independent idiosyncratic disturbances. Assuming that the common factors have unit variance, Eq. (2) implies that the covariance structure of Y is given by Eq. (1), where the diagonal elements of Ψ correspond to the variances of the idiosyncratic disturbances, commonly referred to as uniquenesses.
Eq. (1) implies that the off-diagonal elements of the covariance matrix of Y are entirely determined by the factor loadings. In contrast, the diagonal elements reflect contributions from both common and idiosyncratic components: the common factors capture variation shared across the variables in Y, whereas the idiosyncratic disturbances represent variable-specific noise. Let λ i t denote the element in the i-th row and t-th column of Λ . The i-th diagonal element of Λ Λ ,
h i = t = 1 k λ i t 2 ,
represents the proportion of the variance of X i explained by the common factors and is referred to as the communality of variable X i .
The matrix Σ need not represent the variance-covariance matrix of a single group of observable data series; rather, it may characterize the covariance structure of residuals arising from a system of equations. To illustrate this point, consider the following structural vector autoregressive (SVAR) model:
Y t = c + s = 1 ζ A s Y t s + Λ F t + ϵ t .
Here, Y t is a p-dimensional vector of observed time series at time t; ζ denotes the lag length; F t represents k common structural shocks; and ϵ t captures p idiosyncratic shocks. The composite disturbance term Λ F t + ϵ t therefore exhibits a two-component error structure. However, the common factors F in Eq. (2) may possess a nonzero mean. For panel data, Eq. (4) can be extended to
Y i t = c i + s = 1 ζ A s Y i , t s + Λ i F t + ϵ i t ,
where Λ i is a cross-section-specific loading matrix and c i is a cross-section-specific intercept vector .
In much of the SVAR literature, idiosyncratic shocks are often ignored, effectively imposing the restriction k = p . Allowing for idiosyncratic shocks, however, substantially reduces both the number of structural shocks and the number of identifying conditions required. From a computational perspective, incorporating an error-component structure—such as a common-factor model—into SVAR systems is straightforward. For example, the parameters c and A s in Eq. (4) can be estimated by least-squares under the assumption that the variance-covariance matrix of Λ F t + ϵ t is time-invariant. As a byproduct, this procedure also delivers an estimate of the variance-covariance matrix, denoted Σ ^ , together with its sampling covariance, acov ( Σ ^ ) . In particular, when Λ F t + ϵ t follows a multivariate normal distribution, Σ ^ has a scaled Wishart distribution. Unfortunately, the uncertainty associated with Σ ^ has not been exploited in defining structural shocks nor incorporated into subsequent inference, such as impulse-response analysis.
This form of dimensionality reduction is particularly valuable in environments characterized by highly interconnected data. In macroeconomic modeling, structural shocks are typically assumed to be independent. Yet, although an economy can be examined from multiple perspectives—thereby generating a large number of interrelated observable time series—its dynamics are generally driven by only a small set of underlying shocks, such as demand, supply, geopolitical, and technological disturbances.
A wide range of factor-extraction methods has been proposed in the literature (Harman, 1976, Lawley & Maxwell, 1971, SAS, 2015), including principal-factor, maximum-likelihood, and least-squares approaches. These methods seek to recover the factor model in Eq. (1) from a given sample covariance or correlation matrix Σ ^ . From a practical perspective, it is important not only to obtain point estimates of factor loadings and uniquenesses, but also to assess the precision of these estimates and to identify substantively meaningful—often rotated—factor structures. Such assessments require knowledge of the asymptotic covariance matrices of the estimators, as well as an understanding of how these covariances are affected by the chosen rotation criterion.
The asymptotic covariance of the estimators depends on both the factor-extraction method and the covariance matrix of Σ ^ . Let n denote the sample size, and suppose that the estimator of the loading matrix can be expressed as Λ = f ( Σ ) for a given extraction method f. Assume that f is differentiable at Σ , with differential d f , which defines a linear mapping from the space of p × p symmetric matrices to the space of p × k matrices. By the delta method (Rao, 2009), the difference between n 1 / 2 d Λ and d f ( n 1 / 2 d Σ ) converges to zero in probability as n . In this setting, d Λ Λ ^ Λ and d Σ Σ ^ Σ represent deviations from the unobservable population parameters Σ and Λ , respectively.
Let Σ = [ σ i j ] and define d σ i j = σ ^ i j σ i j , where σ ^ i j denotes the sample covariance (or correlation) between X i and X j . The asymptotic covariance between elements of the estimated loading matrix Λ ^ can then be expressed as
acov ( λ ^ i r , λ ^ j s ) = m , n , x , y = 1 p λ i r σ m x acov ( σ ^ m x , σ ^ n y ) λ j s σ n y .
Similarly, let ψ = ( ψ 1 , , ψ p ) denote the vector of diagonal elements of the uniqueness matrix Ψ . The asymptotic covariance of the uniqueness estimators is given by
acov ( ψ ^ i , ψ ^ j ) = m , n , x , y = 1 p ψ i σ m x acov ( σ ^ m x , σ ^ n y ) ψ j σ n y .
In the following five sections, we derive explicit expressions for the differentials d Λ and d ψ as linear functions of d Σ for several widely used factor-extraction methods. Given the covariance matrix of Σ ^ , these expressions allow practitioners to compute standard errors for the estimated factor loadings Λ ^ and uniquenesses ψ ^ . For data drawn from a multivariate normal population, the covariance of Σ ^ is well established (Girshick, 1939). For nonnormal data, asymptotic results for sample covariances or correlations are also available in the literature (de Leeuw, 1983, Hsu, 1949). Moreover, delta-method-based inference for rotated factor loadings (Archer & Jennrich, 1973, Jennrich, 1973, Yung & Hayashi, 2001) relies on these differentials through the chain rule of differentiation. While earlier studies have examined the asymptotic properties of principal component analysis (Anderson, 1963, Girshick, 1939) and derived standard-error formulas for maximum-likelihood factor analysis (Jennrich & Thayer, 1973, Lawley, 1967), our focus is on least-squares, iterative principal component, principal, alpha, and image factor solutions as implemented in SAS (2015).
Despite the availability of maximum-likelihood (ML) factor analysis, non–ML extraction methods remain widely used and broadly implemented in statistical and econometric software (Lin et al., 2023, Raykov et al., 2026). In many empirical applications, factor analysis is applied to discrete or qualitative data, which often exhibit high-frequency dynamics, high-dimensionality, or heavy-tailed distributions. Although recent studies have emphasized robustness, high-dimensional asymptotics, and software-based inference for rotated solutions, explicit closed-form asymptotic covariance formulas for non-ML factor extraction methods remain largely unavailable. This methodological gap substantially limits formal statistical inference and rigorous model comparison.

3. Iterative Principal Component and Principal Factor Analysis

We begin by introducing notation. For a p × k matrix Λ , let Λ denote its vectorization obtained by stacking the columns of Λ , that is,
Λ = ( λ 11 , , λ p 1 , λ 12 , , λ p 2 , , λ 1 k , , λ p k ) .
When no confusion arises, we continue to use double subscripts to refer to individual elements of Λ . In particular, the ( i , r ) -th element of Λ , denoted by λ i r , corresponds to the ( ( r 1 ) p + i ) -th element of Λ . For a square matrix M, let diag ( M ) denote the vector formed by the diagonal elements of M, and let Diag ( M ) denote the diagonal matrix whose diagonal is given by diag ( M ) . We also use δ i j to denote the Kronecker delta, which equals 1 if i = j and 0 otherwise.
Both iterative principal component factor analysis and principal factor analysis construct the loading matrix Λ from the principal eigenvectors of Σ Ψ :
Σ Ψ = Λ Λ = r = 1 k λ r λ r ,
where λ r denotes the r-th column of Λ = [ λ 1 , λ 2 , , λ k ] . These columns are ordered according to the eigenvalues of Σ Ψ , with λ r corresponding to the r-th largest eigenvalue. Without loss of generality, we assume that these eigenvalues are distinct, which is typically the case for a sample variance-covariance matrix. Consequently, the inner product between any two distinct columns of Λ is zero. In particular, Eq. (5) implies
σ i i ψ i = t = 1 k λ i t 2 , i = 1 , . . . , p .
For the sample variance-covariance matrix Σ ^ , Eq. (5) may, in general, include additional items of the form λ ^ r λ ^ r for r > k . These items are omitted in accordance with the model specification in Eq. (1).
Using Eq. (5), we obtain
Σ Ψ λ r = Λ Λ λ r = s = 1 k λ s λ s λ r = s = 1 k λ s λ s λ r = λ r λ r λ r .
This result holds for any 1 r k . Now fix indices 1 i p and 1 r k . Applying Eq. (6) to the i-th components of both sides of Eq. (7) yields
j i σ i j λ j r + λ i r t = 1 k λ i t 2 = λ i r j = 1 p λ j r 2 .
Taking differentials on both sides, we obtain
r e s i z e b o x w i d t h 418.45818 p t j i σ i j d λ j r + λ j r d σ i j + t = 1 k λ i t 2 d λ i r + 2 λ i r t = 1 k λ i t d λ i t = j = 1 p λ j r 2 d λ i r + 2 λ i r j = 1 p λ j r d λ j r .
After algebraic simplification, this expression reduces to
j i σ i j 2 λ i r λ j r d λ j r + t = 1 k λ i t 2 j = 1 p λ j r 2 d λ i r + 2 λ i r t r λ i t d λ i t = j i λ j r d σ i j .
Collecting all linear equations derived above, the system can be written compactly in matrix form as
A d Λ = B d Σ ,
where the coefficient of d λ j t in the ( i , r ) -th equation is given by
A i r , j t = s = 1 k λ i s 2 s = 1 p λ s r 2 , if j = i and t = r , 2 λ i r λ i t , if j = i but t r , σ i j 2 λ i r λ j r , if j i but t = r , 0 , otherwise ,
for all 1 i , j p and 1 r , t k . Similarly, the coefficient of d σ x y in the same equation is
B i r , x y = δ i x ( 1 δ x y ) λ y r , 1 i , x , y p , 1 r k .
Finally, we obtain
d Λ = A 1 B d Σ
and, by the delta method, the asymptotic covariance matrix of Λ ^ is given by
acov Λ ^ = A 1 B acov Σ ^ A 1 B .

4. Least-Squares Factor Analysis

In least-squares factor analysis (Harman, 1976), the objective is to minimize the sum of squared off-diagonal elements of the residual covariance matrix Σ ^ Λ ^ Λ ^ . Define the objective function
g ( Λ ) = 1 y < x p σ x y t = 1 k λ x t λ y t 2 ,
that is, the sum of squares of the strictly lower-triangular elements of Σ Λ Λ .
The first-order optimality conditions yield p × k constraints on the loading matrix Λ . Specifically, we require
g ( Λ ) λ i r = 0
for all 1 i p and 1 r k . Straightforward differentiation gives
1 2 g λ i r ( Λ ) = 1 y < x p σ x y t = 1 k λ x t λ y t t = 1 k ( δ i x δ r t λ y t + δ i y δ r t λ x t ) = 1 y < x p ( δ i x λ y r + δ i y λ x r ) σ x y t = 1 k λ x t λ y t = 1 y < i λ y r σ i y t = 1 k λ i t λ y t + i < x p λ x r σ x i t = 1 k λ x t λ i t = j i λ j r σ i j t = 1 k λ i t λ j t .
For any fixed 1 i p and 1 r k , the first-order condition reduces to
j i λ j r σ i j = j i λ j r t = 1 k λ i t λ j t .
The differential form of Eq. (8) is given by
j i λ j r d σ i j + σ i j d λ j r = j i t = 1 k λ i t λ j t d λ j r + j i λ j r t = 1 k λ j t d λ i t + λ i t d λ j t .
Grouping terms and simplifying yield
r e s i z e b o x w i d t h 418.45818 p t j i λ j r d σ i j = j i t = 1 k λ i t λ j t σ i j + λ j r λ i r d λ j r + t = 1 k j i λ j r λ j t d λ i t + j i t r λ j r λ i t d λ j t ,
which completes the differential characterization of the first-order condition.
We rewrite the preceding system of equations in matrix form as
B d Σ = D d Λ ,
where the coefficient of d λ j t in the ( i , r ) -th equation is defined as
D i r , j t = z i λ z r λ z t , if j = i , z = 1 k λ i z λ j z σ i j + λ i r λ j r , if j i and t = r , λ j r λ i t , if j i and t r .
These definitions apply for all 1 i , j p and 1 r , t k . Finally, solving for d Λ yields
d Λ = D 1 B d Σ .

5. Alpha Factor Analysis

Let H be the diagonal matrix whose i-th diagonal element is h i . Clearly, h i > 0 ; otherwise, Eq. (3) would imply that the i-th row of the loading matrix Λ is a zero vector, rendering the corresponding factor redundant. Moreover, H 1 ( Σ Ψ ) H 1 is the correlation matrix for Λ F and Eq (3) implies that
d h i = 2 t = 1 k λ i t d λ i t .
To obtain the spectral decomposition of this correlation matrix, let θ r denote its r-th largest eigenvalue, and let β r be the corresponding eigenvector with length θ r 1 / 2 . Thus,
θ r = β r β r ,
and
H 1 ( Σ Ψ ) H 1 β r = θ r β r , 1 r k .
In alpha factor analysis (Kaiser & Caffrey, 1965), the loading matrix Λ = [ λ 1 , λ 2 , , λ k ] is defined as
Λ = H [ β 1 , β 2 , , β k ] .
Additionally, let Γ = [ γ 1 , γ 2 , , γ k ] , where γ r = H 1 β r . Then, for each r = 1 , , k , the following relationships—corresponding to Eqs. (10)-(12)—hold:
λ r = H β r = H 2 γ r ,
θ r = β r β r = γ r H 2 γ r = γ r λ r ,
( Σ Ψ ) γ r = ( Σ Ψ ) H 1 β r = H H 1 ( Σ Ψ ) H 1 β r = θ r H β r = γ r λ r λ r .
In the remainder of this section, we first express d Γ in terms of d Λ using Eq. (10). We then derive a matrix equation relating d Σ , d Γ , and d Λ from Eq. (12).
From Eq. (10), for any 1 i p and 1 r k , we have
λ i r = h i γ i r ,
whose differential is given by
d λ i r = h i d γ i r + 2 γ i r t = 1 k λ i t d λ i t .
Collecting these expressions yields the following matrix representation:
d Γ = E d Λ ,
where the elements of the matrix E are defined by
E i r , j t = γ i r λ j t = δ i j ( δ r t 2 γ i r λ i t ) / h i ,
for all 1 i , j p and 1 r , t k .
For any fixed 1 i p and 1 r k , the i-th component of Eq. (12) can be written as
j i σ i j γ j r + h i γ i r = λ i r j = 1 p λ j r γ j r .
Taking differentials of Eq. (14) yields
r e s i z e b o x w i d t h 418.45818 p t j i σ i j d γ j r + γ j r d σ i j + h i d γ i r + γ i r d h i = λ i r j = 1 p λ j r d γ j r + γ j r d λ j r + j = 1 p λ j r γ j r d λ i r .
Substituting Eq. (9) into the above expression and rearranging terms, we obtain
r e s i z e b o x w i d t h 418.45818 p t j i γ j r d σ i j = j i ( λ i r λ j r σ i j ) d γ j r + ( λ i r 2 h i ) d γ i r + j i λ j r γ j r d λ i r 2 γ i r t r λ i t d λ i t + λ i r j i γ j r d λ j r .
We express all the differential forms above in matrix notation as
F d Σ = G d Γ + J d Λ .
The coefficient matrices F, G, and J are defined as follows.
  • Matrix F: The coefficient of d σ x y in the ( i , r ) -th equation of Eq. (15) is
    F i r , x y = δ i x ( 1 δ i y ) γ y r .
  • Matrix G: The coefficient of d γ j t is given by
    G i r , j t = λ i r 2 h i , if j = i and t = r , λ i r λ j r σ i j , if t = r but j i , 0 , otherwise .
  • Matrix J: The coefficient of d λ j t is
    J i r , j t = s i λ s r γ s r , if j = i and t = r , 2 γ i r λ i t , if j = i but t r , λ i r γ j r , if t = r but j i , 0 , otherwise .
Combining Eqs. (13) and (15), we obtain
F d Σ = G d Γ + J d Λ = G E d Λ + J d Λ = G E + J Λ .
Consequently,
d Λ = ( G E + J ) 1 F d Σ .

6. Image Factor Analysis

Let Δ = Diag Σ 1 1 . In image factor analysis (Jöreskog, 1969), the covariance matrix Σ is modeled as
Σ = Λ Λ + τ Δ ,
where τ > 0 is an unknown scalar parameter. Within this framework, one may write Ψ = τ Δ . Jöreskog (1969) proposed a maximum-likelihood procedure for estimating this model, under which the covariance matrix of the estimators follows directly from standard maximum-likelihood theory.
For the principal-factor procedure (SAS, 2015), the loading matrix Λ is obtained from the k largest eigenvalues and their corresponding eigenvectors of Σ τ Δ . The parameter τ is simultaneously estimated as the coefficient in the regression of diag ( Σ Λ Λ ) on diag ( Δ ) , namely,
τ = diag ( Δ ) , diag ( Σ Λ Λ ) diag ( Δ ) , diag ( Δ )
where · , · denotes the standard inner product.
Let σ x y denote the element at the x-th row and y-th column of Σ 1 . Since Σ Σ 1 is the p × p identity matrix, differentiating this identity using the product rule yields
d Σ 1 = Σ 1 ( d Σ ) Σ 1 .
In particular, for each i = 1 , . . . , p ,
d σ i i = x , y = 1 p σ i x d σ x y σ y i = x , y = 1 p σ i x σ i y d σ x y .
Consequently,
d 1 σ i i = 1 ( σ i i ) 2 d σ i i = x , y = 1 p σ i x σ i y ( σ i i ) 2 d σ x y .
Eq. (17) implies that
τ diag ( Δ ) , diag ( Δ ) = diag ( Δ ) , diag ( Σ Λ Λ ) ,
and therefore
τ i = 1 p 1 σ i i 2 = i = 1 p 1 σ i i ( σ i i h i ) .
Differentiating both sides yields
i = 1 p 1 σ i i 2 d τ + i = 1 p 2 τ σ i i d 1 σ i i = i = 1 p 1 σ i i ( d σ i i d h i ) + ( σ i i h i ) d 1 σ i i .
Substituting Eq. (18) into the above expression, we obtain
r e s i z e b o x w i d t h 418.45818 p t i = 1 p 1 σ i i 2 d τ = x , y = 1 p i = 1 p σ i i h i 2 τ σ i i σ i x σ i y ( σ i i ) 2 d σ x y + i = 1 p 1 σ i i d σ i i 2 t = 1 k λ i t d λ i t .
These expressions can be written compactly as
d τ = μ d Σ + η d Λ ,
where μ and η are vectors of dimensions p 2 and p k , respectively. The coefficient of d σ x y is given by
μ x y = δ x y σ x x + i = 1 p σ i i h i 2 τ σ i i σ i x σ i y ( σ i i ) 2 i = 1 p 1 σ i i 2
and the coefficient of d λ i r is
η i r = 2 λ i r σ i i z = 1 p 1 σ z z 2 .
Next, let θ r denote the r-th largest eigenvalue of Σ τ Δ , and let λ r be the associated eigenvector with length θ r 1 / 2 . Then,
λ r λ r = θ r ,
and
Σ τ Δ λ r = θ r λ r = λ r λ r λ r .
For any fixed 1 i p , the i-th component of this equation is given by
j = 1 p σ i j λ j r τ λ i r σ i i = λ i r j = 1 p λ j r 2 .
Differentiating both sides with respect to the underlying parameters yields
j = 1 p σ i j d λ j r + λ j r d σ i j τ σ i i d λ i r λ i r σ i i d τ τ λ i r d 1 σ i i = 2 λ i r j = 1 p λ j r d λ j r + j = 1 p λ j r 2 d λ i r .
Applying Eq. (18) and collecting terms involving d Σ , d Γ , and d τ , we obtain
r e s i z e b o x w i d t h 418.45818 p t j = 1 p λ j r d σ i j τ λ i r x , y = 1 p σ i x σ i y ( σ i i ) 2 d σ x y = j i ( 2 λ i r λ j r σ i j ) d λ j r + 2 λ i r 2 σ i i + τ σ i i + j = 1 p λ j r 2 d λ i r + λ i r σ i i d τ .
The matrix representation of the above system of equations can be written as
L d Σ = P d Λ + π d τ
where L, P, and π are appropriately defined coefficient matrices. In particular, the coefficient of d σ x y in the ( i , r ) -th equation is given by
L i r , x y = δ i x λ y r τ λ i r σ i x σ i y ( σ i i ) 2 .
The coefficient of d λ j t is
P i r , j t = 2 λ i r 2 σ i i + τ σ i i + z = 1 p λ z r 2 , if j = i and t = r , 2 λ i r λ j r σ i j , if t = r but j i , 0 , otherwise ,
and the coefficient of d τ is
π i r = λ i r σ i i ,
for any 1 i , j , x , y p and 1 r , t k . Note that π is a vector of dimension p k .
Combining Eqs. (19) and (20) yields
d Λ = P + π η 1 L π μ d Σ .
Substituting this expression back into Eq. (19) gives the differential of τ in terms of d Σ :
d τ = μ + η P + π η 1 L π μ d Σ .

7. Standard Error of Uniqueness Estimator ψ ^

Assume that d Λ = M d Σ for some p k × p 2 matrix M. To derive the standard errors of the uniqueness estimator ψ ^ , we begin with the generic relationship from Eq. (6): ψ i = σ i i t = 1 k λ i t 2 . Differentiating both sides yields
d ψ i = d σ i i 2 t = 1 k λ i t d λ i t , i = 1 , 2 , , p .
Eq. (22) can be written more compactly in matrix form as
d ψ = T d Σ + Z d Λ ,
where the coefficient of d σ x y in the i-th equation of Eq. (23) is given by
T i , x y = δ i x δ i y
and the coefficient of d λ j t in the same equation is
Z i , j t = 2 δ i j λ j t ,
for all 1 i , j , x , y p and 1 t k . Consequently, we obtain
d ψ = ( T + Z M ) d Σ .
When Σ is a correlation matrix, d σ i i = 0 , and the expression simplifies further to d ψ = Z M d Σ .
In image factor analysis, the relationship ψ ^ i = σ ^ i i t = 1 k λ ^ i t 2 may no longer hold. Instead, we rewrite Eq. (21) as d τ = ϕ d Σ for some vector ϕ . Using the identity ψ i = τ σ i i and applying Eq. (18), we obtain
d ψ i = d τ σ i i + τ d 1 σ i i = 1 σ i i x , y = 1 p ϕ x y d σ x y + τ x , y = 1 p σ i x σ i y ( σ i i ) 2 d σ x y ,
for all 1 i p . In matrix form, this expression can be written compactly as
d ψ = Q d Σ ,
where the coefficient of d σ x y in the i-th equation is given by
Q i , x y = ψ i σ x y = σ i i ϕ x y + τ σ i x σ i y σ i i 2 , 1 i , x , y p .

8. An Empirical and Simulation Study

In this section, we apply the least-squares factor analysis method to a sample correlation matrix consisting of nine variables measured on 211 subjects. The data were originally reported in page 43 of Lawley and Maxwell (1971), and the sample is assumed to be drawn from an approximately multivariate normal population. For simplicity, we set the number of common factors to k = 2 . This choice is supported by both Horn (1965)’s parallel analysis and Velicer (1976)’s minimum average partial test . Estimates of the factor loading matrix Λ ^ and the uniqueness vector ψ ^ are obtained using the iterative procedure described by Harman (1976).
To compute the asymptotic covariances of these estimators, we first apply Girshick (1939)’s formula (3.23) to derive the covariances of the sample correlation coefficients. We then use the results developed in Section 4 and Section 7 to compute the asymptotic covariance matrix of Λ ^ , as well as the standard errors of ψ ^ . To illustrate the effect of factor rotation, we orthogonally rotate Λ ^ using the varimax criterion and apply the method of Jennrich and Thayer (1973) to obtain the standard errors of the rotated factor loadings. Table 1 reports the parameter estimates for the two-factor least-squares model; standard errors are reported in parentheses.
To verify the theoretical derivation of the asymptotic covariance, acov ( ψ ^ ) , presented in Section 7, we conduct a simulation study. The sample correlation matrix reported above is treated as the population correlation matrix Σ . Using this matrix, we generate a large number m of random correlation matrices from a Wishart distribution with parameter ( Σ , 211 ) . For each simulated correlation matrix, the uniqueness parameters are estimated using the least-squares method with k = 2 .
Based on the resulting collection of simulated estimates of ψ , we compute empirical standard errors for the uniqueness estimates. For m = 50 , 100 , 500 , 1000 , 2000 , the corresponding empirical standard errors are reported in Table 2. The final column of the table presents the theoretical asymptotic standard errors computed using the formulas developed in Section 7. The results demonstrate clear convergence of the empirical standard errors toward their theoretical counterparts, although the convergence occurs at a relatively slow rate.

9. Conclusion with Discussions

Recent years have witnessed renewed interest in formal statistical inference for factor-analytic models. This resurgence is driven by several practical considerations, including the prevalence of non-normal data, heightened concerns about robustness, and the growing complexity of covariance structures encountered in empirical applications. Broadly speaking, the contemporary literature can be organized into three main strands. The first strand emphasizes likelihood-based approaches to factor analysis and covariance structure modeling, building on classical frameworks and their subsequent extensions (Jennrich & Thayer, 1973, Lawley, 1967, Raykov et al., 2026). A second strand focuses on high-dimensional and dynamic factor models, in which both cross-sectional and temporal dimensions expand, giving rise to new challenges for estimation and inference (Bai, 2003, Fan et al., 2021). The third strand concentrates on software-oriented implementations of standard error estimation for rotated factor solutions, typically relying on numerical differentiation or sandwich-type approximations (Archer & Jennrich, 1973, Yung & Hayashi, 2001).
In contrast, explicit asymptotic covariance formulas for widely used non-maximum-likelihood extraction methods—such as least-squares, principal, alpha, and image factor analysis—remain relatively scarce, despite their continued prevalence in empirical research and standard statistical software (Harman, 1976, SAS, 2015). This paper helps bridge this gap by deriving closed-form, analytically tractable asymptotic covariance expressions for several classical no-ML factor solutions. By casting these results within a unified implicit-differentiation framework, the proposed approach accommodates non-normal data and can be readily integrated with existing delta-method results for factor rotation (Archer & Jennrich, 1973, Jennrich, 1973, Yung & Hayashi, 2001). From a practical standpoint, the framework enables transparent computation of standard errors without reliance on numerical differentiation or resampling methods. From a methodological perspective, it complements recent advances in robust and high-dimensional factor modeling by providing a rigorous inferential foundation for classical extraction methods that remain central to applied factor analysis.
In the preceding sections, the derived asymptotic covariances of Λ ^ and ψ ^ are conditional on the assumed number of factors. Consequently, anomalous standard errors may reflect model misspecification or systematic errors—such as an incorrect choice of factor dimensionality—rather than genuine sampling variability. In particular, such distortions may arise from either over-factoring (when k is chosen too large) or under-factoring (when k is chosen too small).
For expositional simplicity, we do not explicitly impose the symmetry condition σ i j = σ j i . In practical implementations, however, exploiting this property is important not only for reducing memory requirements but also for improving computational efficiency, especially when working with large-scale datasets.
For the SVAR model in Eq. (4), target or Procrustean rotation is unnecessary when the structural shocks F t are properly aligned with the endogenous variables Y t in the loading matrix. Such alignment naturally imposes identifying restrictions—such as zero or sign restrictions—on the elements of Λ , which may require corresponding adjustments in its estimation. Nevertheless, the formulas developed in this paper remain valid, as they treat the estimated Λ as given inputs. Moreover, these formulas not only assess the precision of coefficient estimates associated with structural shocks, but also facilitate the evaluation of uncertainty in impulse response functions, forecast-error variance decompositions, historical decompositions, and the estimated factors F t .
From an econometric perspective, the results developed in this paper are particularly relevant for empirical settings in which factor extraction serves as an intermediate step rather than an end in itself. In applied macroeconomics and finance, factor-analytic structures commonly arise in dynamic factor models, large Bayesian or classical VARs, factor-augmented regressions, and error-component representations of high-dimensional systems. Within these frameworks, factor loadings and idiosyncratic components are typically treated as nuisance parameters, and the uncertainty associated with their estimation is rarely propagated into subsequent stages of inference.
The explicit asymptotic covariance formulas derived in this paper enable researchers to systematically account for factor-estimation uncertainty in econometric inference. In particular, they facilitate the construction of delta-method-based standard errors for impulse response functions, forecast-error variance decompositions, counterfactual simulations, and other nonlinear functionals that depend on estimated factors. This consideration is especially important in structural VAR applications with latent common shocks, where factor estimates play a central role in both the identification and interpretation of structural disturbances.
Moreover, because the proposed expressions take the covariance of the sample covariance or correlation as inputs, they remain valid under heteroskedasticity, serial dependence, and non-Gaussian sampling schemes—features that are pervasive in macroeconomic and financial data. As a result, the framework naturally aligns with heteroskedasticity- and autocorrelation-robust covariance estimators widely used in econometrics, without relying on likelihood-based assumptions. For example, when the data are multivariate binary, the asymptotic covariance matrix, acov ( Λ ^ ) , remains well defined. More generally, for categorical data—common in applied factor-analysis settings—the computation of acov ( Λ ^ ) is unlikely to pose practical difficulties.
Finally, the availability of closed-form expressions provides a valuable diagnostic tool for model specification in econometric practice. Abnormally large estimated standard errors may indicate misspecification of factor dimensionality, weak factor structure, or inadequate separation between common and idiosyncratic components. Moreover, a large proportion of statistically insignificant coefficients in Λ ^ may simply reflect over-factoring. From this perspective, asymptotic covariance analysis not only supports formal statistical inference but also offers practical guidance for model selection and validation in factor-based econometric models.

Author Contributions

All authors have equally contributed to conceptualization, methodology, and validation. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Σ Unobservable population variance-covariance matrix of the random vector Y
Σ ^ Sample variance-covariance matrix of the random vector Y
Λ Loading matrix in the variance-covariance structure model Σ = Λ Λ + Ψ
Ψ Uniqueness, a diagonal matrix
Λ ^ , Ψ ^ Estimates of the loading matrix and uniqueness, respectively
p and k Dimensions of Λ ; p is also the dimension for Y and, consequently, for Σ and Ψ
δ i j The Kronecker delta, that equals 1 when i = j and 0 otherwise
Λ The vectorization of the matrix Λ , ordered by columns. Similar for Σ
ψ The diagonal vector of Ψ
θ r The r-th largest eigenvalue of a symmetric matrix
σ x y The element at the x-th row and y-th column of Σ
σ x y The element at the x-th row and y-th column of Σ 1 , the inverse of Σ
diag ( M ) The vector of diagonal elements of the square matrix M
Diag ( M ) The diagonal matrix with diagonal elements in M
ACOV Asymptotic covariance
SVAR Structural Vector AutoRegressive model

References

  1. Anderson, T.W. 1963. Asymptotic theory for principal components analysis. Annals of Mathematical Statistics 34: 122–148. [Google Scholar] [CrossRef]
  2. Archer, C.O., and R.I. Jennrich. 1973. Standard errors for orthogonally rotated factor loadings. Psychometrika 38: 581–592. [Google Scholar] [CrossRef]
  3. Bai, J. 2003. Inferential theory for factor models of large dimensions. Econometrica 71, 1: 135–171. [Google Scholar] [CrossRef]
  4. de Leeuw, J. 1983. Models and methods for the analysis of correlation coefficients. Journal of Econometrics 22: 113–137. [Google Scholar] [CrossRef]
  5. Fan, J., K. Wang, Y. Zhong, and Z. Zhu. 2021. Robust high-dimensional factor models with applications to statistical machine learning. Statistical Science 36, 2: 303–327. [Google Scholar] [CrossRef] [PubMed]
  6. Girshick, M.A. 1939. On the sampling theory of roots of determinantal equations. Annals of Mathematical Statistics 10: 203–224. [Google Scholar] [CrossRef]
  7. Harman, H.H. 1976. Modern Factor Analysis (Third Edition). Chicago: University of Chicago Press. [Google Scholar]
  8. Horn, J. 1965. A rationale and test for the number of factors in factor analysis. Psychometrika 30: 179–185. [Google Scholar] [CrossRef] [PubMed]
  9. Hsu, P.L. 1949. The limiting distribution of functions of sample means and application to testing hypotheses. Proceedings of the First Berkeley Symposium in Mathematical Statistics and Probability; pp. 359–402. [Google Scholar]
  10. Jennrich, R.I. 1973. Standard errors for obliquely rotated factor loadings. Psychometrika 38: 593–604. [Google Scholar] [CrossRef]
  11. Jennrich, R.I., and D.T. Thayer. 1973. A note on Lawley’s formulas for standard errors in maximum likelihood factor analysis. Psychometrika 38: 571–580. [Google Scholar] [CrossRef]
  12. Jöreskog, K.G. 1969. Efficient estimation in image factor analysis. Psychometrika 34: 51–75. [Google Scholar] [CrossRef]
  13. Kaiser, H.F., and J. Caffrey. 1965. Alpha factor analysis. Psychometrika 30: 1–14. [Google Scholar] [CrossRef] [PubMed]
  14. Lawley, D.N. 1967. Some new results in maximum likelihood factor analysis. Proceedings of the Royal Society of Edinburgh 67A: 256–264. [Google Scholar] [CrossRef]
  15. Lawley, D.N., and A.E. Maxwell. 1971. Factor Analysis as a Statistical Method. New York: American Elsevier. [Google Scholar]
  16. Lin, T.I., I.A. Chen, and W.L. Wang. 2023. A robust factor analysis model based on the canonical fundamental skew-t distribution. Stat Papers 64: 367–393. [Google Scholar] [CrossRef]
  17. Rao, C.R. 2009. Linear Statistical Inference And Its Applications (Second Edition). New York: Wiley. [Google Scholar]
  18. Raykov, T., S. Penev, and B. Zhang. 2026. Fitting covariance structure models with arbitrary observed distributions: strong convergence of the asymptotically distribution-free estimator. Structural Equation Modeling: A Multidisciplinary Journal 33:1: 32–38. [Google Scholar] [CrossRef]
  19. SAS. 2015. SAS/STAT® 14.1 User’s Guide The FACTOR Procedure. Cary, NC: SAS Institute Inc. [Google Scholar]
  20. Velicer, W.F. 1976. Determining the number of components from the matrix of partial correlations. Psychometrika 41: 321–327. [Google Scholar] [CrossRef]
  21. Yung, Y.F., and K. Hayashi. 2001. A computational efficient method for obtaining standard error estimates for the promax and related solutions. British Journal of Mathematical and Statistical Psychology 54: 125–138. [Google Scholar] [CrossRef] [PubMed]
Table 1. Estimates and Standard Errors for Two-Factor Least-Squares Factor Analysis
Table 1. Estimates and Standard Errors for Two-Factor Least-Squares Factor Analysis
Unrotated Factors Rotated Factors
Variable I II I II Uniqueness
X 1 .6639(.0397) .3285(.0502) .6745(.0453) .3063(.0538) .4512(.0557)
X 2 .6879(.0381) .2388(.0545) .6202(.0487) .3815(.0555) .4698(.0539)
X 3 .4956(.0536) .2831(.0657) .5328(.0583) .2047(.0658) .6743(.0599)
X 4 .8470(.0249) -.3037(.0372) .3007(.0381) .8481(.0295) .1904(.0412)
X 5 .7035(.0392) -.3179(.0637) .1990(.0493) .7459(.0399) .4040(.0551)
X 6 .8037(.0297) -.3581(.0659) .2312(.0404) .8490(.0353) .2258(.0526)
X 7 .6686(.0440) .3889(.0654) .7242(.0412) .2717(.0513) .4018(.0546)
X 8 .4236(.0609) .2552(.0813) .4656(.0639) .1666(.0695) .7555(.0578)
X 9 .7718(.0347) .4398(.0598) .8289(.0328) .3194(.0444) .2109(.0443)
Table 2. Simulation of Standard Errors for Uniqueness Estimates
Table 2. Simulation of Standard Errors for Uniqueness Estimates
Empirical S.E. of Uniqueness for Simulated Correlations Theoretical
Uniqueness m = 50 100 500 1000 2000 S.E.
ψ 1 .0499849 .0511051 .0549498 .0560314 .0556181 .0556623
ψ 2 .0513352 .0520701 .0545346 .0555878 .0549598 .0538818
ψ 3 .0666649 .0660852 .0599577 .0603062 .0598762 .0598473
ψ 4 .0405467 .0433619 .0432832 .0428048 .0415737 .0411420
ψ 5 .0481342 .0538724 .0524052 .0530417 .0544235 .0550669
ψ 6 .0501591 .0543790 .0541692 .0550827 .0546638 .0526150
ψ 7 .0524787 .0565299 .0547573 .0553479 .0546232 .0546297
ψ 8 .0507511 .0562004 .0576808 .0582864 .0580424 .0578393
ψ 9 .0394781 .0409714 .0430909 .0448280 .0438135 .0443326
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated