Preprint
Article

This version is not peer-reviewed.

Estimation for Longitudinal Varying Coefficient Partially Nonlinear Models Based on QR Decomposition

A peer-reviewed article of this preprint also exists.

Submitted:

29 October 2025

Posted:

10 November 2025

You are already at the latest version

Abstract
Aiming at the estimation efficiency problem of multicollinearity and longitudinal data correlation in the varying coefficient partially nonlinear models, a method based on QR decomposition and quadratic inference function (QIF) is proposed to obtain the orthogonality estimation of parameter components and varying coefficient functions. QR decomposition eliminates the pathology of the design matrix, and combines the adaptive weighting of the relevant structures within the group by QIF to effectively capture the complex correlation structure of longitudinal data. The theoretical analysis proves the asymptotic nature of the estimator, and the efficiency of the estimation method proposed in this paper is verified by simulation experiments.
Keywords: 
;  ;  ;  

1. Introduction

In recent years, longitudinal data have attracted much attention due to their wide applications in fields such as biomedicine, economics and social sciences. This type of data is obtained through repeated observations of the same subject at different time points. As a result, multiple measurement results of the same subject exhibit temporal or spatial dependence, which makes the traditional independent and identically distributed assumption no longer applicable. To address this issue, the varying coefficient partially nonlinear model has become one of the research hotspots, as it combines the interpretability of parametric models and the flexibility of nonparametric models. Li and Mei [1] put forward a profile nonlinear least squares estimation approach to estimate the parameter vector and coefficient function vector of the varying coefficient partially nonlinear model, and further derived the asymptotic characteristics of the obtained estimators.
Important progress has been made in model estimation in existing studies. Yu and Zhu et al. [2] constructed robust estimators for the partially linear additive model under functional data. Yan and Tan et al. [3] explored the empirical likelihood inference targeting the partially linear errors-in-variables model under longitudinal data. Varying coefficient partially nonlinear model based on mode regression, a robust two-stage method for estimation and variable selection was proposed by Xiao and Liang [4]. Zhao and Zhou et al. [5] put forward a new orthogonality-based empirical likelihood inference method through orthogonal estimation techniques and empirical likelihood inference methods, which is used to estimate the parametric and nonparametric components in a class of varying coefficient partially linear instrumental variable models for longitudinal data. Liang and Zeger [6] applied the generalized estimating equation to longitudinal data analysis, which effectively handles the correlation issue of longitudinal data. For the varying coefficient partially nonlinear quantile regression model under randomly left-censored data, Xu and Fan et al. [7] introduced a three-stage estimation weighting method.
However, significant challenges still exist in the existing methods for estimating the varying coefficient partially nonlinear model. For instance, the estimation of nonlinear parameters is easily disturbed by nonparametric components and longitudinal correlation structures, leading to reduced estimation efficiency; when there is multicollinearity among explanatory variables, the ill-posed nature of the design matrix will further decrease the stability of estimation. Such issues severely limit the performance and practicality of the model in practical applications.
To address the above issues, existing studies have proposed a variety of improved methods. For example, Qu and Lindsay et al. [8] used the quadratic inference function to improve the estimating equation, and this method maintains optimal performance even when the working correlation structure is misspecified. Bai and Fung et al. [9] applied the quadratic inference function to handle longitudinal data, and the results demonstrate that the proposed estimation method exhibits excellent asymptotic properties. Schumaker [10] used B-spline basis functions to approximate the varying coefficient part, which improves the computational efficiency. Wei and Wang et al. [11] proposed a model averaging method for linear models with randomly missing response variables. Jiang and Ji et al. [12] proposed an estimation method for the varying coefficient partially nonlinear model based on the exponential squared loss function. Xiao and Chen [13] proposed a procedure for local bias-corrected cross-sectional nonlinear least squares estimation. For left-censored data, Xu and Fan et al. [14] established a quantile regression estimation and variable selection method for the varying coefficient partially nonlinear model. Based on the varying coefficient partially nonlinear model where the nonparametric part has measurement errors, Qian and Huang [15] discussed a modified profile least squares estimation method. Wang and Zhao et al. [16] proposed an inverse probability weighted cross-sectional least squares method to estimate unknown parameters and nonparametric functions for the varying coefficient partially nonlinear model.
This paper proposes an orthogonal estimation framework that integrates QR decomposition and the QIF. QR decomposition can effectively eliminates the ill-posed nature of the design matrix and improves numerical stability. The QIF method avoids the limitations of traditional generalized estimating equations by adaptively weighting the intra-group correlation structure, significantly enhancing estimation efficiency. This study not only provides a new theoretical tool for longitudinal data analysis but also offers a feasible solution to complex modeling problems in practical applications.
The structure of this paper is arranged as follows: Section 2 introduces the model specification and estimation method, including the specific implementation of QR decomposition and QIF; Section 3 discusses the asymptotic properties of the estimators; Section 4 verifies the superiority of the proposed method through simulation experiments; The proof process of the main conclusions is provided in the Appendix.

2. Models and Methods

Consider the varying coefficient partially nonlinear model introduced by Li and Mei [1]:
Y = X T α U + g Z , β + ε ,
where, Y R is the response variable, X R p , U R , and Z R r are covariates, α U = α 1 U , α 2 U , , α p U T and α k U , k = 1 , 2 , , p are unknown smooth functions, g ( · , · ) is a nonlinear function with a known form, β = β 1 , β 2 , , β q T denotes a q-dimensional unknown parameter vector, and ε is the model error with mean zero and variance σ 2 .

2.1. Estimation of Parameter Vector

Considering the model (1) under longitudinal data, suppose the j-th observation of the i-th individual satisfies:
Y i j = X i j T α U i j + g Z i j , β + ε i j , i = 1 , 2 , , n , j = 1 , 2 , , n i ,
among them, ε i j has a mean of zero, and X i j R p , Z i j R r , U i j are covariates.
Based on the idea of Schumaker [10], the unknown function α k · , k = 1 , 2 , , p can be approximated using basis functions. The B-spline basis functions are denoted by the vector B u = B 1 u , B 2 u , , B L u T . The dimension L is defined as L = K + M , where K and M represent the number of interior knots and the order of the spline, respectively. Then α k u can be approximately expressed as
α k u B u T γ k , k = 1 , 2 , , p ,
here γ k = γ k 1 , γ k 2 , , γ k L T are the coefficient vector for the B-spline basis functions. model (2) can be expressed as
Y i j W i j T γ + g Z i j , β + ε i j , i = 1 , 2 , , n , j = 1 , 2 , , n i ,
define W i j = I p B U i j · X i j , where ⊗ is the Kronecker product, and let γ = γ 1 T , γ 2 T , , γ p T T , Y i = Y i 1 , Y i 2 , , Y i n i T , W i = W i 1 , W i 2 , , W i n i T , g ( Z i , β ) = g ( Z i 1 , β ) , g ( Z i 2 , β ) , ,   g ( Z i n i , β ) T , ε i = ε i 1 , ε i 2 , , ε i n i T . Then model (4) can be expressed as
Y i W i γ + g Z i , β + ε i , i = 1 , 2 , , n .
Suppose that for all i = 1 , 2 , , n , the matrices W i have full column rank, their QR decomposition can be expressed as:
W i = Q i R i 0 , i = 1 , 2 , , n ,
where Q i is an n i × n i orthogonal matrix, R i is a p L × p L triangular matrix, and 0 denotes a n i p L × p L zero matrix. The matrix Q i can be divided into two parts as Q i = Q i 1 , Q i 2 , where Q i 1 is a n i × p L matrix and Q i 2 is a n i × n i p L matrix. Substitute the decomposition of Q i into (6) to obtain W i = Q i 1 R i , from the properties of orthogonal matrices, Q i 2 T Q i 1 = 0 can be derived, and then Q i 2 T W i = Q i 2 T Q i 1 R i = 0 is obtained. Multiplying both sides of Equation (5) by Q i 2 T , and we get:
Q i 2 T Y i Q i 2 T g Z i , β + Q i 2 T ε i , i = 1 , 2 , , n .
It is easy to see that model (7) is a regression model containing only unknown parameters. Following Liang [6], the generalized estimating equation for β can be defined as:
i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 V i 1 Q i 2 T ( Y i g ( Z i , β ) ) = 0 ,
among them, g ( 1 ) ( Z i , β ) = ( g ( 1 ) ( Z i 1 , β ) , , g ( 1 ) ( Z i n i , β ) ) T , V i = Q i 2 T Σ i Q i 2 , Σ i is the covariance matrices of Y i , and the structure of V i can also be expressed as V i = A i 1 2 R ρ A i 1 2 according to the method of Liang [6], where A i is a diagonal matrix, R ρ is a working correlation matrix, and ρ is a correlation parameter. Since a consistent estimator of ρ is not always available in practice, we adopt the QIF method to approximate the working correlation matrix R 1 ( ρ ) through a set of basis matrices, thereby avoiding direct specification of the correlation structure. Let M 1 , M 2 , , M s be a set of basis matrices and a 1 , a 2 , , a s be coefficients satisfying:
R 1 ( ρ ) a 1 M 1 + a 2 M 2 + + a s M s .
Substituting (9) into (8), we obtain:
i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 ( a 1 M 1 + + a s M s ) A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) = 0 .
Define the extended score function as follows:
G n ( β ) = 1 n i = 1 n η i ( β ) 1 n i = 1 n g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M 1 A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M 2 A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) g ( 1 ) ( Z i , β ) T Q i 2 A i 1 2 M s A i 1 2 Q i 2 T ( Y i g ( Z i , β ) ) .
Thus, the QIF for β can be defined as:
M n ( β ) = G n T ( β ) Φ n 1 ( β ) G n ( β ) ,
where Φ n ( β ) = 1 n n i = 1 η i ( β ) η i T ( β ) , in this case, the estimator β ^ of β is obtained by minimizing the objective function M n ( β ) :
β ^ = arg min β M n ( β ) .

2.2. Estimation of the Coefficient Functions

The QIF effectively handles the correlation in longitudinal data and improves estimation efficiency by constructing a set of estimating equations. Therefore, after obtaining the initial estimates of the parameters β , we also use the QIF method to estimate the coefficient functions α k u , k = 1 , 2 , , p .
Substitute the estimator of into Equation (5), resulting in:
Y ˜ i W i γ + ε i , i = 1 , 2 , , n ,
where Y ˜ i = Y i g i Z i , β ^ , assuming that the covariance matrix of Y ˜ i is Ψ i , and its structure is expressed following Liang [6] as Ψ i = A i 1 2 R ρ A i 1 2 , where A i is a diagonal matrix, R ρ is a working correlation matrix, and ρ is a correlation parameter. Assuming V i is known, construct the following estimating equation:
n i = 1 W i T Ψ i 1 Y ˜ i W i γ = 0 .
In practical applications, V i is usually unknown. Based on this, we still adopt the QIF method and use M 1 , M 2 , , M s to approximate R i 1 , then we have:
n i = 1 W i T A i 1 2 a 1 M 1 + a 2 M 2 + + a s M s A i 1 2 Y i W i γ = 0 .
Define the extended score function as follows:
G n ( γ ) = 1 n i = 1 n φ i ( γ ) 1 n i = 1 n W i T A i 1 2 M 1 A i 1 2 Y i W i γ W i T A i 1 2 M 2 A i 1 2 Y i W i γ W i T A i 1 2 M s A i 1 2 Y i W i γ .
Thus, the QIF for γ can be defined as:
Q n γ = G n T γ C n 1 ( γ ) G n γ ,
where C n ( γ ) = 1 n n i = 1 φ i γ φ i T γ . In this case, the estimate γ ^ of γ can be obtained by minimizing the objective function Q n ( γ ) :
γ ^ = arg min Q n γ .
Thus, the estimate of the coefficient function α k u can be expressed as:
α ^ k u B u T γ ^ k , k = 1 , 2 , , p .
where γ ^ k is the component corresponding to the k-th coefficient function in γ ^ .

3. Main Conclusions

This section studies the asymptotic properties of the estimators β ^ and α ^ k u , k = 1 , 2 , , p , assuming that β 0 and α 0 ( · ) are the true values of β and α ( · ) respectively, γ 0 is the true value of γ , and α k 0 ( · ) and γ k 0 are the k-th components of α 0 ( · ) and γ respectively. First, we present some common regularity conditions in longitudinal data analysis:
(C1) The random variable U has a bounded support Ω , and its probability density function f u · is positive and has continuous second-order derivatives.
(C2) The varying coefficient function α 1 u , , α p u is continuously differentiable of order r on 0 , 1 , where r > 2 .
(C3) For any Z, g Z , β is continuous with respect to β , and g Z , β has continuous partial derivatives of order r.
(C4) sup i E ( ε i ε i T ) < holds, and there exists some δ > 0 satisfying E ( ε i 2 + δ ) < .
(C5) The covariates X i j and Z i j are assumed to satisfy the following conditions: sup i j E X i j 4 < , sup i j E Z i j 4 < , i = 1 , 2 , n , j = 1 , 2 , , n i .
(C6) Let t 1 , , t k be interior nodes on 0 , 1 . Furthermore, let t 0 = 0 , t k + 1 = 1 , ξ i = t i t i 1 , then there exists a constant C 0 such that: max ξ i min ξ i C 0 , max ξ i + 1 ξ i = o K 1
(C7) Define P i k = Q i 2 A i 1 2 M k A i 1 2 Q i 2 T , then we have:
1 n i = 1 n E g ( 1 ) ( Z i , β ) T P i k g ( 1 ) ( Z i , β ) P D k , k = 1 , 2 , , s .
1 n i = 1 n E g ( 1 ) ( Z i , β ) T P i l T V i P i l g ( 1 ) ( Z i , β ) P Ω l k , l , k = 1 , 2 , , s .
Among them, D k and Ω l k are constant matrices, and P denotes convergence in probability. Define Ω 0 = ( Ω l k ) p s × p s and Γ 0 = ( D 1 T , , D s T ) T , additionally, assume that Ω 0 and Γ 0 T Ω 0 1 Γ 0 are both invertible.
It is noted that, conditions (C1) to (C5) are common conditions in varying coefficient partially nonlinear models components, Condition (C6) indicates that t 0 , , t k + 1 is a uniform partition sequence over the interval 0 , 1 , and Condition (C7) is used for subsequent proofs.
Theorem 1. 
Under conditions (C1) to (C7), and when n , it follows that:
n ( β ^ β 0 ) L N ( 0 , ) ,
where the matrix = ( Γ 0 T Ω 0 1 Γ 0 ) 1 , L denotes convergence in distribution.
Theorem 2. 
Under conditions (C1) to (C7), the number of nodes K = O n 1 2 r + 1 , and when n , it follows that:
α ^ k · α k 0 · = O p n r 2 r + 1 , k = 1 , 2 , , p .
where · denotes the L 2 norm of the function.

4. Simulation Study

This section evaluates the finite-sample performance of the proposed orthogonality estimation method based on QR decomposition and quadratic inference functions in varying coefficient partially nonlinear models through a Monte Carlo simulation study. Consider the following model:
Y = X T α U + g Z , β + ε ,
among them, the covariates X , Z = ( Z 1 , Z 2 ) T both follow normal distributions, the nonlinear function g ( Z , β ) is defined as g ( Z , β ) = β 1 Z 1 + β 2 exp ( Z 2 ) , with the parameter vector β = ( β 1 , β 2 ) T = ( 0.8 , 0.5 ) T . Additionally, U U ( 0 , 1 ) , the coefficient function α ( U ) = sin ( 2 π U ) , the error term ε follows an AR(1) process, and its structure is: ε i , j = 0.5 ε i , j 1 + v i , j , where v i , j N ( 0 , 0.5 1 0.5 2 ) to ensure the error variance is 0.5.
The sample size is set to n = 100 , 150 , 200 , for the i-th subject, the number of repeated measurements is m i 8 , and 1000 simulation runs are conducted for each case. The method combining QR decomposition and QIF proposed in this paper (OQIF) is compared with the profile nonlinear least squares method (PNLS) proposed by Li and Mei [1]. After sorting out, the following Table 1 and Table 2 are obtained.
It can be seen from the results in Table 1 that as the sample size increases, the bias and standard deviation of both methods decrease. However, the bias and standard deviation of the OQIF method are smaller, indicating more accurate estimation. Although a larger sample size improves the estimation accuracy of both methods, the OQIF method performs better in bias control.
The results in Table 2 show that the average length of the confidence interval of the OQIF method is significantly shorter than that of the PNLS method, demonstrating higher estimation efficiency, the coverage rate of the OQIF method is closer to the ideal 95%, while although the coverage rate of the PNLS method has improved, it is still below 95%.
In conclusion, when the sample size increases from 100 to 200, the OQIF method demonstrates higher estimation accuracy and stronger robustness across all evaluation indicators. These trend analyses further confirm the theoretical advantages and practical application value of the OQIF method in handling nonlinear longitudinal data.
We further conducted 1000 simulations and plotted boxplots of the 1000 RMSE values for parameters β ^ 1 and β ^ 2 , as shown in Figure 1 and Figure 2. From the figures, we can observe the following: the RMSE values of β ^ 1 and β ^ 2 for both methods decrease as the sample size increases, however, the OQIF method already exhibits good performance with a small sample size (n=100), while the PNLS method requires a larger sample size to achieve similar accuracy. The boxplots of the OQIF method are more symmetric and compact, indicating that the distribution of its estimators is closer to the normal distribution and has better statistical properties. Additionally, the OQIF method has significantly fewer outliers than the PNLS method, demonstrating stronger robustness against abnormal data. This suggests that the overall performance of the OQIF method in parameter estimation is superior to that of the PNLS method, with more pronounced advantages especially in cases of finite samples.
Next, we estimate the coefficient function α ^ U , simulate 1000 times and draw the box plot of the RMSE of α ^ U under different samples, resulting in Figure 3 below.
As can be seen from the above boxplots: with the increase in sample size, the error distributions of both the OQIF method and the PNLS method become more concentrated, but the error of the OQIF method is significantly smaller than that of the PNLS method. This further verifies the superiority of the OQIF method in the varying coefficient partially nonlinear model.

Author Contributions

Conceptualization, J.G., X.Z. and C.W.; methodology, J.G., X.Z. and C.W.; software, J.G, and C.W.; validation, J.G., X.Z. and C.W.; data curation, J.G. and X.Z.; writing—original draft preparation, J.G.; writing—review and editing, X.Z.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Shandong Province (Grant No.ZR2022MA065).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
QIF Quadratic inference function
OQIF The combination of QR decomposition and QIF
PNLS Profile nonlinear least squares

Appendix A

To prove Theorems 1 and 2, we first present the following lemmas:
Lemma A1. 
Let c be a positive constant, and let ξ i , i = 1 , 2 , n . be a sequence of independent and identically distributed random variables. If it satisfies conditions E ( ξ i ) = 0 and E ( ξ i 2 ) < c < , then we have max 1 k n i = 1 k ξ i = O p ( n log n ) .
Proof. 
The result has been proven in Lemma A.2 of Zhao and Xue [17]. □
Lemma A2. 
Assume that the conditions (C1) to (C7) hold, when n , we have:
n G n ( β 0 ) L N ( 0 , Ω 0 ) ,
where Ω 0 is defined in Condition (C7).
Proof. 
Let R ( U i ) = ( X i 1 T r ( U i 1 ) , , X i n i T r ( U i n i ) ) T , r ( u ) = ( r 1 ( u ) , , r p ( u ) ) T , r k ( u ) = α k ( u ) B ( u ) T γ k , combining with Q i 2 T W i = 0 , we can obtain:
Q i 2 T Y i = Q i 2 T g ( Z i , β 0 ) + Q i 2 T W i γ 0 + Q i 2 T R ( U i ) + Q i 2 T ε i = Q i 2 T g ( Z i , β 0 ) + Q i 2 T R ( U i ) + Q i 2 T ε i .
Let G n , k ( β ) denote the k-th component of G n ( β ) , combining with Equation (11), we obtain:
n G n , k ( β 0 ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ( Y i g ( Z i , β 0 ) ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T R ( U i ) + 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ε i .
Conclusion R ( U i ) = O ( K r ) can be derived from Conditions (C2), (C5), and Corollary 6.21 of Schumaker [10], further combining with Lemma A1, we obtain:
1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T R ( U i ) = O p ( n 1 2 n 1 2 log n K r ) = o p ( 1 ) .
Let ζ i k = g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T ( Y i g ( Z i , β 0 ) ) , combining with E ( ε i X i , Z i ) = 0 and cov ( ε i X i , Z i ) = V i , we can thus conclude E ( ζ i k X i , Z i ) = 0 and cov ( ζ i k X i , Z i ) = g ( 1 ) ( Z i , β 0 ) T P i k V i P i k T g ( 1 ) ( Z i , β 0 ) , where P i k is defined in Condition (C7). We further let ζ i = ( ζ i 1 T , ζ i 2 T , , ζ i s T ) T , it follows that E ( ζ i X i , Z i ) = 0 and
cov ( ζ i X i , Z i ) = N i 11 N i 1 s N i s 1 N i s s ,
where N i l k = g ( 1 ) ( Z i , β 0 ) T P i l V i P i k T g ( 1 ) ( Z i , β 0 ) , l , k = 1 , 2 , , s . combining with Condition (C7) and the Law of Large Numbers, we conclude that:
1 n i = 1 n N i l k P Ω l k .
Furthermore, for any constant vector a R s p that satisfying condition a T a = 1 , we have E a T ζ i = 0 and sup i E a T ζ i 2 + δ a 2 + δ sup i E ζ i 2 + δ C sup i E ε i 2 + δ < , where C is a positive constant. Therefore, a T ζ i satisfies the Lyapunov condition of the Central Limit Theorem. Thus, we obtain: i = 1 n a T ζ i ( a T i = 1 n cov ( ζ i ) a ) 1 2 L N ( 0 , 1 ) . Further combining with Equations (A4) and (A5), we can obtain:
1 n i = 1 n cov ( ζ i ) P Ω 0 .
Further combining with Equations (A2), (A3), and the Central Limit Theorem, we can obtain:
n G n ( β 0 ) = 1 n i = 1 n ζ i + o p ( 1 ) L N ( 0 , Ω 0 ) .
The proof of Lemma A2 is completed. □
Lemma A3. 
Assume that the conditions (C1) to (C7) hold, we have G ˙ n ( β 0 ) P Γ 0 and Ω n ( β 0 ) P Ω 0 , where Γ 0 and Ω 0 are defined in Condition (C7), and G ˙ n ( β 0 ) is the first-order derivative of G n ( β 0 ) with respect to β 0 .
Proof. 
After simple calculations, we obtain:
G ˙ n ( β 0 ) = 1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M 1 A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M s A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) .
Combining with Condition (C7) and the Law of Large Numbers, we have that:
1 n i = 1 n g ( 1 ) ( Z i , β 0 ) T Q i 2 A i 1 2 M k A i 1 2 Q i 2 T g ( 1 ) ( Z i , β 0 ) P D k , k = 1 , 2 , s .
Further combining with Equation (A8), we obtain: G ˙ n ( β 0 ) P Γ 0 . The proof of Lemma 2 shows that:
Ω n ( β 0 ) = 1 n i = 1 n ζ i ζ i T + o p ( 1 ) = 1 n i = 1 n ζ i 1 ζ i 1 T ζ i 1 ζ i s T ζ i s ζ i 1 T ζ i s ζ i s T + o p ( 1 ) .
Similar to the proof of Equation (A5), we can obtain:
1 n i = 1 n ζ i l ζ i k T P Ω l k .
Therefore, Ω n ( β 0 ) P Ω 0 , The proof of Lemma 3 is completed. □
Lemma A4. 
Assume that the conditions (C1) to (C7) hold, we obtain:
M ˙ n ( β 0 ) 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = O p ( n 1 ) ,
and
M ¨ n ( β 0 ) 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G ˙ n ( β 0 ) = o p ( 1 ) .
Where M ˙ n ( β 0 ) and M ¨ n ( β 0 ) are the first-order derivative and the second-order derivative of M n ( β 0 ) , respectively.
Proof. 
According to the definition of M n ( β ) , we can obtain the following result through calculation:
M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) .
It follows from Lemma A2 that: G n ( β 0 ) = O p ( n 1 2 ) . Further, from Condition (C7) and Lemma A3, we can obtain:
G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = O p ( n 1 ) .
Combining the Equations (A12) and (A13), we can obtain:
M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + O p ( n 1 ) .
Similarly, we can obtain the following through calculation:
M ¨ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) G ˙ n ( β 0 ) + R n ,
where,
R n = 2 G ¨ n T ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) 4 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) + 2 G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) Ω ˙ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) G n T ( β 0 ) Ω n 1 ( β 0 ) Ω ¨ n ( β 0 ) Ω n 1 ( β 0 ) G n ( β 0 ) = I 1 + I 2 + I 3 + I 4 .
Combining Lemma A2 and Lemma A3, we can obtain the following: I 1 and I 2 are both O p ( n 1 2 ) , while I 3 and I 4 are both O p ( n 1 ) , thus, R n = o p ( 1 ) . The proof of Lemma A4 is completed. □
Proof of Theorem 1. 
Noting that β ^ = arg min β M n ( β ) , thus M ˙ n ( β ^ ) = 0 , according to the Lagrange Mean Value Theorem, we have:
M ˙ n ( β 0 ) + M ¨ n ( β * ) ( β ^ β 0 ) = 0 ,
where β * lies between β ^ and β 0 , thus, we have:
n ( β ^ β 0 ) = n M ¨ n 1 ( β * ) M ˙ n ( β 0 ) .
Combining Lemma A3 and Lemma A4, we can obtain:
M ¨ n 1 ( β * ) = 2 G ˙ n T ( β * ) Ω n 1 ( β * ) G ˙ n ( β * ) + o p ( 1 ) 1 = 1 2 Γ 0 T Ω 0 1 Γ 0 1 + o p ( 1 ) ,
and
n M ˙ n ( β 0 ) = 2 G ˙ n T ( β 0 ) Ω n 1 ( β 0 ) n G n ( β 0 ) + o p ( 1 ) = 2 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) .
Combining Equations (A18) to (A20) and Slutsky’s Theorem, we can obtain:
n ( β ^ β 0 ) = 1 2 Γ 0 T Ω 0 1 Γ 0 1 + o p ( 1 ) 2 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) = Γ 0 T Ω 0 1 Γ 0 1 Γ 0 Ω 0 1 n G n ( β 0 ) + o p ( 1 ) L N ( 0 , Σ ) ,
where, Σ = Γ 0 T Ω 0 1 Γ 0 1 , the proof of Theorem 1 is completed. □
Proof of Theorem 2. 
Let τ = n r 2 r + 1 , γ = γ 0 + τ s . That is, we need to prove that for any given ε > 0 , there exists a constant C such that the following holds:
P { sup s = C Q n ( γ ) > Q n ( γ 0 ) } < 1 ε .
Let Δ ( γ ) = n K 1 [ Q n ( γ ) Q n ( γ 0 ) ] . By Taylor’s Formula, we can obtain:
Δ ( γ ) = n K 1 [ Q n ( γ 0 + τ s ) Q n ( γ 0 ) ] = n K 1 τ s T Q n ( γ 0 ) + 1 2 n K 1 τ 2 s T Q n ( γ 0 ) s + 1 6 n K 1 τ 3 ( γ { s T Q n ( γ * ) s } ) T s = J 1 + J 2 + J 3 ,
where γ * lies between γ and γ 0 . Noting that K = O n 1 2 r + 1 , we can obtain the following through the assumption conditions and some calculations: J 1 = O p ( n τ K 1 ) s = o p ( 1 ) s , J 2 = O p ( n τ 2 K 1 ) s 2 = O p ( 1 ) s 2 , J 3 = O p ( n τ 3 K 1 ) s 3 = o p ( 1 ) s 3 . Thus, there exists a sufficiently large constant C such that when s = C , J 2 can control J 1 and J 3 ; therefore, Equation (A8) holds, and there further exists a maximum point γ ^ satisfying:
γ ^ γ 0 = O p ( τ ) = O p ( n r 2 r + 1 ) ,
then we can obtain:
α ^ k ( · ) α ^ k 0 ( · ) = O p ( n r 2 r + 1 ) , k = 1 , 2 , , p .
The proof of Theorem 2 is completed. □

References

  1. Li, T.; Mei, C. Estimation and inference for varying coefficient partially nonlinear models. Journal of Statistical Planning and Inference 2013, 143, 2023–2037. [Google Scholar] [CrossRef]
  2. Yu, P.; Zhu, Z.; Shi, J.; Ai, X. Robust estimation for partial functional linear regression model based on modal regression. Journal of Systems Science and Complexity 2020, 33, 527–544. [Google Scholar] [CrossRef]
  3. Yan, L.; Tan, X.y.; Chen, X. Empirical likelihood for partially linear errors-in-variables models with longitudinal data. Acta Mathematicae Applicatae Sinica, English Series 2022, 38, 664–683. [Google Scholar] [CrossRef]
  4. Xiao, Y.; Liang, L. Robust estimation and variable selection for varying-coefficient partially nonlinear models based on modal regression. Journal of the Korean Statistical Society 2022, 51, 692–715. [Google Scholar] [CrossRef]
  5. Zhao, P.; Zhou, X.; Wang, X.; Huang, X. A new orthogonality empirical likelihood for varying coefficient partially linear instrumental variable models with longitudinal data. Communications in Statistics-Simulation and Computation 2020, 49, 3328–3344. [Google Scholar] [CrossRef]
  6. Liang, K.Y.; Zeger, S.L. Longitudinal data analysis using generalized linear models. Biometrika 1986, 73, 13–22. [Google Scholar] [CrossRef]
  7. Xu, H.X.; Fan, G.L.; Liang, H.Y. Quantile regression for varying-coefficient partially nonlinear models with randomly truncated data. Statistical Papers 2024, 65, 2567–2604. [Google Scholar] [CrossRef]
  8. Qu, A.; Lindsay, B.G.; Li, B. Improving generalised estimating equations using quadratic inference functions. Biometrika 2000, 87, 823–836. [Google Scholar] [CrossRef]
  9. Bai, Y.; Fung, W.K.; Zhu, Z.Y. Penalized quadratic inference functions for single-index models with longitudinal data. Journal of Multivariate Analysis 2009, 100, 152–161. [Google Scholar] [CrossRef]
  10. Schumaker, L. Spline Functions: Basic Theory. Wiley 1981. [Google Scholar]
  11. Wei, Y.; Wang, Q.; Liu, W. Model averaging for linear models with responses missing at random. Annals of the Institute of Statistical Mathematics 2021, 73, 535–553. [Google Scholar] [CrossRef]
  12. Jiang, Y.; Ji, Q.; Xie, B. Robust estimation for the varying coefficient partially nonlinear models. Journal of Computational and Applied Mathematics 2017, 326, 31–43. [Google Scholar] [CrossRef]
  13. Xiao, Y.T.; Chen, Z.S. Bias-corrected estimations in varying-coefficient partially nonlinear models with measurement error in the nonparametric part. Journal of Applied Statistics 2018, 45, 586–603. [Google Scholar] [CrossRef]
  14. Xu, H.X.; Fan, G.L.; Liang, H.Y. Quantile regression for varying-coefficient partially nonlinear models with randomly truncated data. Statistical Papers 2024, 65, 2567–2604. [Google Scholar] [CrossRef]
  15. Qian, Y.; Huang, Z. Statistical inference for a varying-coefficient partially nonlinear model with measurement errors. Statistical Methodology 2016, 32, 122–130. [Google Scholar] [CrossRef]
  16. Wang, X.; Zhao, P.; Du, H. Statistical inferences for varying coefficient partially non linear model with missing covariates. Communications in Statistics-Theory and Methods 2021, 50, 2599–2618. [Google Scholar] [CrossRef]
  17. Zhao, P.; Xue, L. Empirical likelihood inferences for semiparametric varying-coefficient partially linear errors-in-variables models with longitudinal data. Journal of Nonparametric Statistics 2009, 21, 907–923. [Google Scholar] [CrossRef]
Figure 1. Box plot of 1000 RMSE values for β ^ 1 under the OQIF (A) and PNLS (B) methods
Figure 1. Box plot of 1000 RMSE values for β ^ 1 under the OQIF (A) and PNLS (B) methods
Preprints 182785 g001
Figure 2. Box plot of 1000 RMSE values for β ^ 2 under the OQIF (A) and PNLS (B) methods
Figure 2. Box plot of 1000 RMSE values for β ^ 2 under the OQIF (A) and PNLS (B) methods
Preprints 182785 g002
Figure 3. Box plot of 1000 RMSE for α ^ U with different sample sizes
Figure 3. Box plot of 1000 RMSE for α ^ U with different sample sizes
Preprints 182785 g003
Table 1. The bias and standard deviation of β ^ 1 and β ^ 2 measured by different methods
Table 1. The bias and standard deviation of β ^ 1 and β ^ 2 measured by different methods
Method Parameter n = 100 n = 150 n = 200
Bias SD Bias SD Bias SD
OQIF β ^ 1 0.00257 0.02012 0.00123 0.01890 0.00079 0.01568
β ^ 2 -0.00390 0.03890 -0.00235 0.03568 -0.00157 0.02946
PNLS β ^ 1 0.01457 0.03012 0.01235 0.02789 0.00988 0.02346
β ^ 2 -0.02789 0.05890 -0.02457 0.05235 -0.02012 0.04457
Table 2. Confidence Interval Length and Coverage Probability Comparison
Table 2. Confidence Interval Length and Coverage Probability Comparison
Method Parameter n=100 n=150 n=200
Length Coverage Length Coverage Length Coverage
OQIF β ^ 1 0.07671 0.943 0.07190 0.947 0.05971 0.951
β ^ 2 0.14853 0.937 0.13551 0.941 0.11329 0.948
PNLS β ^ 1 0.11329 0.911 0.10065 0.918 0.08541 0.925
β ^ 2 0.20995 0.899 0.18773 0.905 0.15946 0.912
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated