Preprint
Article

This version is not peer-reviewed.

Limit Theorem for Kernel Estimate of the Conditional Hazard Function with Weakly Dependent Functional Data

A peer-reviewed article of this preprint also exists.

Submitted:

25 August 2025

Posted:

26 August 2025

You are already at the latest version

Abstract
This paper investigates the asymptotic behavior of the conditional hazard function by kernel method , with particular focus on functional weakly dependent data. Specifically, we establish the asymptotic normality for the proposed estimator when the covariate be a functional quasi-associated process. This result contributes to the broader framework of nonparametric inference under weak dependence and functional data analysis. The estimator is constructed using kernel smoothing techniques inspired by the classical Nadaraya-Watson approach, and its theoretical properties are rigorously derived under suitable regularity conditions. To assess its practical performance, we carry out an extensive simulation study, comparing finite-sample results to their asymptotic counterparts. The findings demonstrate the robustness and reliability of the estimator across various settings, confirming the validity of the stated limit theorem in empirical contexts.
Keywords: 
;  ;  ;  ;  

1. Introduction

Recent advances in computational technology and data acquisition systems have made it possible to store and process massive datasets that vary over time, including curves and images. These types of observations are commonly referred to as functional data. Effectively analyzing and modeling such data presents both a challenge and an opportunity for statisticians, leading to the development of powerful statistical tools—chief among them, nonparametric estimation methods.
Pioneering contributions by Bosq and Lecoutre [1], Ferraty and Vieu [2], Ferraty, Mas, and Vieu [3], and Laksaci and Mechab [4] have laid the foundation for nonparametric estimation in the functional data context. Their works have significantly influenced the theoretical aspect and practical implementation of kernel method, making them key references in the domain of nonparametric functional statistics.
Numerous researchers have addressed the study of nonparametric models from both theoretical and practical perspectives. For instance, in the context of kernel estimation, Ferraty and Vieu [5], and Ferraty, Goia, and Vieu [6] investigated regression operators for functional data. Laksaci and Mechab [7] explored the asymptotic behavior of regression functions under weak dependence, Azzi et al. [8] present the function modal regression for functional data, while Hyndman and Yao [9] proposed estimation techniques and symmetry tests for conditional density functions. Daoudi et al. [10] show conditional density estimation under weak dependence and Censorship phenomena. Other notable contributions include those of Attaoui, Laksaci, and Ould Said [11], as well as Xu [12] and Abdelhak et al. [13], who studied single-index models, Daoudi and Mechab [14], who focused on estimating the conditional distribution function under quasi-association assumptions.
These contributions, centered around kernel-based methods for conditional models, have led to significant insights into the asymptotic properties of estimators related to prediction, conditional distribution functions, and their derivatives, particularly the conditional density function. Furthermore, Bulinski and Suquet [15] examined random fields with both positive and negative dependence structures, Bouaker et al. [16] examine the consistency of the kernel estimator for conditional density in high dimensional statistics, and Newman [17] investigated asymptotic independence and limit theorems in similar settings.
Regarding the hazard function, several works have addressed its estimation in dependent contexts. Ferraty, Rabhi, and Vieu [18], Laksaci and Mechab [19], and Gagui and Chouaf [20] obtained the asymptotic normality results in the case of α -mixing conditions. Belguerna et al. [21] further analyzed the MSE of the conditional hazard function estimator.
The concept of quasi-association for real-valued stochastic processes, presented in [22] as a particular form of weak dependence, was later extended by Bulinski and Suquet [23]. Further contributions were made by Kallabis and Neumann [24], who derived exponential inequalities under weak dependence conditions.
More recently, a number of studies have investigated nonparametric models involving quasi-associated random variables. Bassoudi et al. [25] evaluate the asymptotic characteristics of the conditional hazard estimator derived from the local linear technique for ergodic data. Attaoui [26], Tabti and Ait Saidi [27], and Douge [28] have contributed to this direction.
Furthermore, recent research has increasingly focused on the asymptotic analysis of conditional functional models under weak dependence structures, particularly quasi-association. Daoudi, Mechab, and Chikr Elmezouar [29], as well as Daoudi et al. [30], have investigated asymptotic properties of estimators of conditional hazard functions in single-index models for quasi-associated data. Similarly, Bouzebda, Laksaci, and Mohammedi [31] studied the single-index regression model, while Rassoul et al. [32] examined the mean squared error of the conditional hazard rate, highlighting its asymptotic behavior under weak dependence assumptions. Daoudi et al. [35] demonstrate asymptotic results of a conditional risk function estimate for associated data case in high-dimensional statistics.
In the same context, additional contributions have strengthened the theoretical foundation of kernel-based nonparametric estimators. For instance, the asymptotic normality and consistency of the conditional density and hazard function estimators have been addressed under quasi-associated and weakly dependent functional data settings [33]. High-dimensional statistics and complex dependence frameworks have also been tackled, including the asymptotic behavior of regression estimator under quasi-associated functional censored time series [35] and [36]. These works confirm the growing interest in developing robust asymptotic results for conditional models in dependent functional data, offering theoretical tools that support their practical implementation.
This study examines the asymptotic characteristics of the conditional hazard function estimator for functional data under quasi-associated phenomena, with the aim of establishing its asymptotic properties. We begin by introducing the functional model along with all necessary notation and mathematical tools. As a first result, we establish the almost complete consistency. Then, we derive its asymptotic normality by employing various analytical techniques and decomposition strategies. All theoretical developments are supported with rigorous proofs.
To validate the theoretical findings, we conduct a numerical study demonstrating the asymptotic normal approximation of the proposed estimator. In this context, we generate three datasets of different sizes to examine the impact of key parameters such as sample size and bandwidth on estimator performance. Graphical comparisons between theoretical and empirical results are provided to illustrate the estimator’s effectiveness and assess the quality of the estimation.
This paper is structured as follows. In Section 2, we present the quasi-associated sequence, outline the model construction, and present the estimator. Section 3 states the necessary assumptions and develops the results concerning almost complete convergence and asymptotic normality of the estimator. Section 4 provides a comprehensive numerical study supporting the theoretical findings and offering asymptotic confidence bounds. Finally, the detailed proofs of our main results are given in Appendices Appendix A and Appendix B.

2. The Model

Let Z i = ( S i , T i ) 1 i n , be an H × R -valued measurable and strictly stationary stochastic process defined on a probability space ( Ω , A , P ) , where ( H , d ) denotes a semi metric space, where H is a normed . Hilbert space, provided with an inner product < . , . > . The semi metric noted d defined by d ( s , s ) = s s , s , s H . We consider a fixed point s in H , N s a fixed neighborhood of s and S be a fixed compact subset of R . We assume the existence of a regular version of the conditional probability distribution of the random variable T given S. Furthermore, for all s N s , we suppose that the conditional distribution function of T given S = s denoted by G s ( · ) is three times continuously differentiable. We denote its corresponding conditional density function by g s ( · ) .
In this paper, we investigate the kernel estimation of the conditional hazard function of T given S = s , denoted λ s ( t ) , for all t R such that G s ( t ) < 1 , is given by:
λ s ( t ) = g s ( t ) 1 G s ( t ) .
In our functional context, the kernel estimate of this function is given by:
λ ^ s ( t ) = g ^ s ( t ) 1 G ^ s ( t ) , t R .
where G ^ s ( · ) is the conditional distribution functional estimator, given by:
G ^ · ( t ) = i = 1 n K θ K 1 d · , S i H θ H 1 t T i i = 1 n K θ K 1 d · , S i , t R
and g ^ s ( · ) is the conditional density functional estimator, given by:
g ^ · ( t ) = i = 1 n K θ K 1 d · , S i H θ H 1 t T i i = 1 n K θ K 1 d · , S i , t R
with K denote a kernel function, H be a given differentiable distribution function with derivative H . The quantities θ K = θ K , n and θ H = θ H , n represent sequences of positive bandwidth parameters. Under this framework, the estimator λ ^ s ( t ) can be expressed as:
λ ^ s ( t ) = g ^ N s ( t ) G ^ D s G ^ N s ( t ) .
where
G ^ D · = 1 n E K 1 ( · ) i = 1 n K i ( · ) G ^ N · ( t ) = 1 n E K 1 ( · ) i = 1 n K i ( · ) H i ( t ) g ^ N · ( t ) = 1 n h H E K 1 ( · ) i = 1 n K i ( · ) H i ( t )
with notational convenience:
K i ( · ) = K θ K 1 d · , S i , and H i ( t ) = H θ H 1 t T i .
Our primary objective is to establish both the consistency and the asymptotic normality of the estimator (4) under suitable hypothesis, where the sequence of variables S n n N verify the quasi-association as defined by Bulinski and Suquet [18].
Definition 1.
Given I 1 and I 2 subsets of N where I 1 I 2 = ϕ , for all lipschitzian f 1 and f 2 functions, we consider S n n N as a sequence of quasi associated random vector if :
Cov f 1 S τ , τ I 1 , f 2 S κ , κ I 2 Lip ( f 1 ) Lip ( f 2 ) τ I 1 κ I 2 u = 1 d u = 1 d Cov S τ u , S κ u
Lip ( f 1 ) = sup s t | f 1 ( s ) f 1 ( t ) | s t , with s 1 , , s n = s 1 + + s n .
S τ u represent the u t h component of S τ , defined as S τ u : = < S τ , e u > , where e u , s 1 is an orthonormal basis.

3. Main Results

3.1. Assumptions

In the sequel, when no confusion will likely to arise, we will denote by α and α strictly positive constants, and by χ u the covariance coefficient defined as:
χ u = s u p v u i j v χ i , j
Where
χ i , j = u > 1 i > 1 Cov ( S i u , S j u ) + u > 1 Cov ( S i u , T j ) + i > 1 Cov ( T i , S j u ) + Cov ( T i , T j ) .
For ρ > 0 , let B ( s , ρ ) : = { s H / d ( s , s ) < ρ } be a small ball, shis center and ρ his radius.
To achieve the desired goal, we begin by stating the following required assumptions.
( H 1 )
P ( S B ( s , θ K ) ) = ϕ s ( θ K ) > 0 , and ϕ s ( · ) is a differentiable at 0. Moreover, β s ( · ) such that
u [ 1 , 1 ] , lim θ K 0 ϕ s u θ K ϕ s θ K = β s ( u ) .
( H 2 )
Assume that the Hölder continuity condition is hold for both functions G s ( t ) and g s ( t ) .
G s 1 ( t 1 ) G s 2 ( t 2 ) α d γ 1 ( s 1 , s 2 ) + | t 1 t 2 | γ 2 g s 1 ( t 1 ) g s 2 ( t 2 ) α d γ 1 ( s 1 , s 2 ) + | t 1 t 2 | γ 2
for all ( s 1 , s 2 ) N s 2 and ( t 1 , t 2 ) S 2 , with constants γ 1 > 0 , γ 2 > 0 , and S be a subset of R ( S compact).
( H 3 )
H is an even and bounded function, with a bounded and Lipschitz continuous derivative H , satisfying:
H ( z ) d z = 1 , | z | γ 2 H ( z ) d z < a n d ( H ( z ) ) 2 d z < .
( H 4 )
For a differentiable, Lipschitz continuous and bounded kernel K, α and α such:
α I [ 0 , 1 ] ( . ) < K ( . ) < α I [ 0 , 1 ] ( . )
I [ 0 , 1 ] ( . ) : is the indicator function on [ 0 , 1 ] , K ( . ) is derivative of K ( . ) with:
< α < K ( z ) < α < 0 for 0 z 1
.
( H 5 )
The random pairs ( S κ , T κ ) , κ N is a quasi-associated sequence with covariance coefficient χ u , u N satisfying :
a > 0 , α > 0 , such that χ u α e a u .
( H 6 )
0 < sup i j P ( S i , S j ) B ( s , θ K ) × B ( s , θ K ) = max i j P d s , S i < θ K , P d s , S j < θ K = O ϕ s 2 ( θ K ) .
( H 7 )
The bandwidths θ K and θ H satisfy :
i-
lim n 1 n θ H ϕ s θ K = 0 , lim n log 5 n n θ H ϕ s θ K = 0 .
ii-
lim n n θ H 5 ϕ s θ K = 0 and lim n n θ H 2 θ K ϕ s θ K = 0 .
iii-
lim n θ H 2 + θ K n ϕ s θ K = 0
( H 8 )
For m { 0 , 2 } , the functions Φ l ( · ) = E m g S ( t ) t m m g s ( t ) t m d ( s , S ) = · and
Ψ ( · ) = E m G S ( t ) t m m G s ( t ) t m d ( s , S ) = · are derivable at 0.

3.2. Comments on the Assumptions

( H 1 ) This assumption specifies conditions governing the probability that S lies within a neighborhood of s, along with the consistence behavior of the corresponding probability ratio as the neighborhood size approaches zero. These conditions are essential for applying local convergence theorems. ( H 2 ) The Hölder continuity assumption imposed on the conditional distribution and their derivatives represents a standard regularity condition in the literature. ( H 3 ), ( H 4 ), and ( H 7 ) are technical, ensuring the convergence of the convolution kernel, and enabling the use of Taylor expansions. ( H 5 ) This quasi-association assumption on the data is interesting, as it covers a more general framework than classical independence. ( H 6 ) This assumption characterizes the asymptotic behavior of the joint distribution. ( H 8 ) This last assumption concerns the control of the joint probability of two S variables in a neighborhood of s, which helps control the covariance terms in the asymptotic developments. Such assumptions are typical in the setting of nonparametric estimation when dealing with functional covariate.

3.3. Main Results

3.3.1. The Almost Consistency

Our goal is to derive the almost complete convergence (a.co.) of λ ^ s ( t ) to λ s ( t ) , and this result is formalized in the following theorem.
Theorem 1.
Under the conditions ( H 1 )-( H 8 ), we have
λ s ^ ( t ) λ s ( t ) = O θ H 2 + θ K + O a . co . log n n θ H ϕ s θ K as n
Proof 
(Proof of theorem 1). needed this decomposition:
λ s ^ ( t ) λ s ( t ) = g ^ N s ( t ) G ^ D s G ^ N s ( t ) g s ( t ) 1 G s ( t ) = 1 G ^ D s G ^ N s ( t ) g ^ N s ( t ) E g ^ N s ( t ) + 1 G ^ D s G ^ N s ( t ) λ s ( t ) E G ^ N s ( t ) G u ( v ) + E g ^ N s ( t ) g s ( t ) + λ s ( t ) G ^ D s G ^ N s ( t ) 1 E G ^ N s ( t ) G ^ D s G ^ N s ( t )
And the following subsequent results.
Lemma 1.Under the assumptions ( H 1 ), ( H 4 )-( H 7 ), and for any fixed t, we have
g ^ N s ( t ) E g ^ N s ( t ) = O a . co . log n n θ H ϕ s θ K as n .
Lemma 2.Assuming that the conditions ( H 1 )-( H 8 ) hold, then for n , we infer:
E G ^ N s ( t ) G s ( t ) = N B G ( s , t ) + o θ H 2 + o θ K
E g ^ N s ( t ) g s ( t ) = N B g ( s , t ) + o θ H 2 + o θ K
where
N B G ( s , t ) = θ H 2 2 v 2 H ( v ) d v 2 G s ( t ) t 2 + θ K Ψ 2 ( 0 ) ω 0 ω 1 N B g ( s , t ) = θ H 2 2 v 2 H ( v ) d v 2 g s ( t ) t 2 + θ K Φ 2 ( 0 ) ω 0 ω 1 ω 0 = K ( 1 ) 0 1 ( u K ( u ) ) β s ( u ) d u ω τ = K τ ( 1 ) 0 1 K τ ( u ) β s ( u ) d u for τ = 1 , 2
Corollary 1. Making use the assumptions ( H 1 )-( H 8 ), we get:
Var G ^ D s = 1 n K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u ϕ s ( θ K ) K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 1 + o 1 n ϕ s ( θ K ) Var G ^ N s ( t ) = G s ( t ) n ϕ s θ K H 2 ( v ) d v K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 + o 1 n ϕ s θ K Cov G ^ D s , G ^ N s ( t ) = G s ( t ) K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u n ϕ s θ K K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 G s ( t ) n + o 1 n ϕ s ( θ K )
Lemma 3. Under the assumptions of Theorem 1
G ^ D s G ^ N s ( t ) 1 G s ( t ) , in probability .
And
n θ H ϕ s θ K σ λ ^ 2 1 E G ^ N s ( t ) G ^ D s G ^ N s ( t ) = O p ( 1 )
with
σ λ ^ 2 = ω 2 λ s ( t ) ω 1 2 1 G s ( t ) ( H ( u ) ) 2 d u .

3.3.2. Asymptotic Normality

Theorem 2.
Under ( H 1 )-( H 7 ), we infer:
n θ H ϕ s θ K λ ^ s ( t ) λ s ( t ) D N 0 , σ λ ^ 2 , s A ; n
where
A = s H , g s ( t ) 1 G s ( t ) 0 .
and σ λ ^ 2 is defined in Lemma 3.
Proof of Theorem 2.
Needed the decomposition (6), the results shown in above Lemmas (2 and 3) and the following subsequent result (Lemma 4):
Lemma 4. Assuming that ( H 1 )-( H 7 ) hold, then
n θ H ϕ s θ K g ^ N s ( t ) E g ^ N s ( t ) D N 0 , σ g 2
with
σ g 2 = ω 2 g s ( t ) ω 1 2 ( H ( u ) ) 2 d u
The rigorous proof and demonstrations of the intermediate results (Lemmas 1,2,3 and 4), and Corollary □

4. Application and Numerical Study

4.1. Confidence Bounds

Constructing reliable confidence bounds is a key aspect of statistical analysis, as they characterize the variability and reliability of model estimators. Proper interpretation of these bounds enables more informed and robust conclusions about the underlying estimator. In the context of survival and hazard function estimation, confidence bounds are particularly valuable since they provide a quantitative assessment of the uncertainty surrounding the estimated conditional hazard function, thereby guiding both theoretical analysis and practical decision-making. Moreover, confidence bounds serve as a diagnostic tool: narrow bounds suggest stable and precise estimators, while wider bounds highlight regions where the estimator is less reliable due to limited data or high variability.
As an application of the result established in Theorem 2, we construct confidence bounds for λ s ( t ) at the confidence level 1 α . To this end, we must first estimate the unknown components of the asymptotic variance as follows: these include the conditional density, the conditional survival function, and kernel-based quantities that appear in the variance expression. Consistent estimation of these components is essential, since any bias or misspecification would directly affect the coverage probability of the resulting bounds. Once these quantities are estimated, the asymptotic normality result in Theorem 2 allows us to approximate the distribution of the estimator and derive pointwise confidence intervals for λ s ( t ) across the range of t. This methodology not only validates the theoretical properties of the proposed estimator but also ensures its applicability in empirical studies where inference on the conditional hazard function is required.
ω ^ q = : 1 n ϕ ^ s ( θ K ) τ = 1 n K τ q , q = 1 , 2 ;
ϕ ^ s ( θ K ) = # τ : d S τ , s θ K n
where # ( A ) represents the cardinality of the set A. Also, σ λ ^ 2 is estimated by:
σ λ ^ 2 ^ = : ω 2 ^ λ ^ s ( t ) ω ^ 1 2 ( H ( u ) ) 2 d u .
Corollary 2.
When the assumptions of Theorem 2 hold, we set:
n θ H ϕ ^ s θ K λ ^ s ( t ) λ s ( t ) d N 0 , σ λ ^ 2 ^ , n .
Moreover, the confidence bounds will be,
λ ^ s ( t ) Z 1 α 2 σ λ ^ 2 ^ n θ H ϕ ^ s θ K 1 / 2 , λ ^ s ( t ) + Z 1 α 2 σ λ ^ 2 ^ n θ H ϕ ^ s θ K 1 / 2
where Z 1 α 2 is the quartile 1 α 2 of N ( 0 , 1 ) .

4.2. Numerical Study

In this part, we conduct a numerical study using R software to illustrate and validate the theoretical results through graphical representations. The aim of this simulation is to assess the finite-sample performance of the proposed estimator and to highlight the extent to which the asymptotic properties established in the theoretical framework are reflected in practice. By generating controlled data under specific dependence structures and censoring mechanisms, we are able to visualize the behavior of the conditional density and hazard estimators, compare them with their theoretical counterparts, and evaluate their accuracy across different sample sizes. This numerical experiment also provides insights into the rate of convergence, the influence of the smoothing parameters, and the robustness of the estimator under varying conditions.
This simulation is based on the following points: we describe the data-generating process and the dependence structure imposed, specify the choice of kernel functions and bandwidths, outline the implementation of the random censoring mechanism, and finally present the graphical outputs and error metrics that allow for a systematic comparison between theoretical and empirical curves. The numerical results obtained will not only complement the theoretical findings but also serve as practical evidence of the efficiency and reliability of the proposed estimation methodology.
  • Definite our model by choosing functional covariate as :
    S κ ( u ) = cos W κ u + sin W κ + u + 0.7 W κ u , u [ 0 , + π ] for κ = 1 , , n
    The process W κ satisfies a specific dependence structure,namely a quasi-associated sequence, which is generated as a non-strong mixing auto-regressive process of order 1. This process is constructed by setting the auto-regressive coefficient ρ = 0.1 and modeling the innovation term as a binomial distribution Binom ( 10 , 0.25 ) [18]. We use 100 discretization points of u to obtain the curves S κ shown below in Figure 1, Figure 2 and Figure 3, corresponding to different sample sizes.
    For the real variable is defined as T = m ( S ) + ϵ . Where m is the nonlinear regression operator,
    m ( S ) = 1 5 × exp 2 1 0 π S ( u ) d u 2
    Where ϵ is standard normal distribution. Its clear that the explicit form of the conditional density given by:
    g s t = 1 2 π e 1 2 t m ( s ) 2
    . In the next, we select the distance in H as:
    d s 1 , s 2 = 0 π s 1 ( u ) s 2 ( u ) 2 d u 1 / 2 s 1 , s 2 H
    Also
    K ( s ) = 15 16 1 s 2 2 , s [ 1 , 1 ] ; H ( s ) = s K ( u ) d u
    .
  • A bandwidth Selection Algorithm: The smoothness of the estimators (2) and (3) is controlled by the smoothing parameter θ K and the regularity of the cumulative distribution function (CDF). Therefore, choosing these parameters plays a critical role in the computational process. An optimal selection leads to effective estimation with a small mean squared error (MSE), which, for the conditional hazard function, is given by:
    MSE ( λ ^ ) = 1 n i = 1 n λ ^ s ( t i ) λ s ( t i ) 2 .
    Let H ( . ) be a distribution function on R and H θ ( s ) = H ( s / θ ) . Note that as θ 0 .
    E H θ t T i S i = s = G s ( t ) + O θ 2 .
    This indicates that G s ( t ) can be interpreted as a regression of H λ t T i on S i , Consequently, we adapt this regression framework to our estimation problem. By combining this approach with the normal reference rule [7], we derive an effective algorithm for selecting the bandwidth parameters.
    i-
    Compute the bandwidth θ H using the normal reference rule.
    ii-
    Given θ H , use cross-validation (as proposed by [1]) to determine the optimal value of θ K (using the function fregre.np.cv in the package fda.usc) for our calculation of θ K .
  • Calculate the estimates of both the conditional distribution and the conditional density functions, and compare these estimates with their theoretical counterparts on the same graphs.(Figure 4 and Figure 5).
    It is apparent that our estimations exhibit high accuracy when optimal bandwidths are selected. To assess the performance of each model more rigorously, we compute the mean squared error, as shown in Table 1.
    For the next step in achieving the desired objective and firmly establishing the normal approximation of λ ^ s ( t ) with high effectiveness, we selected the sample that produced an estimate with the smallest MSE (sample size n = 1000 ), and followed the subsequent steps:
    • We compute the conditional hazard function estimator using (1), the asymptotic variance σ λ ^ 2 ^ defined in (7), and the empirical estimate ϕ ^ s θ K .
    • Under the condition:
      lim n n θ H ϕ s θ K N B λ ( s , t ) = 0 ,
      we can ignore the bias term N B λ ( s , t ) and compute the quantity referred to in (8), namely QNH, as:
      n θ H ϕ ^ s ( θ K ) σ λ ^ 2 ^ λ ^ s ( t ) λ s ( t ) .
    • Plot a histogram of QNH and compare it with the standard normal density. (Figure 6).
    • Finally building a confidence bounds (see Figure 7).

5. Conclusion and Some Perspectives

Due to the complexity of the conditional hazard function estimate λ ^ s ( t ) by the kernel approach, we began by decomposing the estimator into three parts, as shown in (6). The first part corresponds to the numerator of the density function estimator, denoted g ^ N s ( t ) , which is considered the dominant component influencing the estimator’s properties. We showed that the denominator of the corresponding term converges in probability to 1 G s ( t ) . The remaining two parts account for the bias in both the conditional distribution and density function estimators, arising from C ^ N s ( t ) and g ^ N s ( t ) .
From a practical perspective, we conducted a simulation study to validate the theoretical findings. Despite the inherent challenges in selecting and tuning parameters such as bandwidths, the proposed estimator demonstrated strong performance, achieving a low mean squared error.
These results are promising and contribute to the growing body of research in this field. However, further investigation is needed to assess the robustness of the method in diverse real-world settings and to explore potential extensions that could improve its applicability and generalizability.
A potential direction for future work is to examine the sensitivity of the estimator to kernel choice and bandwidth selection procedures. Exploring data-driven or adaptive bandwidth strategies may enhance finite-sample performance. Additionally, investigating the behavior of the estimator in high-dimensional or complex data settings could broaden its practical utility.
This study enhances both the theoretical and empirical comprehension of the asymptotic characteristics of the proposed estimator, laying the foundation for future developments and applications in areas where estimating conditional functional parameters is essential.

Author Contributions

Conceptualization, A.R., Z.C.E., H.D. and A.B.; methodology, Z.C.E. and H.D.; software, A.R.; validation, Z.C.E., A.B., and H.D.; formal analysis, A.R., Z.C.E., F.A. and H.D.; investigation, A.R., Z.C.E., F.A., A.B., and H.D.; resources, A.R., Z.C.E., F.A. and H.D.; data curation, A.R.; writing original draft preparation, H.D. and A.R.; writing review and editing, A.R.,Z.C.E., A.B, and H.D.; visualization, A.R. and H.D.; supervision, Z.C.E. and F.A.; project administration, Z.C.E.; funding acquisition, Z.C.E. and F.A. All authors have read and approved the final version of the manuscript for publication.

Funding

This research project was funded by (1) Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2023R358), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia, and (2) The Deanship of Research and Graduate Studies at King Khaled University for funding this work through small under grant numbergroup research RGP1/163/45

Data Availability Statement

The data used to support the findings of this study are available on request from the corresponding author.

Acknowledgments

The authors thank and express their sincere appreciation to the funders of this work: (1) Princess Nourah bint Abdulrahman University Researchers Supporting Project (Project Number: PNURSP2025R358), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia; (2) The Deanship of Scientific Research at King Khalid University, through the Research Groups Program under grant number R.G.P. 1/118/46.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Lemma 1.
The proof of this lemma is based on the exponential inequality of Kallabis and Nymann which is given in the following lemma:
Lemma A1. [19] Let Z 1 , . . . , Z ρ real random variables with E ( Z τ ) = 0 and P ( Z τ M ) = 1 for all τ = 1 , . . . , ρ and M < , let σ ρ 2 = Var ( τ = 1 ρ Z τ ) .
Assume that there exist Υ < and η > 0 such that the following inequality is true for all u-tuplets ( l 1 , . . . , l s ) N s and r-tuplets ( v 1 , . . . , v r ) N r with 1 l 1 . . . l u v 1 . . . v r ρ .
Cov ( Z l 1 . . . Z l u , Z v 1 . . . Z v r ) Υ 2 M u + r 2 r e η ( v 1 l u )
Then,
P τ = 1 ρ Z τ > ϵ exp ϵ 2 / 2 A ρ + B ρ 1 3 r 5 3
for some A ρ σ ρ 2
B ρ = 16 ρ Υ 2 9 A ρ ( 1 e η ) 1 2 ( Υ M ) 1 e η .
We use Lemma 1 on the variables:
Z n ^ τ ( s , t ) = 1 n θ H E K 1 ( s ) Γ τ ( s , t ) E Γ τ ( s , t ) , 1 τ n
Where Γ τ ( s , t ) = K τ ( s ) H τ ( t ) , s H , t R . moreover, we have:
E Z n ^ τ = 0 Z n ^ τ 2 K H n θ H ϕ s ( θ K ) Lip Z n ^ τ 2 α θ K 1 H Lip ( K ) + θ H 1 K Lip ( H ) n θ H ϕ s ( θ K )
And
g ^ N s ( t ) E g ^ N s ( t ) = τ = 1 n Z n ^ τ
In order to apply Lemma 1 we have to choose the variable A n which is conditioned by the variance of Z n ^ , starting by upper bound the Var τ = 1 n Z n ^ τ as follow:
Var τ = 1 n Z n ^ τ = 1 n θ H E K 1 ( s ) 2 τ = 1 n j = 1 n Cov Γ τ ( s , t ) , Γ j ( s , t ) = 1 n θ H E K 1 ( s ) 2 Var Γ 1 ( s , t ) + 1 n θ H E K 1 ( s ) 2 τ = 1 n j = 1 τ j n Cov Γ τ ( s , t ) , Γ j ( s , t ) = 1 n θ H E K 1 ( s ) 2 n Var Γ 1 ( s , t ) + τ τ j Cov Γ τ ( s , t ) , Γ j ( s , t ) = 1 n θ H E K 1 ( s ) 2 n A 1 + A τ j
Where,
A 1 = Var Γ 1 ( s , t ) = E K 1 2 ( s ) H 1 2 ( t ) E K 1 ( s ) H 1 ( t ) 2
Thus, under ( H 2 ) and ( H 3 ), and by integration on the real component z It follows that
E K 1 2 ( s ) H 1 2 ( t ) = E K 1 2 ( s ) E H 1 2 ( t ) S 1 = E K 1 2 ( s ) H 2 t z θ H g S ( z ) d z ; taking v = t z θ H = θ H E K 1 2 ( s ) H 2 ( v ) g S t θ H v d v ,
We show for n enough large using Taylor expansion (first order)
g S t θ H v = g S ( t ) + O θ H = g S ( t ) + o ( 1 )
Hence
E K 1 2 ( s ) H 1 2 ( t ) = θ H H 2 ( v ) d v E K 1 2 ( s ) g S ( t ) + o θ H
We denote by φ m ( S , t ) : = m g S ( t ) t m for m { 0 } , then
E K 1 2 ( s ) φ m S , t = φ m ( s , t ) E K 1 2 ( s ) + E K 1 2 ( s ) φ m ( S , t ) φ m ( s , t ) = φ m ( s , t ) E K 1 2 ( s ) + E K 1 2 ( s ) Φ m ( d ( s , S ) )
In accordance with Ferraty et al.[2] (refer to Lemma 1, page 26), we establish:
E K 1 2 ( s ) Φ m ( d ( s , S ) ) = K 1 2 d ( s , S ) θ K Φ m d ( s , S ) d μ d ( s , S ) ( v ) = K 1 2 v θ K Φ m ( v ) d μ d ( s , S ) ( v ) = K 1 2 ( v ) Φ m ( θ K v ) d μ d ( s , S ) θ K ( v ) = θ K Φ m ( 0 ) v K 1 2 ( v ) d μ d ( s , S ) θ K ( v ) + o ( θ K ) 0 for m = 0
The last line is justified by the 1 order of the Taylor expansion for Φ around 0, and Φ 0 ( 0 ) = 0 . Additionally, we employ the results of Lemma 2 on page 27 in Ferraty et al.[2].
E K 1 2 ( s ) = ϕ s θ K K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u + o ( 1 ) .
Then, (A4) become,
E K 1 2 ( s ) φ 0 S , t = E K 1 2 ( s ) g S ( t ) = φ 0 ( s , t ) E K 1 2 ( s ) . = ϕ s θ K g s ( t ) K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u + o ϕ s ( θ K ) .
This enables us to deduce
E K 1 2 ( s ) H 1 2 ( t ) = θ H H 2 ( v ) d v ϕ s θ K g s ( t ) K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u + o θ H ϕ s θ K .
For the 2nd term in the equation(A3), by the same steps followed above, ( H 3 ) satisfy, we demonstrate
E K 1 ( s ) H 1 ( t ) = θ H ϕ s θ K g s ( t ) K ( 1 ) 0 1 ( K ( u ) ) β s ( u ) d u + o θ H ϕ s θ K .
Which implies that
Var Γ 1 ( s , t ) = θ H ϕ s θ K g s ( t ) H 2 ( v ) d v K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u + o θ H ϕ s θ K .
And
1 n θ H E K 1 ( s ) 2 Var Γ 1 ( s , t ) = O 1 n θ H ϕ s θ K .
For the 2nd term A τ j , we decompose the sum in 2 sets by m n with m n , as n .
A τ v = τ = 1 n j = 1 τ j n Cov Γ τ ( s , t ) , Γ j ( s , t ) = τ = 1 n j = 1 0 < | τ j | m n n Cov Γ τ ( s , t ) , Γ j ( s , t ) + τ = 1 n j = 1 | τ j | > m n n Cov Γ τ ( s , t ) , Γ j ( s , t ) = : I n + I I n .
Under assumptions ( H 1 ), ( H 3 ) and ( H 5 ), we infer for τ j
I n = τ = 1 n j = 1 0 < | τ j | m n n Cov Γ τ ( s , t ) , Γ j ( s , t ) n m n max τ j E Γ τ ( s , t ) Γ j ( s , t ) + E Γ 1 ( s , t ) 2 α n m n max τ j E K τ H τ K j H j + E K 1 H 1 2 α n m n θ H 2 ϕ s 2 θ K + θ H ϕ s θ K 2 α n m n θ H 2 ϕ s 2 θ K .
Now, under the assumptions ( H 3 )-( H 5 ), we set
I I n = τ = 1 n j = 1 | τ j | > m n n Cov Γ τ ( s , t ) , Γ j ( s , t ) Lip ( K ) θ K + Lip ( H ) θ H 2 τ = 1 n j = 1 | τ j | > m n n χ τ , j α n Lip ( K ) θ K + Lip ( H ) θ H 2 χ m n α n Lip ( K ) θ K + Lip ( H ) θ H 2 e a m n .
Then, by (A12) and (A13), we get
A τ j = τ = 1 n j = 1 τ j n Cov Γ τ ( s , t ) , Γ j ( s , t ) α n m n θ H 2 ϕ s 2 θ K + Lip ( K ) θ K + Lip ( H ) θ H 2 e a m n
Taking m n = 1 γ log γ θ K 1 Lip ( K ) + θ H 1 Lip ( H ) 2 θ H 2 ϕ s 2 θ K , we get
1 n θ H E K 1 ( s ) 2 A τ j 0 , as n
Again, by (A10), and (A14), we show
Var τ = 1 n Z n ^ τ = O 1 n θ H ϕ s θ K
Now, we need to evaluate the the covariance term Cov Z n ^ l 1 Z n ^ l u , Z n ^ v 1 Z n ^ v r , for all l 1 , , l u N u and v 1 , , v r N r with 1 l 1 l u v 1 v r n . For that we distingue the following cases:
  • If v 1 = l u . Using the result (A7), we obtain
    Cov Z n ^ l 1 Z n ^ l u , Z n ^ v 1 Z n ^ v r 1 n θ H E K 1 u + r E Z n ^ l 1 Z n ^ v 1 2 Z n ^ v r α K H n θ H ϕ s ( θ K ) u + r E K v 1 2 H v 1 2 α n θ H ϕ s θ K u + r θ H ϕ s θ K
  • If v 1 > l u , quasi-association, under ( H 5 ), we obtain:
    Cov Z n ^ l 1 Z n ^ l u , Z n ^ v 1 Z n ^ v r 4 Lip ( K ) θ K + Lip ( H ) θ H n θ H ϕ s θ K 2 × 2 α K H n θ H ϕ s θ K u + r 2 τ = 1 u j = 1 r χ l τ , v j Lip ( K ) θ K + Lip ( H ) θ H 2 α n θ H ϕ s θ K u + r ( u r ) χ v 1 l u Lip ( K ) θ K + Lip ( H ) θ H 2 α n θ H ϕ s θ K u + r r e a v 1 l u .
    And, by ( H 6 ) we hold,
    Cov Z n ^ l 1 Z n ^ l u , Z n ^ v 1 Z n ^ v r α K H n θ H ϕ s θ K u + r 2 Cov Z n ^ v 1 , Z n ^ l u α K H n θ H ϕ s θ K u + r 2 E Z n ^ l u Z n ^ v 1 + E Z n ^ l u E Z n ^ v 1 α K H n θ H ϕ s θ K u + r 2 α n θ H ϕ s θ K 2 × θ H 2 sup l m P S l , S m B s , θ K × B s , θ K + P S 1 B s , θ K 2 α n θ H ϕ s θ K u + r θ H 2 ϕ s 2 θ K .
    Furthermore, taking a (1- η )-power of (A16), η -power of (A17), with 1 / 4 < η < 1 / 2 , for the tree terms we drive an upper-bound as follows:
    for 1 l 1 l u v 1 v r n :
    Cov Z n ^ l 1 Z n ^ l u , Z n ^ v 1 Z n ^ v r Lip ( K ) θ K + Lip ( H ) θ H 2 η α n θ H ϕ s θ K u + r × θ H ϕ s θ K 2 ( 1 η ) r η e a η v 1 l u Lip ( K ) θ K + Lip ( H ) θ H η α n θ H ϕ s θ K θ H ϕ s θ K ( 1 η ) 2 × α n θ H ϕ s θ K u + r 2 r e a η v 1 l u
The variables Z n ^ τ , τ = 1 , , n fulfill the requirements of Lemma 1 for:
Υ = Lip ( K ) θ K + Lip ( H ) θ H η α n θ H ϕ s θ K θ H ϕ s θ K ( 1 η ) M = α n θ H ϕ s θ K ; A ρ = 1 n θ H ϕ s θ K B ρ = 16 ρ Υ 2 9 A ρ ( 1 e η ) 1 2 ( Υ M ) 1 e η = 1 n θ H ϕ s θ H
Thus,
P g ^ N s ( t ) E g ^ N s ( t ) > η log n n θ H ϕ s θ K = P τ = 1 n Z n ^ τ > η log n n θ H ϕ s θ K exp η 2 log n 2 n θ H ϕ s θ K D ( n )
Where
D ( n ) = 1 n θ H ϕ s θ K + 1 n θ H ϕ s θ H 1 3 η 2 log n n θ H ϕ s θ K 5 6
Hence
P g ^ N s ( t ) E g ^ N s ( t ) > η log n n θ H ϕ s θ K exp η 2 log n 2 + ( η 2 log ( n ) ) 5 / 6 n θ H ϕ s θ K 1 / 6 α exp η 2 log ( n )
Finally, for a suitable choice of η , the proof is achieved. □
Proof 
(Proof of Lemma 2). Taking v = t z θ H , using the stationarity property, we can write:
  • For the bias term of G ^ N s ( t )
    E G ^ N s ( t ) = E 1 n E K 1 ( s ) i = 1 n K i ( s ) H i ( t ) = 1 E K 1 ( s ) E K 1 ( s ) E H 1 ( t ) S
    With
    E H 1 ( t ) S = R H 1 t z θ H g S z d z = 1 θ H R H 1 t z θ H G S z d z = R H 1 ( v ) G S t θ H v d v
    Using a Taylor expansion of the function G S t θ H v :
    G S t θ H v = G S ( t ) θ H v G S ( t ) t + θ H 2 v 2 2 2 G S ( t ) t 2 + o θ H 2
    Under (A21) and hypothesis ( H 3 ), we deduce:
    E H 1 ( t ) S = G S ( t ) + θ H 2 2 2 G S ( t ) 2 t v 2 H ( v ) d v + o θ H 2 .
    Insert (A22) in (A20)
    E G ^ N s ( t ) = 1 E K 1 ( s ) E K 1 ( s ) G S ( t ) + θ H 2 2 v 2 H ( v ) d v E K 1 ( s ) 2 G S ( t ) 2 t + o θ H 2 .
    Denote by ψ m ( S , t ) : = m G S ( t ) t m for m { 0 , 2 } , then
    E G ^ N s ( t ) = E K 1 ( s ) ψ 0 ( S , t ) E K 1 ( s ) + E K 1 ( s ) ψ 2 ( S , t ) E K 1 ( s ) θ H 2 2 v 2 H ( v ) d v + o λ H 2 .
    Where
    E K 1 ( s ) ψ m S , t = ψ m ( s , t ) E K 1 ( s ) + E K 1 ( s ) ψ m ( S , t ) ψ m ( s , t ) = ψ m ( s , t ) E K 1 ( s ) + E K 1 ( s ) Ψ m ( d ( s , S ) )
    With the same steps following to evaluate (A5), we set:
    E K 1 ( s ) Ψ m ( d ( s , S ) ) = K 1 d ( s , S ) θ K Ψ m d ( s , S ) d μ d ( s , S ) ( v ) = K 1 v θ K Ψ m ( v ) d μ d ( s , S ) ( v ) = K 1 ( v ) Ψ m ( θ K v ) d μ d ( s , S ) θ K ( v ) = θ K Ψ m ( 0 ) v K 1 ( v ) d μ d ( s , S ) θ K ( v ) + o ( θ K )
    The last line justifies by the 1 order Taylor expansion for Ψ around 0. Additionally, we employ the results of Ferraty et al.[2] (Lemma 2, page 27).
    E K 1 ( s ) = ϕ s θ K K ( 1 ) 0 1 K ( u ) β s ( u ) d u + o ( 1 ) . v K 1 ( v ) d μ d ( s , S ) θ K ( v ) = K ( 1 ) 0 1 ( u K ( u ) ) β s ( u ) d u
    Which allow us under (A25) to set:
    E K 1 ( s ) Ψ m ( d ( s , S ) ) = θ K ϕ s ( θ K ) Ψ m ( 0 ) K ( 1 ) 0 1 ( u K ( u ) ) β s ( u ) d u + o ( 1 )
    Using (A24), (A26) and the fact that Ψ 0 ( 0 ) = 0 ,
    E K 1 ( s ) ψ 0 ( S , t ) E K 1 ( s ) = ψ 0 ( s , t ) + o ( θ K ) E K 1 ( s ) ψ 2 ( S , t ) E K 1 ( s ) = ψ 2 ( s , t ) + θ K Ψ 2 ( 0 ) K ( 1 ) 0 1 ( u K ( u ) ) β s ( u ) d u + o ( 1 ) K ( 1 ) 0 1 K ( u ) β s ( u ) d u + o ( 1 ) + o ( θ K )
    Hence,
    E G ^ N s ( t ) = G s ( t ) + θ H 2 2 v 2 H ( v ) d v 2 G s ( t ) t 2 + θ K Ψ 2 ( 0 ) K ( 1 ) 0 1 ( u K ( u ) ) β s ( u ) d u K ( 1 ) 0 1 K ( u ) β s ( u ) d u + o θ H 2 + o θ K .
  • For the bias term of g ^ N s ( t ) , we start by writing:
    E g ^ N s ( t ) = 1 E K 1 ( s ) E K 1 ( s ) E θ H 1 H 1 ( t ) S with θ H 1 E H 1 ( t ) S = R H ( v ) g S t θ H v d v .
    Using a Taylor expansion under ( H 3 ), we infer:
    λ H 1 E H 1 ( t ) S = g S ( t ) + θ H 2 2 2 g S ( t ) 2 t v 2 H ( v ) d v + o θ H 2 .
    The same steps used to studying E G ^ N s ( t ) can be followed ( see Rassoul et al. [36] page 16) to infer that:
    E g ^ N s ( t ) g s ( t ) = B N g ( s , t ) + o θ H 2 + o θ K .
Proof of Corollary 1.
Using (A1), we can write:
Var g ^ N s ( t ) = Var g ^ N s ( t ) E g ^ N s ( t ) = Var τ = 1 n Z n ^ τ
Then, to calculate Var G ^ N s ( t ) , we replace H τ by H τ and follow the same steps used in the evaluation of (A2), in conclusion, we get:
Var g ^ N s ( t ) = g s ( t ) n θ H ϕ s θ K K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 H 2 ( v ) d v + o 1 n θ H ϕ s θ K Var G ^ N s ( t ) = G s ( t ) n ϕ s θ K K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 H 2 ( v ) d v + o 1 n ϕ s θ K
For the second result about Var ( G ^ D s ) , keeping the same notation with respect the definition of ( G ^ D s ) in (5), we set
Var G ^ D s = 1 n E K 1 ( s ) 2 Var i = 1 n K i ( u ) = 1 n E K 1 ( s ) 2 Var K 1 ( s ) + 1 n E K 1 ( s ) 2 i = 1 n j = 1 i j n Cov K i ( s ) , K j ( s ) = V K 1 + V K 2
Moreover,
V K 1 = E ( K 1 2 ( s ) ) E K 1 ( s ) 2 n E K 1 ( s ) 2 = 1 n E ( K 1 2 ( s ) ) E K 1 ( s ) 2 1 = 1 n K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u ϕ s ( θ K ) K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 1 + o 1 n ϕ s ( θ K ) .
Furthermore, the 2nd term V K 2 , we decompose as follows:
i = 1 n j = 1 i j n Cov K i ( s ) , K j ( s ) = i = 1 n j = 1 0 < | i j | m n n Cov K i ( s ) , K j ( s ) + i = 1 n j = 1 | i j | > m n n Cov K i ( s ) , K j ( s ) = : J 1 + J 2 .
Now, under the assumption ( H 6 ), we have
J 1 = i 0 < | i j | m n Cov K i ( s ) , K j ( s ) n m n max i j E K i ( s ) K j ( s ) + E K 1 ( s ) 2 α n m n ϕ s 2 θ K
From the condition ( H 5 ), we infer that
J 2 = i | i j | > m n Cov K i ( s ) , K j ( s ) α Lip ( K ) λ K 2 i | i j | > m n χ i , j α n θ K 2 e a m n
This implies that
i = 1 n j = 1 i j n Cov K i ( s ) , K j ( s ) i = 1 n i j Cov K i ( s ) , K j ( s ) α n m n ϕ s 2 θ K + θ K 2 e a m n
Next, taking
m n = 1 γ log γ θ K 2 ϕ s 2 θ K
which allows to write that:
V K 2 = 1 n ϕ s ( θ K ) 2 i = 1 n j = 1 i j n Cov K i ( s ) , K j ( s ) 0
Finally, we get:
Var G ^ D s = 1 n K 2 ( 1 ) 0 1 ( K ( u ) ) 2 β s ( u ) d u ϕ s θ K K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 1 + o 1 n ϕ s ( θ K ) .
Now we evaluate Cov G ^ D s , G ^ N s ( t ) as follows:
Cov G ^ D s , G ^ N s ( t ) = 1 n E K 1 ( s ) 2 Cov K 1 ( s ) , Γ 1 ( t ) + 1 n E K 1 ( s ) 2 i j Cov K i ( s ) , Γ j ( t ) = C V 1 + C V i j with Γ i ( s , t ) = K i ( s ) H i ( t )
Where
C V 1 = E K 1 2 ( s ) H 1 ( t ) n E K 1 ( s ) 2 E K 1 ( s ) H 1 ( t ) n E K 1 ( s )
For the first term in right-hand of (A31), we have
E K 1 2 ( s ) H 1 ( t ) = E K 1 2 ( s ) E H 1 ( t ) S = E K 1 2 ( s ) R H 1 t z θ H g S ( z ) d z = 1 θ H E K 1 2 ( s ) R H 1 t z θ H G S z d z = E K 1 2 ( s ) R H 1 v G S t θ H v d v = E K 1 2 ( s ) G S ( t ) + o 1 .
The first order Taylor expansion of G around t for n large enough justified the last line, furthermore, replacing K by K 2 in (A24), and following the same steps and techniques, then by the fact that Ψ 0 ( 0 ) = 0 , the second term E K 1 2 ( s ) Φ m ( d ( s , S ) ) 0 allow us to get:
E K 1 2 ( s ) H 1 ( t ) = G s ( t ) E K 1 2 ( s ) + o 1 = G s ( t ) ϕ s θ K K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u + o ϕ s θ K
Then,
E K 1 2 ( s ) H 1 ( t ) n E K 1 ( s ) 2 = G s ( t ) n ϕ s θ K K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 + o 1 n ϕ s θ K
By the same technique used above in (A32) and (A33), we can write:
E K 1 ( s ) H 1 ( t ) n E K 1 ( s ) = G s ( t ) n + o 1 n = O 1 n
Finally, by (A34) and (A35), we infer
C V 1 = 1 n E K 1 ( s ) 2 Cov K 1 ( s ) , Γ 1 ( t ) = G s ( t ) K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u n ϕ s θ K K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 G s ( t ) n + o 1 n ϕ s θ K
Furthermore, for the 2nd term C V i j , we follow the steps used to analysis Var τ = 1 n Z n ^ τ in lemma 1, we divide the sum below:
i = 1 n j = 1 i j n Cov K i ( s ) , Γ j ( s , t ) = i = 1 n j = 1 0 < | i j | m n n Cov K i ( s ) , Γ j ( s , t ) + i = 1 n j = 1 | i j | > m n n Cov K i ( s ) , Γ j ( s , t ) = : P 1 + P 2 .
We keep the same notation and under hypothesis ( H 1 ), ( H 3 ) and ( H 6 ), we infer, i j
P 1 = i = 1 n j = 1 0 < | i j | m n n Cov K i ( s ) , Γ j ( s , t ) i = 1 n j = 1 0 < | i j | m n n E K i ( s ) Γ j ( s , t ) α i = 1 n j = 1 0 < | i j | m n n E K j H j α i = 1 n j = 1 0 < | i j | m n n E K j E ( H j S ) θ n m n ϕ s 2 θ K
Both K and H are bounded, so :
P 2 = i = 1 n j = 1 | i j | > m n n Cov K i ( s ) , Γ j ( s , t ) Lip ( K ) θ K 2 + Lip ( H ) θ H i = 1 n j = 1 | i j | > m n n χ i , j α n Lip ( K ) θ K 2 + Lip ( H ) θ H χ m n α n Lip ( K ) θ K 2 + Lip ( H ) θ H e a m n .
Then, by (A37) and (A38), we get
i j Cov K i ( s ) , Γ j ( s , t ) α n m n ϕ s 2 θ K + Lip ( K ) θ K 2 + Lip ( H ) θ H e a m n
Taking m n = 1 γ log θ K 1 Lip ( K ) 2 + θ H 1 Lip ( H ) γ ϕ s 2 θ K , we get:
C V i j = 1 n E ( K 1 ( s ) ) 2 i j Cov K i ( s ) , Γ j ( s , t ) 0
Combining the results (A36) and (A39) we get:
Cov G ^ D s , G ^ N s ( t ) = G s ( t ) K 2 ( 1 ) 0 1 K 2 ( u ) β s ( u ) d u n ϕ s θ K K ( 1 ) 0 1 K ( u ) β s ( u ) d u 2 G s ( t ) n + o 1 n ϕ s θ K
Proof of Lemma 3.
It is clear that the result of Lemma 2 and corollary 1 permits to write:
E G ^ D s G ^ N s ( t ) 1 + G s ( t ) 0
And
Var G ^ D s G ^ N s ( t ) 1 + G s ( t ) 0
Then, by Markov Inequality:
G ^ D s G ^ N s ( t ) 1 + G s ( t ) 0 in probability .
Finally, by combining this result with the fact that
E G ^ D s G ^ N s ( t ) 1 + E G ^ N s ( t ) = 0
we get the required result. □

Appendix B

Proof 
(Proof of Lemma 3). By the definition of g ^ N s ( t ) in (5), it follows that
n θ H ϕ s θ K g ^ N s ( t ) E g ^ N s ( t ) = τ = 1 n Z n τ ( s , t ) = V n
Where
Z n τ ( s , t ) = ϕ s θ K n θ H E K 1 ( s ) ( Γ τ ( s , t ) E Γ τ ( s , t ) )
And
Γ τ ( s , t ) = K τ ( s ) H τ ( t ) , s H , t R , 1 τ n
The result is:
V n N ( 0 , σ g 2 )
To do this, we employ Doob’s basic technique ( pages 228-231)[26]. We indeed choose two sequences of natural numbers that extend to infinity.
p = O ( n ϕ s ( θ K ) ) , q = o ( p )
And we divide V n into
V n = R n + R n + ζ k with R n = τ = 1 k η τ , and R n = τ = 1 k ξ τ
Where
η τ = τ I τ Z n τ ( s , t ) , ξ τ = τ J τ Z n τ ( s , t ) , ζ k = τ = k ( p + q ) + 1 Z n τ ( s , t )
With
I τ = ( τ 1 ) ( p + q ) + 1 , , ( τ 1 ) ( p + q ) + p J τ = ( τ 1 ) ( p + q ) + p + 1 , , τ ( p + q ) .
Remark that, for k = n p + q , . the integral part , we have k q n 0 , k p n 1 and q n 0 , which means as p n 0 as n . Our asymptotic result is now founded on:
E ( R n ) 2 + E ( ζ k ) 2 0
and
R n N ( 0 , σ g 2 )
Proof of (A41). Stationary gives us:
E ( R n ) 2 = k Var ( ξ 1 ) + 2 1 τ < s k Cov ( ξ τ , ξ s )
And
k Var ( ξ 1 ) q k Var ( Z n 1 ( s , t ) ) + 2 k 1 τ < s q Cov ( Z n τ ( s , t ) , Z n s ( s , t ) )
Using (A10) and the fact that k q n 0 . we set
q k Var ( Z n 1 ( s , t ) ) = q k ϕ s θ K 1 n θ H E 2 K 1 ( s ) Var ( Γ 1 ( s , t ) ) = O ( k q n ) 0 , as n .
In the other side, we have
k 1 τ < j q Cov ( Z n τ ( s , t ) , Z n j ( s , t ) ) = k ϕ s θ K n θ H E 2 K 1 ( s ) 1 τ < j q Cov ( Γ τ ( s , t ) , Γ j ( s , t ) )
Similarly to (A11), we write for this last covariance;
τ = 1 q j = 1 τ j q Cov Γ τ ( s , t ) , Γ j ( s , t ) = τ = 1 n j = 1 0 < | τ j | m n q Cov Γ τ ( s , t ) , Γ j ( s , t ) + τ = 1 q j = 1 | τ j | > m n q Cov Γ τ ( s , t ) , Γ j ( s , t ) = I q + I I q
Thus, by the same steps using to evaluate (A11) we obtain:
1 τ < j q Cov ( Γ τ ( s , t ) , Γ j ( s , t ) ) = o ( q θ H ϕ s θ K ) .
Then
k 1 τ < j k Cov ( Z n τ ( s , t ) , Z n j ( s , t ) ) = O ( k q n ) 0 , as n .
From (A44) - (A48) we get:
k Var ( ξ 1 ) 0 , as n .
For the second term of (A43), we use the stationary to evaluate the right-hand side.
1 τ j k Cov ( ξ τ , ξ j ) = m = 1 k 1 ( k m ) Cov ( ξ 1 , ξ m + 1 ) k m = 1 k 1 Cov ( ξ 1 , ξ m + 1 ) k m = 1 k 1 ( τ , j ) J 1 * J m + 1 Cov Z n τ ( s , t ) , Z n j ( s , t )
For all ( τ , j ) J 1 * J s , we have | τ j | p + 1 > p , then
1 τ < j k Cov ( ξ τ , ξ j ) k α ϕ s θ K ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 n θ H E 2 K 1 ( s ) τ = 1 p j = 2 p + q + 1 τ j > p k ( p + q ) χ τ , j θ k p ϕ s θ K ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 n θ H E 2 K 1 ( s ) χ p α k p ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 n θ H ϕ s θ K e a p α k p n θ H 3 ϕ s 3 θ K e a p 0 .
By this last result and (A49) we set:
E ( R 1 ) 2 0 as n .
By the fact ( n k ( p + q ) ) p we get for the sequence ζ k ;
E ( ζ k ) 2 ) ( n k ( p + q ) ) Var ( Z n 1 ( s , t ) ) + 2 1 τ < j k Cov ( Z n τ ( s , t ) , Z n j ( s , t ) ) p Var ( Z n 1 ( s , t ) ) + 2 1 τ < j k Cov ( Z n τ ( s , t ) , Z n j ( s , t ) ) p ϕ s θ K n θ H E 2 K 1 ( s ) Var ( Γ 1 ( s , t ) ) + α ϕ s θ K n θ H E 2 K 1 ( s ) 1 τ < j k Cov ( Γ τ ( s , t ) , Γ j ( s , t ) ) o ( 1 ) α p n + o ( 1 ) .
Hence,
E ( ζ k ) 2 0 as n
Combining with (A49) the proof of (A41) is complete. □
Proof of (A42). based in the following two results
E ( e i t j = 1 k η j ) j = 1 k E ( e i t η j ) 0
And
k Var ( η 1 ) σ g 2 ; k E ( η 1 2 I { η 1 > ϵ σ g 2 } ) 0
roof of (A52).
E ( e i t j = 1 k η j ) j = 1 k E ( e i t η j ) E ( e i t j = 1 k η j ) E ( e i t j = 1 k 1 η j ) E ( e i t η k ) + E ( e i t j = 1 k 1 η j ) j = 1 k 1 E ( e i t η j ) = Cov ( e i t j = 1 k 1 η j , e i t η k ) + E ( e i t j = 1 k 1 η j ) j = 1 k 1 E ( e i t η j )
And successively, we have:
E ( e i t j = 1 k η j ) j = 1 k E ( e i t η j ) Cov ( e i t j = 1 k 1 η j , e i t η k ) + Cov ( e i t j = 1 k 2 η k 1 , e i t η j ) + + Cov ( e i t η 2 , e i t η 1 ) .
Once again we apply Lemma 1, to write:
Cov ( e i t η 2 , e i t η 1 ) α ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 ϕ s θ K n θ H E 2 K 1 ( s ) τ I 1 j I 2 χ τ , j .
Applying (A56) to each term on the right-hand side of (A55):
E ( e i t j = 1 k η j ) j = 1 k E ( e i t η j ) α t 2 ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 ϕ s θ K n θ H E 2 K 1 ( s ) × τ I 1 j I 2 χ τ , j + τ I 1 I 2 j I 3 χ τ , j + + τ I 1 . . . I k 1 j I k χ τ , j .
For all 2 r k 1 , ( τ , j ) I r * I r + 1 , we have | τ j | q + 1 > q , then
τ I 1 I k 1 j I k χ τ , j p χ q .
Therefore, inequality (A54) becomes
E ( e i t j = 1 k η j ) j = 1 k E ( e i t η j ) = α t 2 ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 ϕ s θ K n θ H E 2 K 1 ( s ) k p χ q = α t 2 ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 ϕ s θ K n θ H E 2 K 1 ( s ) k p e a q = α t 2 ( θ K 1 L i p ( K ) + θ H 1 L i p ( H ) ) 2 1 n θ H ϕ s θ K k p e a q = α t 2 k p n θ H 3 ϕ s 3 θ K e a q 0 .
Proof of (A53). By the definition of η 1 and Z n 1 ( s , t ) , we have
k Var ( η 1 ) = k p Var ( Z n 1 ( s , t ) ) = k p ϕ s θ K n θ H E 2 K 1 ( s ) Var ( Γ 1 ( s , t ) )
Using the result of (A9) and the fact that k p n 1 , Which imply that
k Var ( η 1 ) σ g 2 .
For the second term of (A53), we conclude by η 1 α p Z n 1 ( s , t ) α p n θ H ϕ s θ K , and Tchebychev’s inequality to get:
k E ( η 1 2 I { η 1 > ϵ σ g 2 } ) α p 2 k n θ H ϕ s θ K P η 1 > ϵ σ g 2 α p 2 k n θ H ϕ s θ K Var ( η 1 ) ϵ 2 σ g 2 = O p 2 n θ H ϕ s θ K .
This concludes Lemma 3’s proof. □

References

  1. D. Bosq, J.B. Lecoutre, Théorie de l’estimation fonctionnelle, Ed. Economica, (1987).
  2. F. Ferraty, P. Vieu, Nonparametric functional data analysis: Theory and practice, Springer Series in Statistics, Springer, Berlin (2006) ISBN 978-0-387-30369-7.
  3. F. Ferraty, A. Mas, And P. Vieu, Nonparametric regression on functional data: inference and practical aspects, Australian and New Zealand Journal of Statistics, 49 (2007), 267–286. [CrossRef]
  4. A. Lakcasi, B. Mechab, Conditional hazard estimate for functional random fields., Journal of Statistical Theory and Practice, 08 (2014), 192-220. [CrossRef]
  5. F. Ferraty, P. Vieu, Nonparametric models for functional data, with application in regression, time series prediction and curve discrimination, Journal of Nonparametric Statistics, 16 (2004), 111–125. [CrossRef]
  6. F. Ferraty, A. Goia, P. Vieu, Régression non-paramétrique pour des variables aléatoires fonctionnelles mélangeantes, Comptes Rendus Mathematique, 334 (2002), 217-220. [CrossRef]
  7. W. Mechab, A. Lakcasi, Nonparametric relative regression for associated random variables, Metron, 74 (2016), 75-97. [CrossRef]
  8. Azzi, A., Belguerna, A., Laksaci, A., & Rachdi, M. (2023). The scalar-on-function modal regression for functional time series data. Journal of Nonparametric Statistics, 36(2), 503–526. [CrossRef]
  9. R.J. Hyndman, Q. Yao, Nonparametric estimation and symmetry tests for conditional density functions, Journal of Nonparametric Statistics, 14 (2002), 259–278. [CrossRef]
  10. H. Daoudi, A. Belguerna, Z. C. Elmezouar, F. Alshahrani, Conditional Density Kernel Estimation Under Random Censorship for Functional Weak Dependence Data. Journal of mathematics. 2159604 (2025). [CrossRef]
  11. S. Attaoui, A. Laksaci, E. Ould Said, A note on the conditional density estimate in the single functional index model, Statistics and Probability Letters, 81 (2011), 45–53. [CrossRef]
  12. N. Ling, Q. Xu, Asymptotic normality of conditional density estimation in the single index model for functional time series data, Statistics and Probability Letters, 82 (2012), 2235-2243. [CrossRef]
  13. K. Abdelhak, A. Belguerna, and Z Laala, Local Linear Estimator of the Conditional Hazard Function for Index Model in Case of Missing at Random Data, Applications and Applied Mathematics: An International Journal (AAM), 17(1), (2022). https://digitalcommons.pvamu.
  14. H. Daoudi, B. Mechab, Asymptotic Normality of the Kernel Estimate of Conditional Distribution Function for the quasi-associated data, Pakistan Journal of Statistics and Operation Research, 15 (2019), 999–1015. [CrossRef]
  15. A. Bulinski, C. Suquet, Asymptotical behaviour of some functionals of positively and negatively dependent random fields, Fundamentalnaya i Prikladnaya Matematika, 04 (1998), 479-492. https://www.mathnet.
  16. I. Bouaker, A. Belguerna, H. Daoudi, The Consistency of the Kernel Estimation of the Function Conditional Density for Quasi-Associated Data in High-Dimensional Statistics, Journal of Science and Arts, 22 (2022), 247–256. [CrossRef]
  17. C.M. Newman, Asymptotic independence and limit theorems for positively and negatively dependent random variables : Inequalities in Statistics and Probability, Lecture Notes-Monograph Series, 05 (1984), 127-140. http://www.jstor.org/stable/4355492.
  18. F. Ferraty, A. Rabhi And F. Vieu, Estimation nonparametric de la fonction de hasard avec variable explicative fonctionnelle, Revue roumaine de mathématiques pures et appliquées, 53 (2008), 1-18.
  19. A. Laksaci, B. Mechab, Estimation non-paramétrique de la fonctionde hasard avec variable explicative fonctionnelle : cas des données spatiales, Revue roumaine de mathématiques pures et appliquées, 55 (2010), 35–51.
  20. A. Gagui, A. Chouaf, On the nonparametric estimation of the conditional hazard estimator in a single functional index, Statistics in Transition new series, 23 (2022), 89–105. [CrossRef]
  21. A. Belguerna, H. Daoudi, K. Abdelhak, B. Mechab, Z. Chikr Elmezouar, F. Alshahrani, A Comprehensive Analysis of MSE in Estimating Conditional Hazard Functions: A Local Linear, Single Index Approach for MAR Scenarios, Mathematics, 12 (2024), 495. [CrossRef]
  22. P. Doukhan, S. Louhichi, A new weak dependence condition and applications to moment inequalities, Stochastic Processes and their Applications, 84 (1999), 313–342. [CrossRef]
  23. A. Bulinski, C. Suquet, approximation for quasi-associated random fields., Statistics and Probability Letters, 54 (2001), 215–226. [CrossRef]
  24. R.S. Kallabis, M.H. Neumann, An exponential inequality under weak dependence, Bernoulli, 12 (2006), 333–350.
  25. M. Bassoudi, A. Belguerna, H. Daoudi And Z. Laala, Asymptotic properties of the conditional hazard function estimate by the local linear method for functional ergodic data, journal of Applied Mathematics and Informatics, 41 (2023), 1341-364. [CrossRef]
  26. S. Attaoui, Sur l’estimation semi paramètrique robuste pour statistique fonctionnelle, Université du Littoral Côte d’Opale Sciences pour l’Ingénieur (Lille); Université Djillali Liabès (Sidi Bel-Abbès, Algérie), (2012).
  27. H. Tabti, A. Ait Saidi, Estimation and simulation of conditional hazard function in the quasi-associated framework when the observations are linked via a functional single-index structure, Communications in Statistics - Theory and Methods, 47 (2017), 816-–838. [CrossRef]
  28. L. Douge, Thèorèmes limites pour des variables quasi-associes hilbertiennes, annales de l’ISUP, 04 (2010), 51–60.
  29. H. Daoudi, B. Mechab, Z. Chikr Elmezouar, Asymptotic normality of a conditional hazard function estimate in the single index for quasi-associated data, Communications in Statistics - Theory and Methods, 49 (2020), 513–530. [CrossRef]
  30. H. Daoudi, Z. Chikr Elmezouar, F. Alshahrani, Asymptotic Results of Some Conditional Nonparametric Functional Parameters in High Dimensional Associated Data, Mathematics, 11 (2023), 4290. [CrossRef]
  31. S. Bouzebda, A. Laksaci, M. Mohammedi, Single Index Regression Model for Functional Quasi-Associated Times Series Data, REVSTAT-Statistical Journal, 20 (2023), 605–631. [CrossRef]
  32. A. Rassoul, A. Belguerna, H. Daoudi, Z. C. Elmezouar, F. Alshahrani, On the Exact Asymptotic Error of the Kernel Estimator of the Conditional Hazard Function for Quasi-Associated Functional Variables, mathematics, 13(13) (2025).
  33. H. Daoudi, B. Mechab, S. Benaissa, A. Rabhi, Asymptotic normality of the nonparametric conditional density function estimate with functional variables for the quasi-associated data. Int. J. Stat. Econ. 20 (2019). http://www.ceser.in/ceserp/index.php/bse/article/view/6131.
  34. H. Daoudi, B. Mechab and A. Belguerna, Asymptotic Results of a Conditional Risk Function Estimate for Associated Data Case in High-Dimensional Statistics, Asymptotic Results of a Conditional Risk Function Estimate for Associated Data Case in High-Dimensional Statistics, Tebessa, Algeria, 2021, pp. 1-6. [CrossRef]
  35. S. Attaoui, O. Benouda, S. Bouzebda, A. Laksaci, Limit theorems for kernel regression estimator for quasi-associated functional censored time series within single index structure, Mathematics, 13 (2025). [CrossRef]
  36. J. L. Doob, Stochastic Processes, Wiley, New York (1953), pp. 228-231.
Figure 1. S κ = 1 , , n ( n = 50 )
Figure 1. S κ = 1 , , n ( n = 50 )
Preprints 173866 g001
Figure 2. S κ = 1 , , n ( n = 200 )
Figure 2. S κ = 1 , , n ( n = 200 )
Preprints 173866 g002
Figure 3. S κ = 1 , , n ( n = 1000 )
Figure 3. S κ = 1 , , n ( n = 1000 )
Preprints 173866 g003
Figure 4. Conditional distribution
Figure 4. Conditional distribution
Preprints 173866 g004
Figure 5. Conditional density
Figure 5. Conditional density
Preprints 173866 g005
Figure 6. Normal approximation of the conditional hazard function estimator.
Figure 6. Normal approximation of the conditional hazard function estimator.
Preprints 173866 g006
Figure 7. Empirical and theoretical conditional hazard function estimation with confidence bounds.
Figure 7. Empirical and theoretical conditional hazard function estimation with confidence bounds.
Preprints 173866 g007
Table 1. Mean square error of our estimators.
Table 1. Mean square error of our estimators.
Mean Square Error n=50 n=200 n=1000
MSE( g ^ ) 1.11645 × 10 04 8.028238 × 10 05 7.623378 × 10 05
MSE( G ^ ) 1.017427 × 10 04 3.718499 × 10 05 3.076726 × 10 05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated