1. Introduction
Nonparametric density estimation is a major issue of statistics and econometrics. Although the kernel density estimator (KDE) of location-scale type, originally proposed by Rosenblatt (1956) and Parzen (1962), is perhaps the most popular in the literature, its boundary bias problem is an important matter of concern when the support of the density to be estimated is not the whole real line . Various remedies for avoiding the boundary bias problem have been discussed, on the basis of renormalization, reflection, and generalized jackknifing (Jones (1993)), transformation (Marron and Ruppert (1994)), advanced reflection (Zhang et al. (1999)), and so on. Recently, there has been a vast literature on this subject, especially, with the renewed interest in revisiting an asymmetric kernel method after Chen (1999, 2000).
For the standard statistical inference problem of nonnegative data, it is assumed that a random sample of a size n is drawn from a population with density , . Practically, one encounters many situations where the available data are observed only under a certain biased sampling scheme (see, e.g., Patil and Rao (1978)). Then, it may be reasonable to regard a functional of the biased distribution as the inferential target. But, some analyses need the estimation referring to the original distribution. Cox (1969) considered estimating the population mean () and cumulative distribution function , . An overview of nonparametric functional estimation under the biased sampling scheme is found in Cristóbal and Alcalá (2001).
Nonparametric density estimation from the biased data started in the late 1980s. Two important papers often quoted are Bhattacharyya et al. (1988) and Jones (1991) (see also Richardson et al. (1991) and Guillamón et al. (1998)). Jones (1991) studied a kernel smoothing on the basis of Cox’s (1969) distribution estimator, without cares about the boundary bias problem of the classical Rosenblatt–Parzen location-scale KDE. The main contribution of this paper is to revisit the nonparametric density estimation under the biased sampling scheme, using asymmetric kernel method which enables us to avoid the boundary bias problem and then have desirable asymptotic properties. Our approach is different from Mnatsakanov and Ruymgaart (2003,2006) on moment-type density estimation motivated by the so-called moment problem, and Chaubey et al. (2010) using Hill’s lemma (Feller (1971; (1.5) of page 220)).
The rest of this paper is organized as follows.
Section 2 describes the length-biased (LB) distribution and illustrates, in detail, the boundary behavior of a convolution integral. After a brief introduction of the asymmetric kernel method, two density estimators from the LB data are proposed, in parallel with those of Bhattacharyya et al. (1988) and Jones (1991).
Section 3 and
Section 4 state the required assumptions and main results of this paper. All proofs are given in the appendix. In
Section 5, some simulations illustrate the finite sample performance of the density estimators.
As usual, we use the notation
for any bounded function
h on
. We write, for
,
(if it exists), and
. For an estimator
of
, where
, the mean squared error (MSE) is defined by
and the mean integrated squared error (MISE);
is a global measure of discrepancy of
from
f.
2. Preliminaries
2.1. LB Density
Nonparametrically, we wish to estimate the density
f with nonnegative support, in a situation where a random sample
is not directly available but a sample
(say) is instead observed from the LB distribution having the density
Throughout this paper, if there is no confusion, the
s are iid copies of the random variable
X having the density
f, whereas the
’s are iid copies of the random variable
Y having the LB density
. We repeatedly use the fact that, for a measurable function
G and a real number
r,
2.2. Boundary Bias Problem and Asymmetric Kernel Method
When , it is well known that a usual approximation (near the origin) of a certain smooth nonnegative function g as the convolution integral does not hold when . We emphasize that the following fact is a starting point in the present paper: Even for the case , the convergence rate near the origin is slower when ; a typical example is .
More precisely, if
k is a symmetric density on
(say) and
is a bandwidth which tends to zero as
(hereafter, we will omit the phrase “as the sample size
n tends to the infinity” unless otherwise stated, and we denote by
x the location where the density estimation is made), then, as shown in, e.g., Jones (1993),
according to
or
(
), because the kernel
creates a mass outside
when the location
x is at or near the origin. This is the motivation that, instead of the location-scale kernel
, we focus on an application of an asymmetric kernel
, whose support matches the support of the target function, where
is a smoothing parameter, with
. It should be remarked that, after Chen (2000), the notation
(rather than
h) is common in the asymmetric kernel method; indeed, the parameter
for the asymmetric kernel under consideration has no meaning of the bandwidth
h as in the classical Rosenblatt–Parzen location-scale kernel with a compact support, although
corresponds to
and controls the bias-variance trade-off. This is why we refer to
as a smoothing parameter (rather then a bandwidth).
We formally require that
if
. To the best of our knowledge, Silverman (1986; page 28) first mentioned a fairly simple idea of using gamma and log-normal (LN) kernels for the nonnegative data, where the kernel shape varies according to
. Perhaps, Chen’s (2000) gamma kernel, defined by
1
plays a central role in this area; however, some simulations reveal that when the the target density at the origin is zero, the gamma kernel is disadvantage compared to the LN kernel (e.g., Igarashi (2016)). Subsequent authors have discussed various kernels like LN, inverse Gaussian (IG), reciprocal IG (RIG), Birnbaum–Saunders (BS), and inverse gamma. The analysts can now choose what they like, among many options available for the kernel with support
. According to Igarashi and Kakizawa (2020) (see also Kakizawa (2021)), let us choose
in the following form:
Definition 1 (Igarashi and Kakizawa (2020)). Given a baselined density
, we set
where the functional form of
p, with nonnegative support, is independent of
and
x.
To make this formulation clear, let
be a symmetrical-based qBS density, associated with a density generator
g, where
,
,
(due to the fact , it is enough to take ). Note that the density on is symmetric about the origin, where . Here, it is common to standardize g so that , without loss of generality. Indeed, as long as for some constant , this standardization can always be imposed with the replacement of by ; the nomalizing constant is then invariant, i.e., .
Given constants
and
, a family of symmetrical-based non-central qBS kernels (Kakizawa (2021))
2 is defined by
where
. Kakizawa (2018) considered the central case
. Such a family
is flexible via the (infinite-dimensional) density generator
g, as well as the parameter
. In some numerical studies of
Section 5, we will put
(symmetrical-based BS kernel) or
(log-symmetrical (LS) kernel), with
, for simplicity, and use the power exponential (PE) generator;
,
, where the particular choice
ensures that the PE density has the variance 1, like the standard normal density. Other generators (Kotz-type, generalized Pearson-type VII, and generalized logistic-type III) are found in Kakizawa (2018,2021). A symmetrical-based central qBS kernel belongs to a family of symmetrical-based qMIG kernels (MIG is an abbreviation of a mixture of IG and RIG), which is enlarged, linking to a class of skew-BS type kernels. See Kakizawa (2018,2021).
Remark 1. Chaubey and Li (2013) applied Scaillet’s (2004) RIG kernel. As pointed out by Igarashi and Kakizawa (2014), an RIG KDE, however, also suffered from the boundary bias problem, so that the re-formulated RIG kernel should be applied.
2.3. Two density estimators under LB sampling scheme
Using
, as well as the relation
our first estimator is defined in parallel with that of Bhattacharyya et al. (1988), as follows:
Except for a technical issue that needs to take a small
, this estimator may be natural in the sense that
,
, is consistently estimable by the asymmetric KDE
(Bhattacharyya et al. (1988) used the classical Rosenblatt–Parzen KDE
). On the other hand, Jones’s (1991) idea;
(see Cox (1969)) is also reasonable. However, in order to solve its boundary bias problem unless
(see Introduction), our second estimator
is proposed, for which some asymptotic properties will be studied in Subsection 4.1.
Before we proceed with description of our required assumptions (
Section 3) and specific asymptotic results (
Section 4), we highlight a novelty, compared with Jones (1991). Suppose that
but
(we have an example
,
). Then,
We notice that is more preferable, since has the factor (it is numerically unstable near the origin); besides, a rigorous error analysis for the (unweighted) MISE of seems to be hard technically (or it might be impossible), although the pointwise MSE (for ) and weighted MISE of are tractable (see Subsection 4.2).
3. Assumptions
We use the notation . In order to prove asymptotic properties of (Subsection 4.1), the following set of assumptions, labeled as F, is imposed for the density f to be estimated:
- F.
-
(i) 1. f is a twice continuously differentiable function on , where , and are bounded;
2. is a Hölder-continuous function (with exponent ) on , i.e., there exists a constant , such that for any .
(ii) 1. is a bounded function on ;
2. is a Hölder-continuous function (with exponent ) on .
(iii) the inverse moment of X; (say) exists (note that ).
(iii) (say) exists for some constant .
(iv) , and exist.
Remark 2. Under the boundedness of , given in F(ii.1), the density f to be estimated must have a constraint (note that F(iii) for some implies ). However, as illustrated earlier, even in the case , Jones’s (1991) estimator suffers from the boundary bias problem when (we have an example , ).
On the other hand, for (Subsection 4.2), we additionally make some assumptions on the corresponding LB density, labeled as F:
- F.
-
1. In addition to F(i.1),
, and
are bounded functions
3 on
;
2. is a Hölder-continuous function (with exponent ) on .
Lastly, high-level conditions on the kernel
, labeled as A, are needed, whose details will be given in the top of Appendix. For notational simplicity, we write
where the constants
, and
(
) appear in Assumption A1.2–3 of Appendix. It is convenient for us to define
,
, and
. Obviously, the following inequalities hold:
It should be remarked that most of the items in Assumption F (or F
) are needed to approximate
, whose error analyse is found in Appendix. Roughly speaking, we obtain
under F(i) and
under F(ii), for the estimator
. Also, for the estimator
, it is shown that, under F
,
where
. Note that
We can see that, under F(i.1 and iv),
(say) and
(say) are well-defined; besides
4,
(say) is well-defined (we assume F(ii.1)).
4. Main Results
We assume that
- B().
, where ℓ is a (positive) slowly varying function.
Note that all powers of and a function approaching a positive limit vary slowly. For the achievement of the optimal rate of the M(I)SE, must be feasible, at least, where is a constant, independent of n.
In what follows, let
, where
is a constant, independent of
n. We write
where
is the indicator function of the set
S.
4.1. Asymptotic Properties of
Theorem 1.
Suppose that AssumptionsA1, A2.1(
)
; see Appendix, and F(i–iii)
hold. UnderB(
, given constants and , we have
where are constants, independent of , and x. Also, we have
Remark 3. (i) The asymptotic bias and variance of when x is near the origin; , where is finite, can be obtained, as in Kakizawa (2018,2021). The details are omitted.
(ii) The pointwise MSE of
is a corollary of Theorem 1(2), as follows: For fixed
,
where
(the equality holds iff ).
(iii) We have ( if ).
Theorem 2.
Suppose that AssumptionsA1, A2.1(); see Appendix, andF(i–iii)hold. UnderB(, we have for fixed (note that ).
Theorem 3.
Suppose that AssumptionsA1, A2.2(); see Appendix, andF(i–iii)hold.
(i)
UnderB(
)
, we have
(ii)
UnderB(
)
, we have
For fixed
, the replacement of
by
is a routine by combining Theorem 3(i) with a bias of Theorem 1(2): If Assumption F(iii
) holds for some
, and if
(for the extreme case
, assume
), then,
Theorem 4.
Suppose that Assumptions A1, A2.1(
), A3(
; see Appendix, and F
hold, where exists for some constant . UnderB(
, we have
(the equality holds iff (say)).
4.2. Asymptotic Properties of
Due to the presence of the factor in the estimator , the case is excluded throughout this subsection; besides, in Theorem 8 (see also Remark 5), we will consider a truncated MISE of , for the global performance.
Theorem 5.
Suppose that AssumptionsA1, A2.1(
)
; see Appendix,F(ii.1 and iii)
, andF
hold. UnderB(
, given constants and , we have, for ,
where are constants, independent of , and x.
Remark 4. As a corollary of Theorem 5, we have, for fixed
,
where
(the equality holds iff ).
Theorem 6.
Suppose that AssumptionsA1, A2.1(); see Appendix,F(ii.1 and iii), andFhold. UnderB(, we have for fixed .
Theorem 7.
Suppose that AssumptionsA1, A2.2(
)
; see Appendix,F(ii.1 and iii)
, andF
hold. UnderB(
)
, we have
For fixed
, the replacement of
by
is a routine by combining Theorem 7 with a bias of Theorem 6: If
(for the extreme case
, assume
), then,
Theorem 8.
Suppose that AssumptionsA1, A2.1(
), A3(
; see Appendix,F(ii.1, iii, and iv)
, andF
hold, where exists for some constant . UnderB(
, we have, for every ,
(the equality holds iff ).
Remark 5. Whether or not there exists a , such that is order (under some additional unnecessary stronger conditions) would be rather technical. We do not pursue the issue any more.
5. Simulation Studies
To demonstrate the finite sample performance of the proposed density estimator
, we generated 1000 random samples of size
from the LB density
, and computed the PE
-based BS/log KDEs (
) and gamma KDE for the original density
. In the simulation, we used the least squared cross-validated (LSCV) smoothing parameter, for each sample. Then, the average integrated squared errors (ISEs),
, were reported in
Table 1, where
is computed from the
ℓth sample.
As expected, all average ISEs decreased as the sample size
n increased, which is in agreement with the MISE result. The BS/LN KDEs (
) were, overall, improved by the estimators with
; such a tendency can be illustrated via the AMISE relative efficiency index
(see Kakizawa (2018,2021)), since
Needless to say, the best implemented smoothing parameter
, given in Theorem 4, depends on the unknown
f, so that the data-driven procedure is crucial. We tried to conduct the LSCV smoothing parameter selection
5. Unlike the direct sample (Kakizawa (2018,2021)), the present LB setting, for small sample size
(not being reported here), produced multiple local minima for the LSCV score (in many cases, it was rather unstable numerically), whereas such an undesirable behavior seemed to be fixed when
. A further issue of considering a plug-in selection with a pilot estimator is left in future.
6. Discussion
Our asymptotic results under the LB sampling can be extended to more general biased sampling, i.e., a weighted distribution for a known (positive) weight function , given by . The LB density is a special case of , and another example is (the area-biased density). Also, the d-variate weighted density is defined by , . Ahmad (1995) extended Jones’s (1991) estimator to the d-variate case. Note that the product-kernel-method, using the product asymmetric kernel , , instead of , or the non-product kernel (Igarashi (2018) and Kakizawa (2022)), can be straightforwardly applied to solve the boundary bias problem of Ahmad’s (1995) estimator.
Funding
The author has been supported in part by the Japan Society for the Promotion of Science; Grant-in-Aid for Scientific Research (C), 20K11700 and 23K11002.
Data Availability Statement
Not available.
Acknowledgments
Some preliminary results were first announced, without face-to-face talking (due to the pandemic of covid-19), at Japanese Joint Statistical Meeting 2021 (Japanese federation of statistical science associations), autumn meeting 2021 (Mathematical Society of Japan), and the 5th International Conference on Econometrics and Statistics (EcoStat2022).
Conflicts of Interest
The author declares no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
| KDE |
kernel density estimator |
| LB |
length-biased |
| MSE |
mean squared error |
| MISE |
mean integrated squared error |
| LN |
log-normal |
| IG |
inverse Gaussian |
| RIG |
reciprocal inverse Gaussian |
| BS |
Birnbaum–Saunders |
| LS |
log-symmetrical |
| PE |
power exponential |
| MIG |
mixture of IG and RIG |
| LSCV |
least squared cross-validated |
| ISE |
integrated squared error |
| ROT |
rule of thumb |
Appendix A
Appendix A.1. Technical Conditions on k(·;β,x)
There are three indispensable requirements on the kernel , :
(I) approximations of and ;
(II) uniform/nonuniform bounds of ;
(III) behavior of the tail integral of with respect to x (the regularity as will be required only for dealing with an asymptotic expansion of the MISE rigorously).
More precisely, we assume (e.g., Igarashi and Kakizawa (2020) and Kakizawa (2021)):
- A1.
-
In addition to Definition 1, there exists a density , such that
- 1.
exists and, for any , () exists, where is a constant, independent of y;
- 2.
given constants
and
, for all sufficiently small
,
implies that
with
,
,
, where
(
) and
are constants, independent of
and
x;
- 3.
exists, and, given constants
and
, for all sufficiently small
,
implies that
where
are constants, independent of
and
x.
- A2().
-
satisfies:
- 1.
, where is a constant, independent of ;
- 2.
for , , where is a constant, independent of and x.
- A3(H).
-
Given a constant
, and for all sufficiently small
,
(assume that exists).
Most of the existing asymmetric kernels satisfy Assumptions A1, A2
, and A3(
H); see Igarashi and Kakizawa (2020) and Kakizawa (2021). For instance, the constants
, and
(given in A1.2–3), associated with
, are given by
,
, and
, independent of
, where
. Of course, we need to impose a set of requirements on the density generator
g, under which A1, A2(0) and A3(
H) hold for the asymmetric kernel
; see Kakizawa (2018,2021). For simplicity, we assume that there exist constants
such that
for every
. It remains to discuss A2
. Note that A2(
) is technically required to prove Theorems 1–4 (i.e., (
A8) and (A11) under A2(
)); indeed, (A11) is crucial for Lemma A3. On the other hand, A2(
) is enough for the proofs of Theorems 5–8.
Property A1.
The kernel satisfies Assumption A2(ν) for any , with
where .
Proof. Recall that, with
(note that
),
This, together with
, yields
It suffices to bound , , in the same manner as Kakizawa (2018). □
Appendix A.2. Auxiliary Lemmas
We mention (without proof) the following basic lemma; (ii) is a slight modification of Kakizawa (2021):
Lemma A1. (i)
Let g be a twice continuously differentiable function on , where , and are bounded; besides, is Hölder-continuous with exponent . Under AssumptionA1.1–2
, given constants and , we have, for all sufficiently small ,
with , where is a constant, independent of β and x.
(ii)
Let g be a bounded and Hölder-continuous function (with exponent ) on . Under AssumptionsA1
andA2(
)
, given constants and , we have, for all sufficiently small ,
with , where is a constant, independent of β and x.
We can verify that, according to, e.g., Igarashi and Kakizawa (2020) and Kakizawa (2021),
if
exists for some constant
, as follows:
- 1.
-
(i) For , using Lemma A1(i.1).
(ii) Take a constant
(
). Using Lemma A1(i.3), we have, for any
,
(iii) Under A3(
), we have, for
,
- 2.
-
(i) For , by A2.1().
(ii) Take a constant
(
). Using Lemma A1(ii.2), we have, for any
,
(iii) Under A3(
), we have, for
,
Appendix B
Appendix B.1. Some basic results for ζ ϵ
For ease of reference, we first mention the tail probability/moment bounds of
. Rewrite
where
, and
is the average of zero-mean independent random variables
with
,
(we also have
). Then, Bernstein’s inequality yields the exponential bound of the tail probability
(hence, if ).
Suppose that
exists. Using
, we have
Furthermore, it is easy to see that
Appendix B.2. Some preliminary results for ζ β (x)
We next list some facts about
, including the tail probability/moment bounds and asymptotic normality. Rewrite
where
is the average of zero-mean independent random variables
with
,
(we also have
). Assumption A2.1(
) and Bernstein’s inequality yield the exponential bound of the tail probability
(hence,
if
, which is implied by, e.g.,
for some constant
). On the other hand, F(iii
), A1.3, and A2.2(
) imply that, for fixed
(assume
), we have, under B(
) (note that
is feasible when F(iii
) holds for some
),
hence,
using Lyapunov’s theorem (for triangular arrays), together with
for fixed
, which will be shown in (A18) below. Similarly, F(iii
), A1.3, and A2.1(
) imply that, for the case
, we have, under B(
) (note that
is feasible when F(iii
) holds for some
(in this case,
), i.e.,
is unfortunately infeasible for
),
hence,
using Lyapunov’s theorem, together with
, which will be shown in (A14) below.
In addition to
,
Lemma A1 implies that, given a constant
,
(obviously,
and
), and that, for
,
Remark A1. If
exists for some constant
(this ensures
exists), in line with, e.g., Kakizawa (2021), Assumption A3(
) about the behavior of the tail integral of
with respect to
x is crucial for proving the negligibility of the integral
(
), as follows: We have, for any constant
,
Then, the approximation
can be verified, since, by taking
,
(we also take ; see the argument after Lemma A1).
Appendix B.3. Some Preliminary Results for ζ β † (x)
We here list some facts about
, including the tail probability/moment bounds and asymptotic normality. Rewrite
where
is the average of zero-mean independent random variables
with
,
. As in, e.g., Igarashi and Kakizawa (2020) and Kakizawa (2021), by Assumption A2.1(
), an application of Bernstein’s inequality yields the exponential bound of the tail probability
(hence,
if
), whereas, if
and
, then, A2.2(
) implies that, for fixed
(assume
),
hence,
using Lyapunov’s theorem, together with
for fixed
, which will be shown in (A26) below.
In addition to
and
Lemma A1 implies that, given a constant
, we have, for
,
where
.
Remark A2. As mentioned in Theorem 8, we take an arbitrary constant
and consider a weighted MISE, with a weight function
(say). In line with, e.g., Kakizawa (2021), if
exists for some constant
, Assumption A3(
), together with
(say) for any constant
, is crucial for proving the negligibility of
(
), as follows: We have, for any constant
,
We can verify that
for every
, since
Appendix B.4. Proofs of Main Results
Before proving Theorems 1–4, we prepare two lemmas (Lemmas A2 and A3):
Lemma A2.
Suppose that exists. Then,
where is a constant, independent of , and x.
Proof. It is easy to see that
where
□
Now, we rewrite (
A1) as
where
We need to evaluate .
Lemma A3.
Suppose that AssumptionsA2.1(
)
andB
hold, and that exists. Under the boundedness of f, we have
where is a constant, independent of , and x
Proof. Considering an event
(say), we have
Using
(say) for all sufficiently large
n, it can be shown that
Then, the result follows from (
A4), (
A6), (
A12), and Lemma A2, under Assumption B (recall that
; in this case,
, where
is a constant, independent of
n). □
We are ready to prove Theorems 1–4.
Proof (
Proof of Theorem 1)
It is shown from (
A28) and (
A30) that
where
is a constant, independent of
, and
x. Using Lemma A3 and
we have
where
are constants, independent of
, and
x. Then, using (
A12)–(
A18), the proof is completed. □
Proof (
Proof of Theorem 2)
Recall (
A1) (also (
A3) and (
A7)). The strong consistency follows from (
A4), (
A5), and (
A8), together with
for fixed
(we also have
). See Lemma A1(i.2–3). □
Proof (
Proof of Theorem 3)
By Lemma A3, we notice that
for fixed
, and that
. Recalling (
A29), where
, we have
The results (i) and (ii) follow from (
A9) and (
A10), respectively. □
Proof (
Proof of Theorem 4)
In the same way as the argument of Remark , a rigorous derivation of the MISE is made by splitting the integral into three parts (we set ), as follows:
- (i)
Theorem 1 yields for .
- (ii)
Take a constant
. The inequality
together with (
A6), enables us to see that, by (
A27),
- (iii)
Taking
and
, Theorem 1 yields
□
Before proving Theorems 5–8, we prepare two lemmas (Lemmas A4 and A5):
Lemma A4.
Suppose that exists. Then,
where is a constant, independent of , and x.
Proof. It is easy to see that
where
□
Now, (
A2) can be rewritten as
where
We need to evaluate .
Lemma A5.
Suppose that AssumptionsA2.1(
)
andB
hold, and that exists. Under the boundedness of , we have
where is a constant, independent of , and x.
Using
(say) for all sufficiently large
n, it can be shown that
Then, the result follows from (
A4), (
A6), (
A22), and Lemma A4, under Assumption B (recall that
; in this case,
, where
is a constant, independent of
n). □
We are ready to prove Theorems 5–8.
Proof (Proof of Theorem 5) We start with
where
It is shown from (
A32) and (
A34) that
where
is a constant, independent of
, and
x. Using Lemma A5 and
we have
where
are constants, independent of
, and
x. Then, using (
A24)–(
A26), the proof is completed. □
Proof Proof of Theorem 6)
Recall (
A2) (also (
A3) and (
A19)). The strong consistency follows from (
A4), (
A5), and (
A20), together with
for fixed
(see (
A24)). □
Proof (Proof of Theorem 7) Recall (
A33), where
. For fixed
, Lemma A5 implies that
, i.e.,
The asymptotic normality follows from (
A21)). □
Proof (
Proof of Theorem 8)
We assume that (say) exists for some constant (then, exists). In the same way as the argument of Remark , a rigorous derivation is made by splitting the integral into two parts (we set ), as follows:
- (i)
Take a constant
. The inequality
together with (
A6) and
(say), enables us to see that, by (
A31),
- (ii)
Taking
, Theorem 5 enables us to see that, for every
,
□
References
- Ahmad, I.A. On multivariate kernel estimation for samples from weighted distributions. Statist.Probab.Lett. 1995, 22, 121–129. [Google Scholar] [CrossRef]
- Bhattacharyya, B.B.; Franklin, L.A.; Richardson, G.D. A comparison of nonparametric unweighted and length-biased density estimation of fibres. Comm.Statist.Theory Methods 1988, 17, 3629–3644. [Google Scholar]
- Chaubey, Y.P.; Li, J. Asymmetric kernel density estimator for length-biased data. In Contemporary Topics in Mathematics and Statistics with Applications; Adhikari, A., Adhikari, M.R., Chaubey, Y.P., Eds.; Asian Books Private Ltd: New Delhi, 2013; pp. 1–28. [Google Scholar]
- Chaubey, Y.P.; Sen, P.K.; Li, J. Smooth density estimation for length-biased data. J.Indian Soc.Agricultual Statist. 2010, 64, 145–155. [Google Scholar]
- Chen, S.X. 1999. Beta kernel estimators for density functions. Comput.Statist.Data.Anal. 1999, 31, 131–145. [Google Scholar] [CrossRef]
- Chen, S.X. Probability density function estimation using gamma kernels. Ann.Inst.Stat.Math. 2000, 52, 471–480. [Google Scholar] [CrossRef]
- Cox, D.R. Some sampling problems in technology. In New Developments in Survey Sampling; Johnson, N.L., Smith, H.Jr., Eds.; Wiley: New York, 1969; pp. 506–527. [Google Scholar]
- Cristóbal, J.A.; Alcalá, J.T. An overview of nonparametric contributions to the problem of functional estimation from biased data. Test 2001, 10, 309–332. [Google Scholar] [CrossRef]
- Feller, W. An Introduction to Probability Theory and its Applications, 2nd ed.; Wiley: New York, 1971; Volume II. [Google Scholar]
- Guillamón, A. , Navarro, J., and Ruiz, J.M. 1998. Kernel density estimation using weighted data. M. 1998. Kernel density estimation using weighted data. Comm.Statist.Theory Methods 27 2123–2135.
- Igarashi, G. Weighted log-normal kernel density estimation. Comm.Statist.Theory Methods 2016, 45, 6670–6687. [Google Scholar] [CrossRef]
- Igarashi, G. Multivariate density estimation using a multivariate weighted log-normal kernel. Sankyhā 2018, 80, 247–266. [Google Scholar] [CrossRef]
- Igarashi, G.; Kakizawa, Y. Re-formulation of inverse Gaussian, reciprocal inverse Gaussian, and Birnbaum–Saunders kernel estimators. Statist.Probab.Lett. 2014, 84, 235–246. [Google Scholar] [CrossRef]
- Igarashi, G.; Kakizawa, Y. Multiplicative bias correction for asymmetric kernel density estimators revisited. Comput.Statist.Data.Anal. 2020, 141, 40–61. [Google Scholar] [CrossRef]
- Jones, M.C. Kernel density estimation for length biased data. Biometrika 1991, 78, 511–519. [Google Scholar] [CrossRef]
- Jones, M.C. Simple boundary correction for kernel density estimation. Stat.Comput. 1993, 3, 135–146. [Google Scholar] [CrossRef]
- Kakizawa, Y. Nonparametric density estimation for nonnegative data, using symmetrical-based inverse and reciprocal inverse Gaussian kernels through dual transformation. J.Statist.Plann.Inference 2018, 193, 117–135. [Google Scholar] [CrossRef]
- Kakizawa, Y. A class of Birnbaum–Saunders type kernel density estimators for nonnegative data. Comput.Statist.Data.Anal. 2021, 161, 107249. [Google Scholar] [CrossRef]
- Kakizawa, Y. Multivariate elliptical-based Birnbaum–Saunders kernel density estimation for nonnegative data. J.Multivariate.Anal. 2022, 187, 104834. [Google Scholar] [CrossRef]
- Marron, J.S. and Ruppert, D. 1994. Transformations to reduce boundary bias in kernel density estimation. J.R.Stat.Soc. 1994, 56, 653–671. [Google Scholar]
- Mnatsakanov, R.M.; Ruymgaart, F.H. Some properties of moment-empirical cdf’s with application to some inverse estimation problems. Math.Methods Statist. 2003, 12, 478–495. [Google Scholar]
- Mnatsakanov, R.M.; Ruymgaart, F.H. On moment-density estimation in some biased models. In Optimality: The Second Erich L. Lehmann Symposium; Institute of Mathematical Statistics, 2006; pp. 322–333. [Google Scholar]
- Parzen, E. On estimation of a probability density function and mode. On estimation of a probability density function and mode. Ann.Math.Statist. 33 1962, 1065–1076. [Google Scholar] [CrossRef]
- Patil, G.P.; Rao, C.R. Weighted distributions and size-biased sampling with applications to wildlife populations and human families. Biometrics 1978, 34, 179–189. [Google Scholar] [CrossRef]
- Richardson, G.D.; Kazempour, M.K.; Bhattacharyya, B.B. Length biased density estimation of fibres. J.Nonparametr.Stat. 1991, 1, 127–141. [Google Scholar] [CrossRef]
- Rosenblatt, M. Remarks on some nonparametric estimates of density functions. Ann.Math.Statist. 1956, 27, 832–837. [Google Scholar] [CrossRef]
- Scaillet, O. Density estimation using inverse and reciprocal inverse Gaussian kernels. J.Nonparametr.Stat. 2004, 16, 217–226. [Google Scholar] [CrossRef]
- Silverman, B.W. Density Estimation for Statistics and Data Analysis; Chapman and Hal: London, UK, 1986. [Google Scholar]
- Zhang, S. , Karunamuni, R.J., and Jones, M.C. 1999. An improved estimator of the density function at the boundary. C. 1999. An improved estimator of the density function at the boundary. J.Amer.Statist.Assoc. 94 1231–1241.
| 1 |
Clearly, the kernel is a rescaled version of the (standard) gamma density , with substitution of for the shape parameter . Here, the shape is limited to be greater than or equal to 1, so as to ensure that the resulting kernel is bounded. |
| 2 |
It is easy to see that
provided that is also a density generator. |
| 3 |
It automatically means that is a Lipschitz-continuous function (i.e., Hölder-continuous function, with exponent ) on , with , and that and are continuous functions on ; besides, under F(ii.1), and . Note that , . |
| 4 |
This, together with F(ii.1), implies that exists.
|
| 5 |
The rule of thumb (ROT) procedure, with the gamma or LN reference, was also considered. But, these results were not reported here, since, not surprisingly, the ROT was non-robust for the misspecification, although the computational speed of the ROT was very high. |
Table 1.
Average ISEs of the estimators using the LSCV selected smoothing parameter. The value in the parenthesis stands for the standard deviation.
Table 1.
Average ISEs of the estimators using the LSCV selected smoothing parameter. The value in the parenthesis stands for the standard deviation.
| |
PE-based BS KDE |
PE-based log KDE |
Gamma KDE |
| n |
|
|
|
|
| 200 |
12.882 |
12.440 |
12.678 |
12.128 |
12.698 |
| |
(14.863) |
(14.960) |
(15.083) |
(14.326) |
(13.601) |
| 300 |
9.767 |
9.699 |
9.809 |
9.833 |
9.939 |
| |
(11.067) |
(11.587) |
(11.414) |
(12.160) |
(10.339) |
| 500 |
6.896 |
6.703 |
7.123 |
6.886 |
7.567 |
| |
(7.873) |
(7.376) |
(8.257) |
(7.973) |
(8.373) |
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).