Preprint
Article

This version is not peer-reviewed.

On Constructive Approximation of Nonlinear Operators

Submitted:

04 December 2025

Posted:

10 December 2025

You are already at the latest version

Abstract
Suppose $K_{_Y}$ and $K_{_X}$ are the image and the preimage of a nonlinear operator $\f:K_{_Y}\rightarrow K_{_X}$. It is supposed that the cardinality of each $K_{_Y}$ and $K_{_X}$ is $N$ and $N$ is large. % large sets of observed and reference signals, respectively, each containing $N$ signals. We provide an approximation to the map $\f$ that requires a prior information only on { a few elements} $p$ from $K_{_Y}$, where $p\ll N$, but still effectively represents $\f(K_{_Y})$. It is achieved under quite non-restrictive assumptions. The device behind the proposed method is based on a special extension of the piecewise linear interpolation technique to the case of sets of stochastic elements. The proposed technique provides {a single} operator that transforms any element from the {arbitrarily large } set $K_Y$. The operator is determined in terms of pseudo-inverse matrices so that it always exists.
Keywords: 
;  ;  ;  

1. Introduction

A purpose of the proposed methodology is to provide an effective way to transform large data sets. The methodology is motivated by the problems arising in signal processing where the nonlinear operator : K Y K X is interpreted as a nonlinear system (or a nonlinear filter) transforming a set of stochastic signals K Y to the set of stochastic signals K X . Therefore, below we refer to this terminology.
The device behind the proposed method is based on a special extension of the piecewise linear interpolation technique to the case of stochastic signal sets. The device is not straightforward and requires the careful substantiation presented in Section 2.3, Section 3.4, Section 4.2 and Section 4.4 below.

1.1. Motivations

The problem under consideration is motivated by the following observations.

1.1.1. Transformation of Large Sets of Signals

Suppose we need to transform a set of signals K Y to another set of signals K X . The signals are represented by finite stochastic vectors1. A major associated difficulty and inconvenience which is common to many known filtering methodologies (see, for example, [1]–9,11,13,22,23,25]) is that they require a prior information on each reference signal to be estimated2. In particular, the filters in [22,23,25] are based on the use of either the reference signal x K X itself, as in [22,23], or its estimate, as in [25]. The Wiener filtering approach (see, for example, [1]–13,23,25]) assumes that covariance matrices formed from a reference signal, x K X , and an observed signal, y K Y , are known or can be estimated. The latter can be done, for instance, from samples of x and y . In particular, this means that the reference signal x can be measured.
In the case of processing large signal sets, such restrictions become much more inconvenient.
The major motivating question for this work is as follows. Let : K Y K X denote a filter that estimates a large set of reference signals, K X , from a large set of observed signals, K Y . Each set contains N signals. Is it possible to construct a filter that requires a prior information only on few signals, p ≪ N, from KX but performs better than the known filters based on a prior information on every reference signal from KX? We denote such a filter by (p−1).
It is shown in Section 2.3 and Section 4.4 that the positive answer is achievable under quite unrestrictive assumptions. The required features of filter ( p 1 ) are satisfied by its special structure described in Section 2.3, Section 3.1 and Section 3.4. The related conditions are also considered in those Sections.

1.1.2. Filtering Based on Idea of Piecewise Function Interpolation

The specific structure of the proposed filter follows from the extension of piecewise function interpolation [14]. This is because the technique of piecewise function interpolation [14] has significant advantages over the methods of linear and polynomial approximation used in known filtering techniques (such as, for example, those in [5,9]).
The structure of the proposed filter is presented in Section 2.3, Section 3.1 and Section 4.2 below.

1.1.3. Exploiting Pseudo-Inverse Matrices in the Filter Model

Most of the known filtering techniques, for example, those ones in [1]–3,6]–8,11,23,25], are based on exploiting inverse matrices in their mathematical models. In the cases of grossly corrupted signals or erroneous measurements those inverse matrices may not exist and, thus, those filters cannot be applied. The examples in Section 5 illustrate this case.
The filter proposed here avoids this drawback since its model is based on exploiting pseudo-inverse matrices. As a result, the proposed filter always exist. That is, it processes any kind of noisy signals. An extension of the filtering techniques to the case of implementation of the pseudo-inverse matrices is done on the basis of theory presented in [5].

1.1.4. Computational Work

Let m and n be the number of components of x K X and of y K Y , respectively, where K X and K Y each contains N signals. The known filtering techniques (e.g. see [1]–8,11,23,25]), applied to x and y , require computation of a product of an m × n matrix and an n × n matrix, as well as computation of an n × n inverse or pseudo-inverse matrix for each pair of signals x K X and y K Y . This requires O ( 2 m n 2 ) and O ( 26 n 3 ) flops, respectively [26]. Thus, for the processing of all signals in K X and K Y , the filters in [1]–8,11,23,25] require O ( 2 m n 2 N ) + O ( 26 n 3 N ) operations.
Alternatively, K X and K Y can be represented by vectors, χ and γ , each with m N and n N components, respectively. In such a case, the techniques in [1]–8,11,23,25] can be applied to χ and γ as opposed to each signals in K X and K Y . computational requirement is then O ( 2 m n 2 N 2 ) and O ( 26 n 3 N 3 ) operations, respectively [26].
In both cases, but especially when N is large, computational work associated with the approaches [1]–8,11,23,25] becomes unreasonable hard.
For the filter ( p 1 ) to be introduced below, the associated computational work is substantially less. This is because ( p 1 ) requires computation of only p pseudo-inverse matrices associated with p selected signals in K X , where p is much less than the number of signals in K X . Therefore, for processing of the signal sets, K X and K Y , ( p 1 ) requires only O ( 2 m n 2 p ) + O ( 26 n 3 p ) flops where p N . This comparison is illustrated in Section 5.

1.2. Relevant works

Some particular filtering techniques relevant to the method proposed below are as follows.

1.2.1. Generic Optimal Linear (GOL) Filter [5]

The generic optimal linear (GOL) filter in [5] is a generalization of the Wiener filter to the case when covariance matrix is not invertible and observable signal is arbitrarily noisy (i.e. when, in particular, noise is not necessarily additive and Gaussian). The GOL filter has been developed for processing an individual stochastic signal. Some ideas from [5] are used in the proof of Theorem 1 below.

1.2.2. Simplicial Canonical Piecewise Linear Filter [23]

A complex Wiener adaptive filter was developed in [23] from the two-dimensional complex-valued simplicial canonical piecewise linear filter [24]. The filter in [23] was developed for the processing of an individual stochastic signal and can be exploited when the reference signal is known and a ‘covariance-like’ matrix is invertible. The latter precludes an application to the signal types considered, for example, in Section 5: the matrices used in [23] are not invertible for the signals as those in Section 5. Similarly, the filters studied in [8,11] were developed for the processing of a single signal when the covariance matrices are invertible.
For the filter proposed here, these restrictions are removed.

1.2.3. Adaptive Piecewise Linear Filter [22]

A piecewise linear filter in [22] was proposed for a fixed image denoising (given by a matrix), corrupted by an additive Gaussian noise. That is, the method involved a non stochastic reference signal and required its knowledge. No theoretical justification for the filter was given in [22].

1.2.4. Averaging Polynomial Filter [10,12]

The averaging polynomial filter proposed in [10,12] was developed for the purpose of processing infinite signal sets. The filter was based on an argument involving the ‘averaging’ over sets of signals under consideration. This device allows one to determine a single filter for the processing of infinite signal sets. At the same time, it leads to an increase in the associated error when signals differ considerably from each other. This effect is illustrated in Section 5 below.

1.2.5. Other Relevant Filters

The technique developed in [13] is an extension of the GOL filter to the constraint problem with respect to the filter rank. It concerns data compression.
The methods in [6,7,15,16] have been developed for deterministic signals. Motivated by the results achieved in [15,16], adaptive filters were elaborated in [17]. A theoretical basis for the device proposed in [15,16] is provided in [18].
We note that the idea of piecewise linear filtering has been used in the literature in several very different conceptual frameworks, despite exploiting some very similar terms (as in [15]–24]). At the same time, a common feature of those techniques is that they were developed for the processing of a single signal, not of large signal sets as in this paper. In particular, piecewise linear filters in [19] have been obtained by arranging linear filters and thresholds in a tree structure. Piecewise linear filters discussed in [20] were developed using so-called threshold decomposition, which is a segmentation operator exploited to split a signal into a set of multilevel components. Filter design methods for piecewise linear systems proposed in [21] were based on a piecewise Lyapunov function.

1.3. Difficulties Associated with the Known Filtering Techniques

Basic difficulties associated with applying the known filtering techniques to the case under consideration (i.e. to processing of large signal sets, K X and K Y ) are that:
(i) they require an information on each reference signal (in the form of a sample, for example),
(ii) matrices used in the known filters can be not invertible (as in the simulations considered below in Section 5) and then the filter does not exist, and
(ii) the associated computation work may require a very long time. For example, in simulations (Section 5), MATLAB was out of memory for computing the GOL filter [5] when each of sets K X and K Y was represented by a long vector (this option has been discussed in Section 1.1.4 above).

1.4. Differences from the Known Filtering Techniques

The differences from the known filtering techniques discussed above are as follows.
(i) We consider a single filter that processes arbitrarily large input-output sets of stochastic signal-vectors. The known filters [1]–9,11,13,15]–25] have been developed for the processing of an individual signal-vector only. In the case of their application to arbitrarily large signal sets, they imply difficulties described in Section 1.1 and Section 1.3 above.
(ii) As a result, our piecewise linear filter model (Section 3), the statement of the problem (Section 3.3 below) and consequently, the device of its solution (Section 4 below) are different from those considered in [15]–24]. In this regard, see also Section 1.2.5.
(iii) The above naturally leads to a new structure of the filter (presented in Section 3.4 and Section 4.2 below) which is very different from the known ones.

1.5. Contribution

In general, for the processing of large data sets, the proposed filter allows us to achieve better results in comparison with the known techniques in [1]–25]. In particular, it allows us to
(i) achieve a desired accuracy in signal estimation3,
(ii) exploit a prior information only on few reference signals, p, from the set K X that contains N p signals or even infinite number of signals,
(iii) find a single filter to process any signal from the arbitrarily large signal set,
(vi) determine the filter in terms of pseudo-inverse matrices so that the filter always exists, and
(v) decrease the computational load compared to the related known techniques.

2. Some Preliminaries

2.1. Notation

The signal sets we consider are, in fact, special representations of time series.
Let ( Ω , Σ , μ ) be a probability space4, and K X and K Y be arbitrarily large sets of signals such that
K X = { x ( t , · ) L 2 ( Ω , R m ) | t T } a n d K Y = { y ( t , · ) L 2 ( Ω , R n ) | t T }
where T : = [ a , b ] R . We interpret x ( t , · ) as a reference signal and y ( t , · ) as an observable signal, an input to the filter studied below5. The variable t T R represents time6. Then, for example, the stochastic signal x(t, ·) can be interpreted as an arbitrary stationary time series.
Let { t k } 1 p T be a sequence of fixed time-points such that
a = t 1 < < t p = b .
Because of the partition (1), the sets K Y and K X are divided in `smaller’ subsets K X , 1 , , K X , p 1 and K Y , 1 , , K Y , p 1 , respectively, so that, for each j = 1 , , p ,
K X , j = { x ( t , · ) | t j t t j + 1 } a n d K Y , j = { y ( t , · ) | t j t t j + 1 } .
Therefore, K Y and K X can now be represented as
K X = j = 1 p 1 K X , j a n d K Y = j = 1 p 1 K Y , j .

2.2. Brief Description of the Problem

Given two arbitrarily large sets of stochastic signals, K Y and K X , find a single filter : K Y K X that estimates the signal x K X with a controlled, associated error. Note that in our formulation the set K Y can be finite or infinite.

2.3. Brief description of the method

The solution of the above problem is based on the representation of the proposed filter in the form of a sum with p 1 terms 1 , , p 1 where each term, j, is interpreted as a particular sub-filter (see (4) and (5) below). Such a filter is denoted by ( p 1 ) : K Y K X .
The sub-filter j transforms signals that belong to ‘piece’ K Y , j of set K Y to signals in ‘piece’ K X , j of K X , i.e. j : K Y , j K X , j . Each sub-filter j depends on two parameters, α j and B j .
The prime idea is to determine j (i.e. α j and B j ) separately, for each j = 1 , , p 1 . The required α j and B j follow from the solutions of the equation (11) and an associated minimization problem (11) (see Section 3.4 and Section 4.2 below). This procedure adjusts j so that the error associated with the estimation of x ( t , · ) K X , j is minimal.
A motivation for such a structure of the filter ( p 1 ) is as follows. The method of determining α j and B j provides an estimate j [ y ( t , · ) ] that interpolates x ( t , · ) K X , j at t = t j and t = t j + 1 . In other words, the filter is flexible to variations in the sets of observed and reference signals K Y and K X , respectively. Due to this way of determining j, it is natural to expect that the processing of a ‘smaller’ signal set, K Y , j , may lead to a smaller associated error than that for the processing of the whole set K Y by a filter which is not specifically adjusted to each particular piece K Y , j .
As a result, ( p 1 ) [ y ( t , · ) ] represents a special piecewise interpolation procedure and, thus, should be attributed with the associated advantages such as, for example, the high accuracy of estimation.
In Section 4.4, this observation is confirmed. In Section 2 and Section 5, it is also shown that the proposed technique allows us to avoid the difficulties discussed in Section 1.3 above.

3. Description of the Problem

3.1. Piecewise Linear Filter Model

Let ( p 1 ) : K Y K X be a filter such that, for each t T ,
( p 1 ) [ y ( t , · ) ] = j = 1 p 1 δ j j [ y ( t , · ) ] ,
where
j [ y ( t , · ) ] = α j + B j [ y ( t , · ) ] a n d δ j = 1 , i f t j t t j + 1 , 0 , o t h e r w i s e .
Here, j is a sub-filter defined for t j t t j + 1 . In (5), α j = [ α j ( 1 ) , , α j ( m ) ] T R m and B j : L 2 ( Ω , R n ) L 2 ( Ω , R m ) is a linear operator given by a matrix B j R m × n , so that
[ B j ( y ) ] ( t , ω ) = B j [ y ( t , ω ) ] .
Thus, j is defined by an operator F j : R n R m such that
F j [ y ( t , ω ) ] = α j + B j [ y ( t , ω ) ] .
Filter ( p 1 ) defined by (4)-(6) is called the piecewise filter7.

3.2. Assumptions

In the known approaches related to filtering of stochastic signals (e.g. see [1]–13,23,25]), it is assumed that covariance matrices formed from the reference signal and observed signal are known or can be estimated.
The assumption used here is similar. The covariance matrices that are assumed to be known or can be estimated, are formed from selected signal pairs { x ( t j , · ) , y ( t j , · ) } with j = 1 , , p and p to be a small number8, p N , where N is the number of signals in K X or K Y .

3.3. The Problem

In (4)-(6), parameters of the filter ( p 1 ) , i.e. vector α j and matrix B j , for j = 1 , , p 1 , are unknown. Therefore, under the assumptions described in Section 3.2, the problem is to determine α j and B j , for j = 1 , , p 1 . The related problem is to estimate an error associated with the filter ( p 1 ) .
Solutions to the both problems are given in Section 4.2 and Section 4.4, respectively. In particular, in the following Section 3.4, interpolation conditions (8) and (11) are introduced that lead to a determination of α j and B j .

3.4. Interpolation Conditions

Let us denote
x ( t j , · ) Ω 2 = Ω x ( t j , ω ) 2 2 d μ ( ω )
where x ( t j , ω ) 2 is the Euclidean norm of x ( t j , ω ) R m .
For t = t 1 , let x ^ ( t 1 , · ) be an estimate of x ( t 1 , · ) determined by known methods [1]–13,23,25]. This is the initial condition of the proposed technique.
For j = 1 , , p 1 , each sub-filter F j in (5)-(6) is defined so that α j and B j satisfy the conditions as follows.
Sub-filter 1: For j = 1 , α 1 and B 1 solve
x ^ ( t 1 , · ) = α 1 + B 1 [ y ( t 1 , · ) ] and min B 1 [ x ( t 2 , · ) α 1 ] B 1 [ y ( t 2 , · ) ] Ω 2 ,
respectively. Then an estimate of x ( t , · ) , x ^ ( t , · ) , for t [ t 1 , t 2 ] , is determined as
x ^ ( t , · ) = 1 [ y ( t , · ) ] = α 1 + B 1 [ y ( t , · ) ] = x ^ ( t 1 , · ) + B 1 [ y ( t , · ) y ( t 1 , · ) ]
where α 1 and B 1 satisfy (8). In particular, α 1 = x ^ ( t 1 , · ) B 1 [ y ( t 1 , · ) ] and
x ^ ( t 2 , · ) = 1 [ y ( t 2 , · ) ] .
Extending this procedure up to j = k 1 , where k = 3 , , p , we set the following. Let x ^ ( t k 1 , · ) be an estimate of x ( t k 1 , · ) defined by the preceding steps as
x ^ ( t k 1 , · ) = k 2 [ y ( t k 1 , · ) ] .
Then sub-filter k 1 is defined as follows.
Sub-filter k 1 : For j = k 1 , α k 1 and B k 1 solve
x ^ ( t k 1 , · ) = α k 1 + B k 1 [ y ( t k 1 , · ) ] and min B k 1 [ x ( t k , · ) α k 1 ] B k 1 [ y ( t k , · ) ] Ω 2 ,
respectively. Then an estimate of x ( t , · ) , x ^ ( t , · ) , for t [ t k 1 , t k ] , is determined as
x ^ ( t , · ) = k 1 [ y ( t , · ) ] = α k 1 + B k 1 [ y ( t , · ) ] = x ^ ( t k 1 , · ) + B 1 [ y ( t , · ) y ( t k 1 , · ) ] .
The conditions (8) and (11) are motivated by the device of piecewise function interpolation and associated advantages [14].
Filter ( p 1 ) of the form (4)-(5) with α j and B j satisfying (8) and (11) is called the piecewise linear interpolation filter. The pair of signals { x ( t k , · ) , y ( t k , · ) } associated with time t k defined by (1) is called the interpolation pair.

4. Main Results

4.1. General Device

In accordance withe the scheme presented in Section 3.1 and Section 3.4 above, an estimate of the reference signal x ( t , · ) , for any t T = [ a , b ] , by the piecewise linear interpolation filter ( p 1 ) , is given by
x ^ ( t , · ) = ( p 1 ) [ y ( t , · ) ] = j = 1 p 1 δ j j [ y ( t , · ) ] ,
where, for each j = 1 , , p 1 , the sub-filter j is given by (5), and is defined from the interpolation conditions (8) and (11).
Below, we show how to determine j to satisfy the conditions (8) and (11).

4.2. Determination of Piecewise Linear Interpolation Filter

Let us denote
z ( t j , t j + 1 , · ) = x ( t j + 1 , · ) x ^ ( t j , · ) a n d w ( t j , t j + 1 , · ) = y ( t j + 1 , · ) y ( t j , · ) .
We need to represent z ( t j , t j + 1 , · ) and w ( t j , t j + 1 , · ) in terms of their components as follows:
z ( t j , t j + 1 , · ) = [ z ( 1 ) ( t j , t j + 1 , · ) , , z ( m ) ( t j , t j + 1 , · ) ] T and w ( t j , t j + 1 , · ) = [ w ( 1 ) ( t j , t j + 1 , · ) , , w ( n ) ( t j , t j + 1 , · ) ] T ,
where z ( j ) ( t j , t j + 1 , · ) L 2 ( Ω , R ) and w ( i ) ( t j , t j + 1 , · ) L 2 ( Ω , R ) are stochastic variables, for all j = 1 , , m .
Then we can introduce the covariance matrix
E z j w j = z ( i ) ( t j , t j + 1 , · ) , w ( k ) ( t j , t j + 1 , · ) i , k = 1 m , n ,
where z ( i ) ( t j , t j + 1 , · ) , w ( k ) ( t j , t j + 1 , · ) = Ω z ( i ) ( t j , t j + 1 , ω ) w ( k ) ( t j , t j + 1 , ω ) d μ ( ω ) .
Below, M is the Moor-Penrose generalized inverse of a matrix M.
Now, we are in a position to establish the main results.
Theorem 1. 
Let
K X = { x ( t , · ) L 2 ( Ω , R m ) | t T = [ a , b ] } a n d K Y = { y ( t , · ) L 2 ( Ω , R n ) | t T = [ a , b ] }
be sets of reference signals and observed signals, respectively. Let t j [ a , b ] , for j = 1 , , p , be such that
a = t 1 < < t p = b .
For t = t 1 , let x ^ ( t 1 , · ) be a known estimate of x ( t 1 , · ) 9. Then, for any t [ a , b ] , the proposed piecewise linear interpolation filter ( p 1 ) : L 2 ( Ω , R n ) L 2 ( Ω , R m ) transforming any signal y ( t , · ) L 2 ( Ω , R m ) to an estimate of x ( t , · ) , x ^ ( t , · ) , is given by
x ^ ( t , · ) = ( p 1 ) [ y ( t , · ) ] = j = 1 p 1 δ j F j [ y ( t , · ) ]
where
j [ y ( t , · ) ] = x ^ ( t j , · ) + B j [ y ( t , · ) y ( t j , · ) ] ,
x ^ ( t j , · ) = j 1 [ y ( t j , · ) ] ( f o r   j = 2 , , p 1 ) ,
B j = E z j w j E w j w j + M B j [ I n E w j w j E w j w j ] ,
and where In is n × n identity matrix and MBj is an m × n arbitrary matrix.
Proof: The proof of Theorem 1 is given in Section A.
It is worthwhile to observe that, due to an arbitrary matrix M B j in (19), the filter ( p 1 ) is not unique. In particular, M B j can be chosen as the zero matrix O similarly to the generic optimal linear [5] (which is also not unique by the same reason).

4.3. Numerical Realization of Filter ( p 1 ) and Associated Algorithm

4.3.1. Numerical Realization

In practice, the set T = [ a , b ] (see Section 2.1) is represented by a finite set { τ k } k = 1 N , i.e. [ a , b ] = [ τ 1 , τ 2 , , τ N ] where a τ 1 < τ 2 < < τ N b .
For k = 1 , , N , the estimate of x ( τ k , · ) , x ^ ( τ k , · ) , and observed signal y ( τ k , · ) are represented by m × q and n × q matrices
X ^ ( k ) = [ x ^ ( τ k , ω 1 ) , , x ^ ( τ k , ω q ) ] a n d Y ( k ) = [ y ( τ k , ω 1 ) , , y ( τ k , ω q ) ] .
The sequence of fixed time-points { t k } 1 p [ a , b ] introduced in (1) is such that
τ 1 = t 1 < < t p = τ N ,
where t 1 = τ n 0 , t 2 = τ n 0 + n 1 , , t p = τ n 0 + n 1 + n p 1 , and where n 0 = 1 and n 1 , , n p 1 are positive integers such that N = n 0 + n 1 + + n p 1 .
For j = 1 , , p , vectors x ^ ( t j , · ) and y ( t j , · ) associated with t j in (21) are represented, respectively, by
X ^ j ( k ) = [ x ^ ( t j , ω 1 ) , , x ^ ( t j , ω N ) ] and Y j = [ y ( t j , ω 1 ) , , y ( t j , ω N ) ] .

4.3.2. Algorithm

As it has been mentioned in Section 3.4, it is supposed that, for t = t 1 , an estimate of X 1 , X ^ 1 , is known and can be determined by the known methods. This is the initial condition of the proposed technique.
On the basis of the results obtained in Section 3.4 and Section 4.2, the performance algorithm of the proposed filter consists of the following steps. For j = 1 , p , we write N j = n 0 + n 1 + + n j 1 .
Initial parameters: Y ( 1 ) , , Y ( N ) , { t j } j = 1 p (see (21)), { E z j w j } j = 1 p , { E w j w j } j = 1 p (see (14) and (15)), X ^ 1 , n 0 = 1 , N 0 = 0 and M B j = O , for j = 1 , , p 1 .
(Possible ways to get estimates of E z j w j and E w j w j are discussed below in Section 4.5.)
Final parameters: X ^ ( 2 ) , X ^ ( 3 ) ,…, X ^ ( N ) .
Algorithm:
• for j = 1 to p do
    begin
B j = E z j w j E w j w j ;
• for k = N j 1 + 1 to N j do
     begin
X ^ ( k ) = X ^ j + B j ( Y ( k ) Y j ) ;
X ^ j + 1 = X ^ ( N j ) ;
      end
    end

4.4. Error Analysis

It is natural to expect that the error associated with the piecewise interpolating filter ( p 1 ) decreases when max j = 1 , , p 1 Δ t j decreases. Below, in Theorem 3, we justify that this observation is true. To this end, first, in the following Theorem 2, we establish an estimate of the error associated with the filter F.
Let us introduce the norm by
x ( t , · ) T , Ω 2 = 1 b a T x ( t , · ) Ω 2 d t .
We also denote x ( t , ω ) T , Ω 2 = x ( t , · ) T , Ω 2 .
Let us suppose that x ( · , ω ) and y ( · , ω ) are Lipschitz continuous signals, i.e. that there exist real non-negative constants λ j and γ j , with j = 1 , , p , such that, for t [ t j , t j + 1 ] ,
x ( t , ω ) x ( t j , ω ) T , Ω 2 λ j Δ t j a n d y ( t , ω ) y ( t j + 1 , ω ) T , Ω 2 γ j Δ t j
where Δ t j = | t j + 1 t j | .
Theorem 2. 
Under the conditions (23) the error associated with the piecewise interpolation filter, x ( t , ω ) F ( p 1 ) [ y ( t , ω ) ] T , Ω 2 , is estimated as follows:
x ( t , ω ) F ( p 1 ) [ y ( t , ω ) ] T , Ω 2 max j = 1 , , p 1 ( λ j + γ j B j 2 ) Δ t j + E z j z j 1 / 2 2 E z j w j ( E w j w j 1 / 2 ) 2 .
Proof: The proof of Theorem 2 is given in Section A. □
Further, to show that the error of the reference signal estimate tends to the zero, we need to assume that, for t [ t 1 , t 2 ] , the known estimate x ^ ( t 1 , ω ) differs from x ( t , ω ) for the value of the order Δ t 1 , i.e. that, for some constant c 1 0 ,
x ( t , ω ) x ^ ( t 1 , ω ) Ω 2 c 1 Δ t 1 , f o r t [ t 1 , t 2 ] .
Theorem 3. 
Let the conditions (23) and (25) be true. Then the error associated with the piecewise interpolating filter F, x ( t , ω ) F ( p 1 ) [ y ( t , ω ) ] T , Ω 2 , decreases in the following sense:
x ( t , ω ) F ( p 1 ) [ y ( t , ω ) ] T , Ω 2 0 a s max j = 1 , , p 1 Δ t j 0 a n d p .
Proof: The proof of Theorem 3 is given in Section A. □
Remark 1. 
We would like to emphasize that the statement of Theorem 3 is fulfilled only under assumptions (23) and (25). At the same time, the assumptions (23) and (25) are not restrictive from a practical point of view. The condition (23) is true for Lipschitz continuous signals x and y , i.e. for very wide class of signals. The condition (25) is achieved by a choosing an appropriate known method (e.g. see [1]–13,23,25]) to find the estimate x ^ ( t 1 , ω ) used in the proposed filter ( p 1 ) (see (8) and Theorem 1).

4.5. Some Remarks Related to the Assumptions of the Method

As it has been mentioned in Section 3.2, for j = 1 , , p , matrices E z j w j and E w j w j in (19) are assumed to be known or can be estimated. Here, p is a chosen number of selected interpolation signal pairs (see Section 3.4). We note that normally p is much smaller than the number of input-output signals x ( t , · ) and y ( t , · ) . Therefore, to estimate any signal x ( t , · ) from an arbitrarily large set K X , only a small number, p, of matrices E z j w j and E w j w j should be estimated (or be known). This issue has also been discussed in Section 1.1.1 and Section 1.1.4.
By the proposed method, x ( t , · ) is estimated for t [ t j , t j + 1 ] . While E w j w j in (19) can be directly estimated from observed signals y ( t j + 1 , · ) and y ( t j , · ) , an estimate of matrix E z j w j depends on the reference signal x ( t j + 1 , · ) (see (14) and (15)) which is unknown (because the estimate is considered for t [ t j , t j + 1 ] ).
Some possible approaches to an estimation of matrix E z j w j could be as follows.
1. In the general case, when x ( t , · ) and y ( t , · ) are arbitrary signals as discussed in Section 2.1 above, matrix E z j w j can be estimated as proposed, for example, in [27], from samples of z j and w j .
2. In the case of incomplete observations, the method proposed in [28,29] can be used.
3. Let E z ^ j w j be a matrix obtained from matrix E z j w j where the term x ( t j + 1 , · ) is replaced by x ^ ( t , · ) with t [ t j 1 , t j ] . Since x ^ ( t , · ) with t [ t j 1 , t j ] is known, matrix E z ^ j w j can be considered as an estimate of E z j w j .
4. In the important case of an additive noise, E z j w j can be represented in the explicit form. Indeed, if
y ( t , · ) = x ( t , · ) + ξ ( t , · )
where ξ ( t , · ) L 2 ( Ω , R m ) is a random noise, then z ( t j , t j + 1 , · ) = y ( t j + 1 , · ) ξ ( t j + 1 , · ) x ^ ( t j , · ) and matrix E z j w j can be represented as follows:
E z j w j = E ( y j + 1 ξ j + 1 ) ( y j + 1 y j ) E x ^ j ( y j + 1 y j )
We note that the RHS of (27) depends only on observed signals y ( t j , · ) , y ( t j + 1 , · ) , estimated signal x ^ ( t j , · ) , and noise ξ ( t j + 1 , · ) , not on the reference signal x ( t j + 1 , · ) . In particular, in (27), the term E ξ j + 1 ( y j + 1 y j ) can be estimated as ± ( E [ ξ j + 1 2 ] ) 1 / 2 ( E [ ( y j + 1 y j ) 2 ] ) 1 / 2 where E [ ξ j + 1 2 ] = Ω [ ξ ( t j + 1 , ω ) ] 2 d μ ( ω ) . It is motivated by the Holder’s inequality for integrals. The second term in (27), E x ^ j ( y j + 1 y j ) , can be estimated from the samples of x ^ ( t j + 1 , · ) and y ( t j + 1 , · ) y ( t j , · ) .
We also note that the first term in the RHS of (27), E ( y j + 1 ξ j + 1 ) ( y j + 1 y j ) , is similar to the related covariance matrix in the Wiener filtering approach [5].
5. Other known ways to estimate E ξ j + 1 ( y j + 1 y j ) can be found in [5], Section 5.3.
In general, an estimation of covariance matrices is a special research topic which is not a subject of this paper. The relevant references can be found, for example, in [5,29].

5. Simulations

5.1. General Consideration

In these simulations, in accordance with Section 4.3.1, signal sets K X and K Y (see Section 2.1) are given by
K X = { x ( τ 1 , · ) , x ( τ 2 , · ) , , x ( τ N , · ) } a n d K Y = { y ( τ 1 , · ) , y ( τ 2 , · , , y ( τ N , · ) } ,
where, for k = 1 , , N , x ( τ k , · ) L 2 ( Ω , R m ) and y ( τ k , · ) L 2 ( Ω , R n ) . In many practical problems (arising, for example, in a DNA analysis the number N is quite large, for instance, N = O ( 10 4 ) .
We set N = 141 and m = n = 116 . Thus, in these simulations, the interval T = [ a , b ] (see Section 2.1 and Section 4.3.1) is modelled as 141 points τ k with k = 1 , , 141 so that [ a , b ] = [ τ 1 , τ 2 , , τ 141 ] .
The sequence of fixed time-points { t k } 1 p T in (1) is now such that
τ 1 = t 1 < < t p = τ 141 .
Below, in Examples 1-12, four particular choices of the specific interpolation signal pairs { x ( t j , · ) , y ( t j , · ) } 1 p (introduced in Section 3.4) are considered, for p = 5 , 8 , 15 and 28. Points t 1 , , t p are as follows.
For p = 5 , 8 , 15 , if j = 1 , , p , then t j = t j ( p ) = τ 1 + ( j 1 ) Δ p , respectively, where Δ 5 = 35 ,   Δ 8 = 20 and Δ 15 = 10 .
For p = 28 , if j = 1 , , p 1 , then t j = t j ( p ) = τ 1 + ( j 1 ) Δ 28 , and if j = p , then t 28 = t 28 ( 28 ) = t 27 + 6 = 141 , where Δ 28 = 5 .
Signals x ( τ k , · ) and y ( τ k , · ) have been simulated as digital images represented by 116 × 256 matrices
X ( k ) = [ x ( τ k , ω 1 ) , , x ( τ k , ω 256 ) ] and Y ( k ) = [ y ( τ k , ω 1 ) , , y ( τ k , ω 256 ) ] ,
respectively, for k = 1 , , 141 , so that X ( k ) represents an image that should be estimated from an observed image Y ( k ) . A column of matrices X ( k ) and Y ( k ) , x ( τ k , ω i ) R 116 and y ( τ k , ω i ) R 116 , for i = 1 , , 256 , represents a realization of signals x ( τ k , · ) and y ( τ k , · ) , respectively.
Note that X ( 1 ) , , X ( 141 ) did not used in the piecewise linear filter F ( p 1 ) below since they are not supposed to be known. They are represented here for illustration purposes only. In particular, X ( 1 ) , , X ( 141 ) are used to compare their estimates by different filters.
Observed noisy signals Y ( 1 ) , , Y ( 141 ) have been simulated in different forms presented by (40), (49), (50) and (51) in the Examples 1-12 below. We note that the considered observed signals are grossly corrupted.
To estimate the signals X ( 1 ) , ..., X ( 141 ) from the observed signals Y ( 1 ) , ..., Y ( 141 ) , the proposed piecewise linear filter F ( p 1 ) , the generic optimal linear (GOL) filters [5] and the averaging polynomial filter [12] have been used.
The filters proposed in [12,13,22,23] have not been applied here by the reasons discussed in Section 1. In particular, the filter in [23] cannot be applied to signals represented by Y ( 1 ) , ..., Y ( 141 ) in the form (40), (49), (50) and (51) below because the associated inverse matrices used in [23] do not exist.
For signals under consideration (given by matrices X ( k ) and Y ( k ) with k = 1 , , 141 ), the filter F ( p 1 ) , the generic optimal linear (GOL) filters [5] and the averaging polynomial filter [10,12] are represented as follows.
(i) Piecewise linear filter F ( p 1 ) .For j = 1 , , p , { X j , Y j } designates an interpolation pair defined similarly to that in Section 3.4. Each X j and Y j is associated with t j in (25) so that
X j = [ x ( t j , ω 1 ) , , x ( t j , ω 256 ) ] a n d Y j = [ y ( t j , ω 1 ) , , y ( t j , ω 256 ) ] .
The estimate X ^ ( k ) of X ( k ) by the filter F ( p 1 ) is given by
X ^ ( k ) = F ( p 1 ) [ Y ( k ) ] ,
where, by (16)-(19) in Section 4.2,
F ( p 1 ) [ Y ( k ) ] = j = 1 p 1 δ j F j [ Y ( k ) ] , δ j = 1 , i f t j τ k t j + 1 , 0 , o t h e r w i s e ,
F j ( p 1 ) [ Y ( k ) ] = X ^ j + B j [ Y ( k ) Y j ] ,
X ^ j = F j 1 [ Y j ] , X ^ 1 i s g i v e n ,
B j = E Z j W j ( E W j W j ) ,
and where E Z j W j and E W j W j are estimates of matrices E z j w j and E w j w j in (19), respectively. In particular, E W j W j can be represented in the form
E W j W j = W j W j T , w h e r e W j = Y j + 1 Y j .
Further, matrix E Z j W j depends on Z j = X j + 1 X ^ j where X j + 1 is unknown. Therefore a determination of E Z j W j is reduced, in fact, to finding an estimate of X j + 1 . Since it is customary to find E Z j W j in terms of signal samples [5], E Z j W j has been presented as
E Z j W j = Z ˜ j W j T , w h e r e Z ˜ j = X ˜ j + 1 X ^ j
and X ˜ j + 1 has been constructed from a sample of X j + 1 as follows. The sample of X j + 1 is a 116 × 128 matrix presented by odd columns of X j + 1 . Then an estimate of X j + 1 is chosen as a 116 × 256 matrix X ˜ j + 1 where each odd column is a related odd column of X j + 1 , and each even column is an average of two adjacent columns. The last column in X ˜ j + 1 is the same as its preceding column.
This way of estimating E Z j W j was chosen for illustration purposes only. Other related methods have been considered in Section 4.5.
The errors associated with the filter F ( p 1 ) are given by
ε k , F ( p 1 ) = X ( k ) F ( p 1 ) [ Y ( k ) ] F 2 , f o r k = 1 , , 141 .
(ii) Generic optimal linear (GOL) filters [5]. To each signal Y ( k ) , an individual GOL filter W k has also been applied, so that W k estimates X ( k ) from Y ( k ) in the form
W k Y ( k ) = E X ( k ) Y ( k ) E Y ( k ) Y ( k ) Y ( k ) ,
for each k = 1 , , 141 . Thus, the GOL filter W k requires an estimate of 141 matrices E X ( k ) Y ( k ) , for each k = 1 , , 141 .
Similarly to matrix E Z j W j in the filter F ( p 1 ) above, the matrix E X ( k ) Y ( k ) has been estimated from samples of each X ( k ) , X ˜ ( k ) , for each k = 1 , , 141 .
One of the advantages of the proposed filter F ( p 1 ) is that F ( p 1 ) requires a smaller number, p, of samples of X j , X ˜ j , to be known (where j = 1 , , p ).
The errors associated with filters W k are given by
ϵ k , w = X ( k ) W k Y ( k ) F 2 .
(iii) Averaging polynomial filters [10,12]. By the methodology in [10], the averaging polynomial filter W is based on the use of the estimates of the covariance matrices, E X Y and E Y Y , in the form
E X Y = 1 141 k = 1 141 X ˜ ( k ) ( Y ( k ) ) T and E Y Y = 1 141 k = 1 141 Y ( k ) ( Y ( k ) ) T .
Then, for each, k = 1 , , 141 , the estimate of X ( k ) is given by
W Y ( k ) = E X Y E Y Y Y ( k ) .
The errors associated with the filter W are given by
ε k W = X ( k ) W Y ( k ) F 2 , f o r k = 1 , , 141 .

5.2. Simulations with Signals Modelled from Images ‘Plant’: Application of Piecewise Interpolation Filter and GOL Filters

Here, results of simulations for reference signals represented by matrices X ( 1 ) ,   , X ( 141 ) (see (29) above) formed from images ‘plant’10 are considered. Typical selected images X ( k ) are shown in Figure 9.
Observed noisy images Y ( 1 ) , , Y ( 141 ) have been simulated in the form
Y ( k ) = X ( k ) randn ( k ) rand ( k ) ,
for each k = 1 , , 141 . Here, • means the Hadamard product, and randn ( k ) and rand ( k ) are 116 × 256 matrices with random entries. The entries of randn ( k ) are normally distributed with mean zero, variance one and standard deviation one. The entries of rand ( k ) are uniformly distributed in the interval ( 0 , 1 ) . A typical example of such images is given in Figure 10 (a).
To demonstrate the effectiveness of the proposed filter F ( p 1 ) , sub-filters F j ( p 1 ) and associated interpolation signal pairs { X j , Y j } j = 1 p have been chosen in four different ways as follows.
Example 1. First, for p = 5 , the interpolation signal pairs are
{ X 1 , Y 1 } = { X ( 1 ) , Y ( 1 ) } , { X 2 , Y 2 } = { X ( 35 ) , Y ( 35 ) } , { X 3 , Y 3 } = { X ( 70 ) , Y ( 70 ) } ,
{ X 4 , Y 4 } = { X ( 105 ) , Y ( 105 ) } , { X 5 , Y 5 } = { X ( 141 ) , Y ( 141 ) } .
The error values { ε k , F ( 4 ) } 1 141 associated with filter F ( 4 ) are evaluated by (37). The graph of { ε k , F ( 4 ) } 1 141 is presented in Figure 11 (a).
Example 2. For p = 8 , the interpolation signal pairs are
{ X 1 , Y 1 } = { X ( 1 ) , Y ( 1 ) } , { X j , Y j } = { X ( 20 ( j 1 ) ) , Y ( 20 ( j 1 ) ) } , f o r j = 2 , , 7 ;
a n d { X 8 , Y 8 } = { X ( 141 ) , Y ( 141 ) } .
The error magnitudes { ε k , F ( 7 ) } 1 141 associated with the piecewise interpolation filter F ( 7 ) constructed by (31)-(36) with the interpolation signal pairs given by (43)-(44) are diagrammatically shown in Figure 11 (b).
It follows from Figure 11 (b) that the errors associated with filter F ( 7 ) is less than those of filter F ( 4 ) . This is a confirmation of Theorem 3.
Example 3. Further, for p = 15 , the interpolation pairs are
{ X 1 , Y 1 } = { X ( 1 ) , Y ( 1 ) } , { X j , Y j } = { X ( 10 ( j 1 ) ) , Y ( 10 ( j 1 ) ) } f o r j = 2 , , 14 ;
and { X 15 , Y 15 } = { X ( 141 ) , Y ( 141 ) } .
In Figure 11 (c), the errors { ε k , F ( 15 ) } 1 141 associated with the piecewise interpolation filter F ( 15 ) are presented. The Figure 11 (c) demonstrates a further confirmation of Theorem 3: the errors associated with the piecewise interpolation filter diminishes as p increases.
Example 4. Finally, the number of interpolation signal pairs { X j , Y j } j = 1 p is p = 29 so that
{ X 1 , Y 1 } = { X ( 1 ) , Y ( 1 ) } , { X j , Y j } = { X ( 5 ( j 1 ) ) , Y ( 5 ( j 1 ) ) } f o r j = 2 , , 28 ;
and { X 29 , Y 29 } = { X ( 141 ) , Y ( 141 ) } .
In this case, when p is grater than in the previous Examples 1-3, the errors { ε k , F ( 29 ) } 1 141 associated with the piecewise interpolation filter F ( 29 ) are smaller than those associated with filters F ( 4 ) , F ( 8 ) and F ( 15 ) - see Figure 11 (d).
The diagrams of errors associated with the GOL filters [5] are also presented in Figure 11. It follows from Figure 11 that proposed filters F ( 4 ) , F ( 8 ) , F ( 15 ) and F ( 29 ) provide the better accuracy then that of the GOL filters.
At the same time, the filter F ( p 1 ) is easer to implement since it requires less initial information compared to GOL filters, as it has been discussed in Section 1.1.1 and Section 1.1.4.
Results of the application of the averaging polynomial filter [10,12] are discussed in Section 5.4 below.

5.3. Simulations with Signals Modelled from Images ‘Boat’: Application of Piecewise Interpolation Filter and GOL Filters

In this section, results of the simulations for a different type of signals than those considered in Section 5.2 above are presented. Here, the reference signals X ( 1 ) , , X ( 141 ) are formed from images ‘boat’11.
Observed noisy signals Y ( 1 ) , , Y ( 141 ) have been simulated in the form
Y ( k ) = X ( k ) randn ( k ) ,
for each k = 1 , , 141 . The noise term is different from that in (40).
Typical selected images X ( k ) and Y ( k ) are shown in Figure 12 and Figure 13, respectively.
As in Section 5.2, the piecewise interpolation filter F ( p 1 ) is constructed by (31)-(36). In Examples 5-8 below, the number p 1 of sub-filters F j ( p 1 ) and associated interpolation signal pairs { X j , Y j } j = 1 p have been chosen in four different ways.
Example 5. First, similar to Example 1, the number of interpolation signal pairs { X j , Y j } j = 1 p has been chosen as p = 5 , and X j and Y j have been presented as in (41)-(42).
The error values { ε k , F ( 4 ) } 1 141 associated with the piecewise interpolation filter F ( 4 ) applied to these data are presented in Figure 14 (a).
Example 6. For the grater number of interpolation signal pairs than that in Example 5, p = 8 , and for X j and Y j ( j = 1 , , 8 ) chosen as in (43)-(44), the error magnitudes { ε k , F ( 7 ) } 1 141 associated with the piecewise interpolation filter F ( 7 ) are diagrammatically shown in Figure 14 (b). A comparison between Figure 14 (a) and (b) demonstrates that the increase in p implies the decrease in the errors associated with the filter F ( p 1 ) .
Example 7. For p = 15 , and for X j and Y j ( j = 1 , , 15 ) chosen as in (45)-(46), the errors { ε k , F ( 14 ) } 1 141 associated with the piecewise interpolation filter F ( 14 ) are further less than those for filters F ( 4 ) and F ( 7 ) . See Figure 14 (c) in this regard.
Example 8. The further increase in p to p = 29 , confirms this tendency. The piecewise interpolation filter F ( 28 ) with X j and Y j ( j = 1 , , 29 ) chosen similar to (47)-(48) produces the associated errors { ε k , F ( 28 ) } 1 141 represented in Figure 14 (d). They are, clearly, less than the errors associated with filters F ( 4 ) , F ( 7 ) and F ( 15 ) .
The errors associated with the GOL filters are also presented in Figure 14 (a)-(d). The figures clearly demonstrate the advantage of the piecewise interpolation filter F ( p 1 ) .
Results of the application of the averaging polynomial filter [12] are discussed in Section 5.4 below.

5.4. Results of Simulations for Averaging Polynomial Filter [10,12]

To further illustrate the effectiveness of the proposed piecewise interpolation filter, in this Section, results of simulations for the averaging polynomial filter [10,12] are presented. The filter has been applied to two different types of data considered in Section 5.2 and Section 5.3.
Example 9. The filter [10,12] applied to signals considered in Section 5.2 gives the associated errors { ϵ k W } k = 1 141 (see (39)) represented in Figure 15 (a). For a comparison, the errors associated with the piecewise interpolation filter F ( 28 ) and the GOL filters [5] are also given in Figure 15 (a).
A typical example of the estimated signal by the averaging polynomial filter [10,12] is presented in Figure 10 (d) above.
Example 10. The averaging polynomial filter [12] applied to signals considered in Section 5.3 produces the associated errors { ϵ k W } k = 1 141 shown in Figure 15 (b). The errors associated with the piecewise interpolation filter F ( 28 ) are much smaller and they are not discerned in Figure 15 (b).
Together with Figure 10, Figure 11, Figure 13 and Figure 14, Figure 15 (a) and (b) illustrate the advantage of the piecewise interpolation filter.
Figure 1. Examples of selected signals to be estimated from observed data.
Figure 1. Examples of selected signals to be estimated from observed data.
Preprints 188126 g001
Figure 2. Examples of the observed signal and the estimates obtained by different filters.
Figure 2. Examples of the observed signal and the estimates obtained by different filters.
Preprints 188126 g002
Figure 3. Illustration of the errors associated with the piecewise interpolation filters F ( p 1 ) and the GOL filters [5] applied to signals described in Examples 1–4.
Figure 3. Illustration of the errors associated with the piecewise interpolation filters F ( p 1 ) and the GOL filters [5] applied to signals described in Examples 1–4.
Preprints 188126 g003
Figure 4. Examples of selected signals to be estimated from observed data considered in Example 5-9.
Figure 4. Examples of selected signals to be estimated from observed data considered in Example 5-9.
Preprints 188126 g004
Figure 5. Examples of the observed signal and the estimates obtained by different filters.
Figure 5. Examples of the observed signal and the estimates obtained by different filters.
Preprints 188126 g005

5.5. Further Simulations with Different Type of Noise

In Examples 11 and 12 below, a different type of noise is considered. Unlike the multiplicative noise in (40) and (49), here, the noise is additive.
Example 11. First, the piecewise interpolation filter F ( 28 ) , the GOL filters [5] and the averaging polynomial filter [12] have been applied to the observed signals given by
Y ( k ) = X ( k ) + 900 × randn ( k ) , for   k = 1 , , 141 .
where X ( k ) is as in Section 5.2, i.e. X ( k ) is formed from the images ‘plant’. In Figure 8 (a), the diagrams of the errors associated with filter F ( 28 ) and the GOL filters [5] are given. The errors associated with the averaging polynomial filter [12], { ϵ k W } k = 1 141 , are much grater (of order O ( 10 9 ) ) and they are not presented in Figure 8 (a).
Figure 6. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) of order p and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 5–8.
Figure 6. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) of order p and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 5–8.
Preprints 188126 g006
Figure 7. Illustration of errors associated with the averaging polynomial filters [10,12] in Examples 9 and 10.
Figure 7. Illustration of errors associated with the averaging polynomial filters [10,12] in Examples 9 and 10.
Preprints 188126 g007
Example 12. In this example, the reference signals X ( 1 ) , , X ( 141 ) are as those in Section 5.3, i.e. they are formed from the image ‘boat’. The observed signals are given by
Y ( k ) = X ( k ) + 1000 × randn ( k ) , for   k = 1 , , 141 .
The piecewise interpolation filter F ( 28 ) and the GOL filters [5] estimate the reference signals with the associated errors represented in Figure 8 (b). As in Example 11 above, in this case, the errors associated with the averaging polynomial filter [10,12] are much grater (of order O ( 10 10 ) ) and they are not presented in Figure 8 (b).
Examples 11 and 12 further demonstrate the advantages of the proposed piecewise interpolation filter.
Figure 8. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 11 and 12.
Figure 8. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 11 and 12.
Preprints 188126 g008
Figure 9. Examples of selected signals to be estimated from observed data.
Figure 9. Examples of selected signals to be estimated from observed data.
Preprints 188126 g009
Figure 10. Examples of the observed signal and the estimates obtained by different filters.
Figure 10. Examples of the observed signal and the estimates obtained by different filters.
Preprints 188126 g010
Figure 11. Illustration of the errors associated with the piecewise interpolation filters F ( p 1 ) and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 1–4.
Figure 11. Illustration of the errors associated with the piecewise interpolation filters F ( p 1 ) and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 1–4.
Preprints 188126 g011
Figure 12. Examples of selected signals to be estimated from observed data considered in Example 5-9.
Figure 12. Examples of selected signals to be estimated from observed data considered in Example 5-9.
Preprints 188126 g012
Figure 13. Examples of the observed signal and the estimates obtained by different filters.
Figure 13. Examples of the observed signal and the estimates obtained by different filters.
Preprints 188126 g013
Figure 14. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) of order p and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 5–8.
Figure 14. Illustration of errors associated with the piecewise interpolation filter F ( p 1 ) of order p and the generic optimal linear (GOL) filters [5] applied to signals described in Examples 5–8.
Preprints 188126 g014
Figure 15. Illustration of errors associated with the averaging polynomial filters [10,12] in Examples 9 and 10.
Figure 15. Illustration of errors associated with the averaging polynomial filters [10,12] in Examples 9 and 10.
Preprints 188126 g015

5.6. Summary of Simulations

The above simulations confirm the theoretical results obtained in Theorems 1–3. In particular, Figs. Figure 11 and Figure 14 demonstrate that the error associated with the piecewise interpolation filter F ( p 1 ) decreases when the number of sub-filters F 1 , , F p , p, increases.
A comparison between the proposed filter F ( p 1 ) and the known related filters [5,10,12,22,23] has been done. The filter F ( p 1 ) estimates the reference signals with the accuracies that are much better than those of the generic optimal linear (GOL) filters [5] and the averaging polynomial filter [10,12]. Further, the filters proposed in [22,23] fail in processing the signals under consideration. This is because the observed signals in (40), (49), (50) and (51) are grossly corrupted and, therefore, the inverse matrices used in the filter structures in [23] do not exist. The technique in [22] requires the use of the reference signal in the proposed filter which is supposed to be unknown in the simulations above.
The filters have been applied to the different signal sets (presented in Section 5.2 and Section 5.3), using different forms of noise (given in (40), (49), (50) and (51)).
computational work associated with the proposed filter F ( p 1 ) is substantially less than that associated with the known filters discussed in Section 1 (in particular, with the filters in [11]–23]). This is because, for the processing of a data set containing N signals, filter F ( p 1 ) requires computation of p covariance matrices with p N while the known filters require computation of N matrices (in the above Examples, p = 5 , 7 , 15 , 28 , respectively, and N = 141 ).

6. Conclusions

The technique of the constructive approximation of a nonlinear operator K Y K X has been provided. Here, K X = { x ( t , · ) L 2 ( Ω , R m ) | t T } and K Y = { y ( t , · ) L 2 ( Ω , R n ) | t T } where T : = [ a , b ] R and Ω = { ω } is the set of outcomes of a probability space. The device behind the proposed method is based on a special extension of the piecewise linear interpolation technique to the case of stochastic sets K X and K Y . The proposed methodology is motivated by the problems arising in signal processing where the nonlinear operator is interpreted as a nonlinear system (or a nonlinear filter) transforming a set of stochastic signals KY to the set of stochastic signals KX Therefore, the provided technique provides an effective way to transform large data sets.
Distinctive features of the approach are as follows.
(i) The proposed filter ( p 1 ) : K Y K X is nonlinear and is presented in the form of a sum with p 1 terms where each term, j : K Y , j K X , j , is interpreted as a particular sub-filter. Here, K Y , j and K X , j are ‘small’ pieces of K Y and K X , respectively.
(ii) The prime idea is to exploit a prior information only onfew reference signals, p, from the set K X that contains N p signals (or even an infinite number of signals) and determine j separately, for each pieces K Y , j and K X , j , so that the associated error is minimal. In other words, the filter ( p 1 ) is flexible to changes in the sets of observed and reference signals K Y and K X , respectively.
(iii) Due to the specific way of determining j, the filter ( p 1 ) provides a smaller associated error than that for the processing of the whole set K Y by a filter which is not specifically adjusted to each particular piece K Y , j . Moreover, the error associated with our filter decreases when the number of its terms, 1 , , ( p 1 ) , increases.
(iv) While the proposed filter ( p 1 ) processes arbitrarily large (and even infinite) signal sets, the filter is nevertheless fixed for all signals in the sets.
(v) The filter ( p 1 ) is determined in terms of pseudo-inverse matrices so that the filter always exists.
(vi) computational load associated with the filter ( p 1 ) is less than that associated with other known filters applied to the processing of large signal sets.

Authors’ Contributions

Conceptualization, methodology, writing original draft - A.T. Numerical simulations - A.T. and P.P. Algorithm - P.P. Matlab codes - P.P. English amelioration - P.P. The authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Proof of Theorem 1: It follows from (8) and (11) that α j , for j = 1 , , p 1 , is given by
α j = x ^ ( t j , ω ) B j [ y ( t j , ω ) ] .
Further, for α j given by (A1),
[ x ( t j + 1 , · ) α j ] B j [ y ( t j + 1 , · ) ] Ω 2
= z ( t j , t j + 1 , · ) B j [ w ( t j , t j + 1 , · ) ) ] Ω 2 = t r { E z j z j E z j w j B j T B j E w j z j + B j E w j w j B j T } = E z j z j 1 / 2 2 E z j w j ( E w j w j 1 / 2 ) 2 + ( B j E z j w j E w j w j ) E w j w j 1 / 2 2
= E z j z j 1 / 2 2 E z j w j ( E w j w j 1 / 2 ) 2 + E z j w j ( E w j w j 1 / 2 ) B j E w j w j 1 / 2 2 ,
where · is the Frobenius norm. The latter is true because
E w j w j E w j w j 1 / 2 = ( E w j w j 1 / 2 )
and
E z j w j E w j w j E w j w j = E z j w j
by Lemma 24 in [5]. Thus, the second expression in (11) is reduced to the problem
min B j E z j w j ( E w j w j 1 / 2 ) B j E w j w j 1 / 2 2 .
It is known (see, for example, [5], p. 304) that the solution of problem (A5) is given by (19). The equation (17) follows from (6) and (A1).
Theorem 1 is proven. □
Proof of Theorem 2: For t [ t j , t j + 1 ] and F j defined by (17)–(19),
x ( t , ω ) F [ y ( t , ω ) ] = x ( t , ω ) F j [ y ( t , ω ) ] = x ( t , ω ) x ^ ( t j , ω ) + B j y ( t j , ω ) B j y ( t , ω ) = [ x ( t , ω ) x ( t j + 1 , ω ) ] + z ( t j , t j + 1 , ω ) B j w ( t j , t j + 1 , ω ) + B j [ y ( t j + 1 , ω ) y ( t , ω ) ] .
Then (A6) implies
x ( t , ω ) F [ y ( t , ω ) ] T , Ω 2 x ( t , ω ) x ( t j + 1 , ω ) T , Ω 2 + z ( t j , t j + 1 , ω ) B j w ( t j , t j + 1 , ω ) Ω 2 + B j [ y ( t j + 1 , ω ) y ( t , ω ) ] T , Ω 2
where z ( t j , t j + 1 , ω ) B j w ( t j , t j + 1 , ω ) Ω 2 = z ( t j , t j + 1 , ω ) B j w ( t j , t j + 1 , ω ) T , Ω 2 .
It follows from (A2) and (A3) that for B j given by (19),
z ( t j , t j + 1 , ω ) B j w ( t j , t j + 1 , ω ) Ω 2 = E z j z j 1 / 2 2 E z j w j ( E w j w j 1 / 2 ) 2 .
Then (16)–(19), (23) and (A6)–(A8) imply that for all t [ a , b ] and ω Ω , (24) is true. □
Proof of Theorem 3: The relation (22) implies that
x ( t , ω ) F [ y ( t , ω ) ] T , Ω 2 = 1 b a j = 1 p 1 t j t j + 1 x ( t , ω ) F j [ y ( t , ω ) ] Ω 2 d t ,
where
x ( t , ω ) F j [ y ( t , ω ) ] Ω 2 = x ( t , ω ) x ^ ( t j , ω ) + B j [ y ( t j , ω ) B j y ( t , ω ) ] Ω 2 x ( t , ω ) x ( t j , ω ) Ω 2 + x ( t j , ω ) x ^ ( t j , ω ) Ω 2 + B j [ y ( t j , ω ) B j y ( t , ω ) ] Ω 2 .
Then
t j t j + 1 x ( t , ω ) F j [ y ( t , ω ) ] Ω 2 d t t j t j + 1 x ( t , ω ) x ( t j , ω ) Ω 2 d t + t j t j + 1 x ( t j , ω ) x ^ ( t j , ω ) Ω 2 d t + B j t j t j + 1 y ( t j , ω ) y ( t , ω ) Ω 2 d t
λ j ( Δ t j ) 2 + x ( t j , ω ) x ^ ( t j , ω ) Ω 2 Δ t j + B j γ j ( Δ t j ) 2
Let us consider an estimate of x ( t j , ω ) x ^ ( t j , ω ) Ω 2 , for j = 1 , , p 1 . To this end, let us denote Δ t = max j = 1 , , p 1 Δ t j .
For j = 1 , i.e. for t [ t 1 , t 2 ] ,
x ( t , ω ) F 1 y ( t , ω ) Ω 2 x ( t , ω ) x ( t 1 , ω ) Ω 2 + x ( t 1 , ω ) x ^ ( t 1 , ω ) Ω 2 + B 1 y ( t 1 , ω ) y ( t , ω ) Ω 2 λ 1 Δ t 1 + c 1 Δ t 1 + B 1 γ 1 Δ t 1 β 1 Δ t ,
where β 1 = λ 1 + c 1 + B 1 γ 1 . In particular, the latter implies
x ( t 2 , ω ) x ^ ( t 2 , ω ) Ω 2 = x ( t 2 , ω ) F 1 y ( t 2 , ω ) Ω 2 β 1 Δ t
For j = 2 , i.e. for t [ t 2 , t 3 ] ,
x ( t , ω ) F 2 y ( t , ω ) Ω 2 x ( t , ω ) x ( t 2 , ω ) Ω 2 + x ( t 2 , ω ) x ^ ( t 2 , ω ) Ω 2 + B 2 y ( t 2 , ω ) y ( t , ω ) Ω 2 λ 2 Δ t 2 + β 1 Δ t + B 2 γ 2 Δ t 2 β 2 Δ t ,
where β 2 = λ 2 + β 1 + B 2 γ 2 . In particular, then it follows that
x ( t 3 , ω ) x ^ ( t 3 , ω ) Ω 2 = x ( t 3 , ω ) F 2 y ( t 3 , ω ) Ω 2 β 2 Δ t .
On the above basis, let us assume that, for j = k 1 with k = 2 , , p 1 , i.e. for t [ t k 1 , t k ] ,
x ( t k , ω ) x ^ ( t k , ω ) Ω 2 = x ( t k , ω ) F k 1 y ( t k , ω ) Ω 2 β k 1 Δ t
where β k 1 is defined by analogy with β 2 .
Then, for j = k with k = 2 , , p 1 , i.e. for t [ t k , t k + 1 ] ,
x ( t , ω ) F k y ( t , ω ) Ω 2 x ( t , ω ) x ( t k , ω ) Ω 2 + x ( t k , ω ) x ^ ( t k , ω ) Ω 2 + B k y ( t k , ω ) y ( t , ω ) Ω 2 λ k Δ t k + β k 1 Δ t + B k γ 2 Δ t k β k Δ t ,
where β k = λ k + β k 1 + B k γ k . Thus, the following is true:
x ( t k + 1 , ω ) x ^ ( t k + 1 , ω ) Ω 2 = x ( t k + 1 , ω ) F k y ( t k + 1 , ω ) Ω 2 β k Δ t .
Therefore, (A10), (A11) and (A12) imply
t j t j + 1 x ( t , ω ) F j [ y ( t , ω ) ] Ω 2 d t λ j ( Δ t j ) 2 + β j 1 ( Δ t j ) 2 + B j γ j ( Δ t j ) 2 η j ( Δ t ) 2
where η j = λ j + β j 1 + B j , and then it follows from (A9)–(A11) and (A13) that for all t [ a , b ] ,
x ( t , ω ) F [ y ( t , ω ) ] T , Ω 2 1 b a j = 1 p 1 η j ( Δ t ) 2 = 1 b a Δ t j = 1 p 1 η j Δ t .
Let us now choose c R and d R so that Δ t = d c p and partition interval [ c , d ] R by points τ 1 , , τ p so that c = τ 1 and τ j = τ 1 + j Δ t with j = 1 , , p . There exists an integrable (bounded) function φ : [ c , d ] R such that, for ξ j ( τ j , τ j + 1 ) , φ ( ξ j ) = η j . Then
lim Δ t j = 1 p 1 η j Δ t = lim Δ t j = 1 p 1 φ ( ξ j ) Δ t = c d φ ( τ ) d τ < + .
Thus,
1 b a Δ t j = 1 p 1 η j Δ t 0 a s Δ t 0 .
As a result, (A14)–(A16) imply (26). □

References

  1. Chen, J.; Benesty, J.; Huang, Y.; Doclo, S. New Insights Into the Noise Reduction Wiener Filter. IEEE Trans. on Audio, Speech, and Language Processing 2006, 14(No. 4), 1218–1234. [Google Scholar] [CrossRef]
  2. Spurbeck, M.; Schreier, P. Causal Wiener filter banks for periodically correlated time series. Signal Processing 2007, 87(6), 1179–1187. [Google Scholar] [CrossRef]
  3. Goldstein, J. S.; Reed, I.; Scharf, L. L. A Multistage Representation of the Wiener Filter Based on Orthogonal Projections. IEEE Trans. on Information Theory 1998, 44, 2943–2959. [Google Scholar] [CrossRef]
  4. Y. Hua, M. Nikpour, and P. Stoica, “Optimal Reduced-Rank estimation and filtering,”IEEE Trans. on Signal Processing,vol. 49, pp. 457-469, 2001.
  5. A. Torokhti and P. Howlett, Computational Methods for Modelling of Nonlinear Systems, Elsevier, 2007.
  6. E. D. Sontag, Polynomial Response Maps, Lecture Notes in Control and Information Sciences, 13, 1979.
  7. Chen, S.; Billings, S. A. Representation of non-linear systems: NARMAX model. Int. J. Control 1989, 49(no. 3), 1013–1032. [Google Scholar] [CrossRef]
  8. V. J. Mathews and G. L. Sicuranza, Polynomial Signal Processing, J. Wiley & Sons, 2001.
  9. Torokhti, A.; Howlett, P. Optimal Transform Formed by a Combination of Nonlinear Operators: The Case of Data Dimensionality Reduction. IEEE Trans. on Signal Processing 2006, 54(No. 4), 1431–1444. [Google Scholar]
  10. Torokhti, A.; Howlett, P. Filtering and Compression for Infinite Sets of Stochastic Signals. Signal Processing, 2009; Volume 89, pp. 291–304. [Google Scholar]
  11. Vesma, J.; Saramaki, T. Polynomial-Based Interpolation Filters - Part I: Filter Synthesis; Circuits, Systems, and Signal Processing, 2007; Volume 26, Number 2, pp. Pages 115–146. [Google Scholar]
  12. Torokhti, A.; Manton, J. Generic Weighted Filtering of Stochastic Signals. IEEE Trans. on Signal Processing 2009, 57(issue 12), 4675–4685. [Google Scholar] [CrossRef]
  13. Torokhti, A.; Miklavcic, S. Data Compression under Constraints of Causality and Variable Finite Memory. Signal Processing 2010, 90(Issue 10), 2822–2834. [Google Scholar] [CrossRef]
  14. Babuska, I.; Banerjee, U.; Osborn, J. E. Generalized finite element methods: main ideas, results, and perspective. International Journal of Computational Methods 2004, 1(1), 67–103. [Google Scholar] [CrossRef]
  15. Kang, S.; Chua, L. A global representation of multidimensional piecewise-linear functions with linear partitions. IEEE Trans. on Circuits and Systems 1978, 25(Issue:11), 938–940. [Google Scholar] [CrossRef]
  16. Chua, L.O.; Deng, A.-C. Canonical piecewise-linear representation. IEEE Trans. on Circuits and Systems 1988, 35(Issue:1), 101–111. [Google Scholar] [CrossRef]
  17. Lin, J.-N.; Unbehauen, R. Adaptive nonlinear digital filter with canonical piecewise-linear structure. IEEE Trans. on Circuits and Systems 1990, 37(Issue:3), 347–353. [Google Scholar] [CrossRef]
  18. J.-N. Lin and R. Unbehauen, Canonical piecewise-linear approximations,IEEE Trans. on Circuits and Systems I: Fundamental Theory and Applications,39 Issue:8, pp. 697 - 699, 1992.
  19. Gelfand, S.B.; Ravishankar, C.S. A tree-structured piecewise linear adaptive filter. IEEE Trans. on Inf. Theory 1993, 39(issue 6), 1907–1922. [Google Scholar] [CrossRef]
  20. Heredia, E.A.; Arce, G.R. Piecewise linear system modeling based on a continuous threshold decomposition. IEEE Trans. on Signal Processing 1996, 44(Issue:6), 1440–1453. [Google Scholar] [CrossRef]
  21. Feng, G. Robust filtering design of piecewise discrete time linear systems. IEEE Trans. on Signal Processing 2005, 53(Issue:2), 599–605. [Google Scholar] [CrossRef]
  22. Russo, F. Technique for image denoising based on adaptive piecewise linear filters and automatic parameter tuning. IEEE Trans. on Instrumentation and Measurement 2006, 55(Issue:4), 1362–1367. [Google Scholar] [CrossRef]
  23. Cousseau, J.E.; Figueroa, J.L.; Werner, S.; Laakso, T.I. Efficient Nonlinear Wiener Model Identification Using a Complex-Valued Simplicial Canonical Piecewise Linear Filter. IEEE Trans. on Signal Processing 2007, 55(Issue:5), 1780–1792. [Google Scholar] [CrossRef]
  24. P. Julian, A. Desages, B. D’Amico, Orthonormal high-level canonical PWL functions with applications to model reduction,IEEE Trans. on Circuits and Systems I: Fundamental Theory and Applications,47 Issue:5, pp. 702 - 712, 2000.
  25. T. Wigren, Recursive Prediction Error Identification Using the Nonlinear Wiener Model,Automatica,29, 4, pp. 1011–1025, 1993.
  26. G. H. Golub and C. F. van Loan, Matrix Computations, Johns Hopkins University Press, Baltimore, 1996.
  27. T. Anderson,An Introduction to Multivariate Statistical Analysis,New York, Wiley, 1984.
  28. L. I. Perlovsky and T. L. Marzetta, Estimating a Covariance Matrix from Incomplete Realizations of a Random Vector,IEEE Trans. on Signal Processing, 40, pp. 2097-2100, 1992.
  29. O. Ledoit and M. Wolf, A well-conditioned estimator for large-dimensional covariance matrices,J. Multivariate Analysis88, pp. 365–411, 2004.
1
We say a stochastic vector x is finite if its realization has a finite number of scalar components.
2
To the best of our knowledge, the exception is the methodology in [10,12] where the filtering techniques exploit information on reference signals in the form of the vector obtained from averaging over reference signal sets.
3
This means that any desired accuracy is achieved theoretically, as is shown in Section 4.4 below. In practice, of course, the accuracy is increased to a prescribed reasonable level.
4
As usually, Ω = { ω } is the set of outcomes, Σ a σ -field of measurable subsets in Ω and μ : Σ [ 0 , 1 ] an associated probability measure on Σ . In particular, μ ( Ω ) = 1 .
5
Hereinafter, we will use a non-curly symbol to denote an operator and associated matrix (e.g., the operator j : L 2 ( Ω , R n ) L 2 ( Ω , R m ) and the associated matrix F j R m × n are denoted by F j ).
6
It is worthwhile to note that it is not assumed that the covariance matrices are known for each signal pair from K X × K Y , { x ( t , · ) , y ( t , · ) } with t [ a , b ] .
7
As it has been mentioned in Section 3.4, x ^ ( t 1 , · ) can be determined by the known methods.
8
The database is available in http://sipi.usc.edu/services/database.html.
9
The database is available in http://sipi.usc.edu/services/database.html.
10
The database is available in http://sipi.usc.edu/services/database.html.
11
The database is available in http://sipi.usc.edu/services/database.html.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated