Preprint
Article

This version is not peer-reviewed.

Sequential Tests and Reliability Integral Theory

Submitted:

03 June 2025

Posted:

04 June 2025

You are already at the latest version

Abstract
We use the data of two papers “Statistical Inference on the Shape Parameter of Inverse Generalized Weibull Distribution” (Zhuang et al.), and “On Designing of Bayesian Shewhart-Type Control Charts for Maxwell Distributed Processes with Application of Boring Machine” (Alshahrani et al.) to compare the above authors findings with ours. From the analysis we get different results: the cause is that they use the Probability Limits of the PI (Probability Interval) as they were the Confidence Limits (Control Limits of the Control Charts, CCs). The Control Limits in the Shewhart CCs are based on the Normal Distribution (Central Limit Theorem, CLT) and are not valid for non-normal distributed data: consequently, the decisions about the “In Control” (IC) and “Out Of Control” (OOC) states of the process are wrong. The Control Limits of the CCs are wrongly computed, due to unsound knowledge of the fundamental concept of Confidence Interval. Minitab and other software (e.g. JMP, SAS) use the “T Charts”, claimed to be a good method for dealing with “rare events”, but their computed Control Limits of the CCs are wrong. The same happens for the Confidence Limits of the parameters of the distribution involved in the papers (Weibull, Inverse Weibull, Gamma, Binomial, Maxwell). We will show that the Reliability Integral Theory (RIT) is able to solve these problems and the Sequential way of dealing with data.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

Since 1989, the author (FG) tried to inform the Scientific Community about the flaws in the use of (“wrong”) quality methods for making Quality [1] and in 1999 about the GIQA (Golden Integral Quality Approach) showing how to manage Quality during all the activities of the Product and Process Development in a Company [2], including the Process Management and Control Charts (CC) for Process Control. Control Charts (CC) use sequentially the collected data to assess if a Production or Service process output is to be considered In Control (IC) or Out Of Control (OOC); the decision is very important for taking Corrective Actions (CA), if needed.
To show our Theory we will use some of the data found in the papers [3,4,5].
But before that we mention the very interesting the statements in the Excerpt 1:
Excerpt 1. From the paper “Misguided Statistical Process Monitoring Approaches”
Excerpt 1. From the paper “Misguided Statistical Process Monitoring Approaches”
Preprints 162230 sch001
We agree with the authors in the excerpt 1, but, nevertheless, they did not realise the problem that we are showing here: wrong Control Limits in CCs for Rare Events, with data exponentially or Weibull or Maxwell distributed. Several papers compute “a-scientific” control limits… See References…
We will show that the Test of Hypotheses and the Confidence Intervals (CI) are intimately related and so equivalent for decision making. Using the data in [3,4,5] with good statistical methods [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33] we give our “reflections on Sequential Methods and Control Charts (CCs)”.
We will try to state that several papers (that are not cited here, but you can find in the “Garden of flowers” [24] and some in the Appendix A) compute in an a-scientific way (see the formulae in the Appendix C) the Control Limits of CCs for “Individual Measures or Exponential, Weibull, Maxwell and Gamma distributed data”, indicated as I-CC (Individual Control Charts); we dare to show, to the Scientific Community, how to compute the True Control Limits (True Confidence Limits). If the author is right, then all the decisions, taken up today, have been very costly to the Companies using those Control Limits; therefore, “Corrective Actions” are needed, according to the Quality Principles, because NO “Preventive Actions” were taken [1,2,27,28,29,30,31,32,33,34,35,36]: this is shown through the suggested published papers. Humbly, given our strong commitment to Quality [34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], we would dare to provide the “truth”: Truth makes you free [hen (“hic et nunc”=here and now)].
On 22nd of February 2024, we found the paper “Publishing an applied statistics paper: Guidance and advice from editors” published in Quality and Reliability Engineering International (QREI-2024, 1-17) [by C. M. Anderson-Cook, Lu, R. B. Gramacy, L. A. Jones-Farmer, D. C. Montgomery, W. H. Woodall; the authors have important qualifications and Awards]; since I-CC is a part of “applied statistics” we think that their hints will help: the authors’ sentence “Like all decisions made in the face of uncertainty, Type I (good papers rejected) and Type II (flawed papers accepted) errors happen since the peer review process is not infallible.” is very important for this paper: the interested readers can see [34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
To let the reader follow our way of approaching the problem of estimation we will use various figures and some data: this is caused by the fact that there are wrong ideas in the literature.
By reading [24] and other papers, the readers are confronted with this type of practical problem: we have a warehouse with two departments
a)
in the 1st of them, we have a sample (the “The Garden of flowers… in [24]”) of “products (papers)” produced by various production lines (authors)
b)
while, in the other, we have some few products produced by the same production line (same author)
c)
several inspectors (Peer Reviewers, PRs) analyse the “quality of the products” in the two departments; the PRs can be the same (but we do not know) for both the departments
d)
The final result, according to the judgment of the inspectors (PRs), is the following: the products stored in the 1st dept. are good, while the products in the 2nd dept. are defective. It is a very clear situation, as one can guess by the following statement of a PR: “Our limits [in the 1st dept.] are calculated using standard mathematical statistical results/methods as is typical in the vast literature of similar papers [24].” See the standard mathematical statistical results/methods in the Appendix A and meditate (see the formulae there)!
Hence, the problem becomes “…the standard … methods as is typical …”: are those standards typical methods (in the “The Garden … in [24]” scientific?
To understand we need to give now “Some ideas on Hypothesis Test and The Statistical Hypotheses with the related risks”; in alternative, you can read [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33,34,35,36].
We define as statistical hypothesis a statement about a population parameter θ (e.g. the ′′true′′ mean, the ′′true′′ shape, the ′′true′′ variance, the ′′true′′ reliability, the ′′true′′ failure rate, …n that we assume to exists and has a value even though it is unknown to us), related to the statistical model F(x|θ) associated with a random variable (RV) X. The set of all the possible values of the parameter is called the parameter space Θ. The goal of a hypothesis test is to decide, based on a sample drawn from the population, which value hypothesized for the population parameter of the parameter space Θ can be accepted as true. Remember: nobody knows the truth…
Generally, two competitive hypotheses are defined, the null hypothesis H0 and the alternative hypothesis H1.
A hypothesis testing procedure (or simply a hypothesis test) is a rule (decision criterion) that specifies
  • for which sample values the decision is made to «accept» H0 as true,
  • for which sample values H0 is rejected and then H1 is accepted as true.
based on managerial/Statistics which defines
  • the test statistic (a formula to analyse the data)
  • the critical region C (rejection region)
to be used for decisions, with the stated risks: decision criterion.
The subset of the sample space for which H0 will be rejected is called rejection region (or critical region). The complement of the rejection region is called the acceptance region.
If θ denotes the population parameter, the general form of the null hypothesis is H0: {θ∈Θ0} versus the alternative hypothesis H1: {θ∈Θ1}, where Θ0 is a subset of the parameter space Θ and Θ1 a subset disjoint from Θ0.; Θ0∪Θ1= Θ and Θ0∩Θ1=∅; before collecting any data, with H0 we accept a probability α of wrong decision, while with H1 we accept a probability β of wrong decision. A hypothesis test of H0: {θ∈Θ0} versus the alternative hypothesis H1: {θ∈Θ1} might make one of two types of errors, traditionally named Type I Error and Type II Error; their probabilities are indicated as α and β.
If «actually (but we do not know)» H0: {θ∈Θ0} is true and the hypothesis test (the rule, the computed quantity S, in the figure 1), due to the collected data, incorrectly decides to reject H0 then the test (and the Experimenter, the Manager, the Researcher, the Scholar who follow the rule) makes a Type I Error, whose probability is α. If, on the other hand, «actually (but we do not know)» θ∈Θ1 but the test (the rule), due to the collected data, incorrectly decides to accept H0 then the test (and the Experimenter, the Manager, the Researcher, the Scholar who follow the rule) makes a Type II Error, whose probability is β.
These two different situations are depicted in the Table 1 (for simple parametric hypotheses).
The framework of a test of hypothesis is depicted in the Figure 1.
Notice that when we decide to “accept the null hypothesis” in reality we use a short-hand statement saying that “we do not have enough evidence to state the contrary”. It is evident that
α = P r e j e c t   H 0 | H 0   t r u e   and   β = P a c c e p t   H 0 | H 0   f a l s e
A likelihood ratio test is any test that has a rejection region of the following form {s(D): q(D)≥c}, where c is any number satisfying 0≤c≤1 and s(D) is the “statistic” by which we elaborate the data of the empirical sample D. This test is a measure of how much the evidence, provided by the data D, supports H0.
This has great importance for Control Charts, as you can see in the Figure 3.
Suppose C is the “critical” (or rejection) region for a test, based on a «statistic s(D)» (the formula to elaborate the sampled data D, providing the value s(D).
Then for testing H0: {θ∈Θ0}, the test makes a mistake if «s(D)∈C», so that the probability of a Type I Error is α=P(S(D)∈C) [S(D) is the random variable giving the result s(D)]. It is important the power of the test 1-β, which is the probability of rejecting H0 when in reality H0 is false
1 β = P r e j e c t   H 0 | H 0   f a l s e
Therefore, the power function of a hypothesis test with rejection region C is the function of θ defined by β(θ)=P(S(D)∈C). The function 1-β(θ), power function, evaluated at the value θ, is often named the Operating Characteristic curve [OC curve].
To find the RV S(D) and the region C, we use the likelihood function L(θ|D={x1, x2, …, xn])
L θ D = 1 n f ( x i )
Let L0 be the Likelihood function L(θ0|D) and L1 be the Likelihood function L(θ1|D): the most powerful test is the one that has the most powerful critical region C={s(D): q(n)=L1/L0≥kα}, where q(n) is the Likelihood Ratio L1/L0 and the quantity kα is chosen in such a way that the Type I Error has a risk (probability) α as in the formula (4), with fixed n (the sample size),
. . q ( n ) k α L 0   d x 1 d x 2 . . d x n = α
The most powerful critical region C has the highest power 1-β(θ).
Let CRn be the “Critical Region” found by (4) and βn be the probability (5), function of n,
β n = . . q ( n ) C R n L 1   d x 1 d x 2 . . d x n
By (4) and (5), increasing n, we arrive to select a final sample size n, such that βn=β, the desired risk.
Usually when an efficient estimator exists, this provides then a powerful statistic, giving the most powerful test. [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]
We will use α = β in the following discussion. After the data analysis, we can decide if the data suggest us to “accept (= not reject)” H0: {θ∈Θ0} or “accept” H1: {θ∈Θ1},and after that we can compute the Confidence Interval, CI=θL-------θU, of the parameter θ, with Confidence Level C L = 1 α = 1 α / 2 + β / 2 = 1 α / 2 + α / 2 .
When we consider the Control Charts we want to test the two Hypotheses H0: {the process is “IC (In Control)”} against H1: { the process is “OOC (Out Of Control)”}, and after the data analysis we can compute the Control Interval (which is actually a Confidence Interval), LCL-------UCL.
If we use the Table 3 data (remission time of 128 bladder cancer patients) it is easy to see that (as said with the above warehouse example) the practical problem becomes a Theoretical one [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58] (all references and Figure 21). Since those data are well “exponentially distributed” we anticipate here, immediately, the wrong formulae (either using the parameter θ = θ 0 or its estimate t ¯ 0 , with α = 0.0027 ) in the formula (6) (as you can find in [24])
L C L = θ 0 l n 1 α / 2 = 0.00135   t ¯ 0           U C L = θ 0 l n α / 2 = 6.6077   t ¯ 0
The readers should understand clearly the Theoretical and Practical Difference between L------U (the Probability Interval) and LCL------UCL (the Confidence Interval), pictorially shown in the Figure 2: the two lines L and U depends on the parameter θ (to be estimated) and on the two probabilities α and β, while the two points L and U depends on the assumed value θ0 of the parameter and on the two chosen probabilities α and β; after the data analysis, we compute the estimate t ¯ 0 of the parameter θ and from that the Confidence Interval LCL------UCL, with Confidence Level C L = 1 α . It is clear now the wrong ideas in the formulae (6).
In the formulae (6), for the interval LCL------UCL (named Control Interval, by the authors [24]), the LCL actually must be L and the UCL actually must be U, vertical interval L------U (figure 2); the actual interval LCL------UCL is the horizontal one in the figure 2, which is not that of the formulae (6). Since the errors have been continuing for at least 25 years, we dare to say that this paper is an Education Advance for all the Scholars, for the software sellers and the users: they should study the books and papers in [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
The readers could think that the I-CCs are well known and well dealt in the scientific literature about Quality. We have some doubt about that: we will show that, at least in one field, the I-CC_TBE (with TBE, Time Between Event data) usage, it is not so: there are several published papers, in “scientific magazines and Journals (well appreciated by the Scholars)” with wrong Control Limits; a sample of the involved papers (from 1994 to January 2024) can be found in [23,24]”. Therefore, those authors do not extract the maximum information from the data in the Process Control. “The Garden…” [24] and the excerpts 1, with the Deming’s statements, constitute the Literature Review.
Excerpt 2. Some statements of Deming about Knowledge and Theory (Deming 1986, 1997)
Excerpt 2. Some statements of Deming about Knowledge and Theory (Deming 1986, 1997)
Preprints 162230 sch002
We hope that the Deming statements about knowledge will interest the Readers (Excerpt 2).
Figure 3. LCL and UCL of Control Charts with their risks.
Figure 3. LCL and UCL of Control Charts with their risks.
Preprints 162230 g003
The good Managers, Researchers, Scholars do not forget that the two risks always are present and therefore they must take care of the power of the test 1-β, they use for the decision (as per the principles F1 and F2) [24,25,26,27,28,29,30].
Such Managers, Researchers, Scholars use the Scientific Method.
It is important to state immediately and in an explicit way that
the risks must be stated,
together with the goals (the hypotheses),
BEFORE any statistical (reliability) test is carried out and data are analysed.
For demonstration of reliability characteristics, with reliability tests, Managers, Students, Researchers and Scholars must take into account, according the F1 principle, the very great importance of W. E. Deming statements (Excerpt 2): from these, unfortunately for Quality, for the Customers, for the Users and for the Society, this devastating result
The result is that hundreds of people are learning what is wrong. I make this statement on the basis of experience, seeing every day the devastating effects of incompetent teaching and faulty applications.
In many occasions and several Conferences on Total Quality Management for Higher Education Institutions, [Toulon (1998), Verona (1999), Derby (2000), Mons (2001), Lisbon (2002), Oviedo (2003), Palermo (2005), Paisley (2006), Florence (2008), Verona (2009)] the author (FG) showed many real cases, found in books and magazines specialized on Quality related to concepts, methods and applications wrong, linked to Quality [21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58]. All the very many documents published (more than 250) by F. Galetto show the profound truth that
facts and figures are useless, if not dangerous, without a sound theory (F. Galetto, 2000),
Brain is the most important asset: let’s not forget it. (F. Galetto, 2003),
All that is particularly important for the analysis of any type of data (quality or reliability).
Sequential sampling
Sequential sampling refers to a routine in which each unit is “measured” about a kind of quantity of interest (length, weight, defectiveness, duration, reliability, failure rate, …) and the “cumulated” quantity is employed in decision taking about the acceptance of the null hypothesis H0, with 1-α as the probability of Accepting H0, when it is true. At any “measurement” 1, 2, …, k, decision rules are required to provide a decision between three alternatives, a) Acceptance of H0, b) Rejection of H0, or c) continuation of sampling (by taking a new unit); this process continues until a decision a) or b) is taken; the number of items then drawn defines the sample size: sequential sampling, in general, leads to an expected sample size smaller than other sampling methods.
As seen before, the likelihood ratio test is used; the likelihood ratio test statistic for testing H0 versus H1 is the ratio q(k)=L1(k)/L0(k), where k is the present (variable) sample size, with the rules (after Wald, 1945), a) if q ( k ) B β / ( 1 α ) then retain H0, b) if q ( k ) A ( 1 β ) / α then choose H1 and reject H0, if B β / ( 1 α ) q ( k ) A ( 1 β ) / α then continue sampling. The two quantities A and B are not computed easily: likely, Wald provided the approximations A ( 1 β ) / α and B β / ( 1 α ) . These rules, under a suitable transformation of scale, lead to two Decision parallel lines, the Acceptance line and the Rejection line: the successive points of q * ( k ) the “transformed value of q(k)” generates a random walk path; when the path reaches a decision line the inspection ceases, while when the path is contained within the two lines, the sampling is continued and new items are tested and measured.

2. Materials and Methods

2.1. A Reduced Background of Statistical Concepts

After the ideas given in the Introduction, we provide the following ones essential to understand the “problems related to I-CC and sequential estimation” as we found in the literature. We suggest it for the formulae given and for the difference between the concepts of PI (Probability Interval) and CI (Confidence Interval): this is overlooked in “The Garden … [24]”
Engineering Analysis is related to the investigation of phenomena underlying products and processes; the analyst can communicate with the phenomena only through the observed data, collected with sound experiments (designed for the purpose): any phenomenon, in an experiment, can be considered as a measurement-generating process [MGP, a black box that we do not know] that provides us with information about its behaviour through a measurement process [MP, known and managed by the experimenter], giving us the observed data (the “message”).
It is a law of nature that the data are variable, even in conditions considered fixed, due to many unknown causes.
MGP and MP form the Communication Channel from the phenomenon to the experimenter.
The information, necessarily incomplete, contained in the data, has to be extracted using sound statistical methods (the best possible, if we can). To do that, we consider a statistical model F(x|θ) associated with a random variable (RV) X giving rise to the measurements, the “determinations” D={x1, x2, …, xn} of the RV, constituting the “observed sample” D; n is the sample size. Notice the function F(x|θ) [a function of real numbers, whose form we assume we know] with the symbol θ accounting for an unknown quantity (or some unknown quantities) that we want to estimate (assess) by suitably analysing the sample D.
We indicate by f x | θ = d F x | θ / d x the pdf (probability density function) and by F x | θ the Cumulative Function, where θ is the set of the parameters of the functions.
We state in the Table 2 a sample of models where θ is a set of parameters:
Two important models are the Normal and the Exponential, but we consider also the others for comparison. When θ = μ , σ 2 we have the Normal model, written as N (x| μ , σ 2 ), with (parameters) mean E[X]=μ and variance Var[X]=σ2 with pdf
f ( x | μ , σ 2 ) = n x μ , σ 2 = 1 2 π σ e x μ 2 / ( 2 σ 2 )
When θ = θ we have Exponential model, E(x|θ), with (the single parameter) mean E[X]= θ = 1 / λ (variance Var[X]= θ 2 = 1 / λ 2 ), whose pdf is written in two equivalent ways f x | θ = e x / θ / θ = λ e λ x = f x | λ .
When we have the observed sample D={x1, x2, …, xn}, our general problem is to estimate the value of the parameters of the model (representing the parent population) from the information given by the sample. We define some criteria which we require a "good" estimate to satisfy and see whether there exist any "best" estimates. We assume that the parent population is distributed in a form, the model, which is completely determinate but for the value θ0 of some parameter, e.g. unidimensional, θ, or bidimensional θ={μ, σ2}, or θ={β,η,ω}) as in the GIW(x|β,η,ω), or θ={β,η,ω, θ }) as in the MPGW(x|β,η,ω, θ ) .
We seek some function of θ, say τ(θ), named inference function, and we see if we can find a RV T which can have the following properties: unbiasedness, sufficiency, efficiency. Statistical Theory allows us the analysis of these properties of the estimators (RVs).
We use the symbols X ¯ and S 2 for the unbiased estimators T1 and T2 of the mean and the variance.
Luckily, we have that T1, in the Exponential model  f x | θ , is efficient [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33], and it extracts the total available information from any random sample, while the couple T1 and T2, in the Normal model, are jointly sufficient statistics for the inference function τ(θ)=(μ, σ2), so extracting the maximum possible of the total available information from any random sample. The estimators (which are RVs) have their own “distribution” depending on the parent model F(x|θ) and on the sample D: we use the symbol φ t ,   θ , n for that “distribution”. It is used to assess their properties. For a given (collected) sample D the estimator provides a value t (real number) named the estimate of τ(θ), unidimensional.
As said before, a way of finding the estimate is to compute the Likelihood Function  L θ D [LF] and to maximise it: the solution of the equation L θ D / θ =0 is termed Maximum Likelihood Estimate [MLE]. Both are used for sequential tests.
The LF is important because it allows us finding the MVB (Minimum Variance Bound, Cramer-Rao theorem) [1,2,6,7,8,9,10,11,12,13,14,15,16,26,27,28,29,30,31,32,33,34,35,36] of an unbiased RV T [related to the inference function τ(θ)], such that
V a r T τ θ 2 E l n L θ D θ 2 = M V B T
The inverse of the MVB(T) provides a measure of the total available amount of information in D, relevant to the inference function τ(θ) and to the statistical model F(x|θ).
Naming IT(T) the information extracted by the RV T we have that [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,26,27,28,29,30,31,32,33,34,35,36]
IT(T)=1/MVB(T) T is an Efficient Estimator.
If T is an Efficient Estimator there is no better estimator able to extract more information from D.
The estimates considered before were “point estimates” with their properties, looking for the “best” single value of the inference function τ(θ).
We recap the very important concept of Confidence Interval (CI) and Confidence Level (CL) [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,26,27,28,29,30,31,32,33,34,35,36].
The “interval estimates” comprise all the values between τL (Lower confidence limit) and τU (Upper confidence limit); the CI is defined by the numerical interval CI={τL-----τU}, where τL and τU are two quantities computed from the observed sample D: when we make the statement that τ(θ)∈CI, we accept, before any computation, that, doing that, we can be right, in a long run of applications, (1-α)%=CL of the applications, BUT we cannot know IF we are right in the single application (CL=Confidence Level).
We know, before any computation, that we can be wrong α% of the times but we do not know when it happens.
The reader must be very careful to distinguish between the Probability Interval PI={L-----U}, where the endpoints L and U depends on the distribution φ t ,   θ , n of the estimator T (that we decide to use, which does not depend on the “observed sample” D) and, on the probability π=1-α (that we fix before any computation), as follows by the probabilistic statement (9) [se the figure 2 for the exponential density, when n=1]
P L T U = L U φ t ,   θ , n d t = 1 α
and the Confidence Interval CI={τL-----τU} which depends on the “observed sample” D.
Notice that the Probability Interval PI={L-----U}, given in the formula (9), does not depend on the data D, as you can pictorially see in fig. 2: L and U are the Probability Limits. Notice that, on the contrary, the Confidence Interval CI={τL-----τU} does depend on the data D, pictorially seen in fig. 2. This point is essential for all the papers in the References.
Shewhart identified this approach, L and U, on page 275 of [19] where he states:
Preprints 162230 i004
The Tchebycheff Inequality: IF the RV X is arbitrary with density f(x) and finite variance σ 2 THEN we have the probability P X μ k σ 1 / k 2 , where μ = E X . This is a “Probabilistic Theorem”.
It can be transferred into Statistics. Let’s suppose that we want to determine experimentally the unknown mean μ within a “stated error ε”. From the above (Probabilistic) Inequality we have P μ ε < X < μ + ε   1 σ 2 / ε 2 ; IF σ ε THEN the event X μ < ε is “very probable” in an experiment: this means that the observed value x of the RV X can be written as μ ε < x < μ + ε and hence x ε < μ < x + ε . In other words, using x as an estimate of μ we commit an error that “most likely” does not exceed ε . IF, on the contrary, σ ε , we need n data in order to write P μ ε < X ¯ < μ + ε   1 σ 2 / ( n ε 2 ) , where X ¯ is the RV “mean”; hence we can derive x ¯ ε < μ < x ¯ + ε ., where x ¯ is the “empirical mean” computed from the data. In other words, using x ¯ as an estimate of μ we commit an error that “most likely” does not exceed ε . See the excerpts 3, 3a, 3b.
Notice that, when we write x ¯ ε < μ < x ¯ + ε , we consider the Confidence Interval CI [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33], and no longer the Probability Interval PI [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33].
These statistical concepts are very important for our purpose when we consider the Sequential tests and the Control Charts, especially with Individual data.
Notice that the error made by several authors [4,5,24] is generated by lack of knowledge of the difference between PI and CI [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33]: they think wrongly that CI=PI, a diffused disease [4,5,24]! They should study some of the books/papers [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33] and remember the Deming statements (excerpt 2).
The Deming statements are important for Quality. Managers, scholars; the professors must learn Logic, Design of Experiments and Statistical Thinking to draw good decisions. The authors must, as well. Quality must be their number one objective: they must learn Quality methods as well, using Intellectual Honesty [1,2,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33]. Using (9), those authors do not extract the maximum information from the data in the Process Control. To extract the maximum information from the data one needs statistical valid Methods [1,2,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33].
As you can find in any good book or paper [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33] there is a strict relationship between CI and Test Of Hypothesis, known also as Null Hypothesis Significance Testing Procedure (NHSTP). In Hypothesis Testing (see the Introduction), the experimenter wants to assess if a “thought” value of a parameter of a distribution is confirmed (or rejected) by the collected data: for example, for the mean μ (parameter) of the Normal  n (x| μ , σ 2 ) density, he sets the “null hypothesis” H0={μ=μ0} and the probability P=α of being wrong if he decides that the “null hypothesis” H0 is true, when actually it is opposite: H0 is wrong. When we analyse, at once, the observed sample D={x1, x2, …, xn} and we compute the empirical (observed) mean x ¯ and the empirical (observed) standard deviation s   , we define the Acceptance interval, which is the CI
L C L = x ¯ t 1 α / 2 s / n < μ < x ¯ + t 1 α / 2 s / n = U C L
Notice that the interval (for the Normal model, μ assumed)
μ t 1 α / 2 σ / n μ t 1 α / 2 σ / n
is the Probability Interval such that P μ t 1 α / 2 σ / n < X ¯ < μ t 1 α / 2 σ / n = 1 α .
A fundamental reflection is in order: the formulae (10) and (11) tempt the unwise guy to think that he can get the Acceptance interval, which is the CI [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23], by substituting the assumed values μ 0 , σ 0 of the parameters with the empirical (observed) mean x ¯ and standard deviation s . This trick is valid only for the Normal distribution.
The formulae (10) can be used sequentially to test H0={μ=μ0} versus H1={μ=μ10}; for any value 2<k≤n; we obtain n-2 CIs, decreasing in length; we can continue until either μ1<LCL or UCL<μ0, or both (verify) μ1<LCL and UCL<μ0.
More ideas about these points can be found in [34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
In the field of Control Charts, with Shewhart, instead of the formula (10), we use (12)
x ¯ z 1 α / 2 s c 4 n < μ < x ¯ + z 1 α / 2 s c 4 n
where the t distribution value t 1 α / 2 is replaced by the value z 1 α / 2 of the Normal distribution, actually z 1 α / 2 =3, and a coefficient c 4 is used to make “unbiased” the estimate of the standard deviation, computed from the information given by the sample.
Actually, Shewhart does not use the coefficient c 4 is as you can see from page 294 of Shewhart book (1931), where X ¯ is the “Grand Mean”, computed from D [named here empirical (observed) mean x ¯ ], σ is “estimated standard of each sample” (named here s, with sample size n=20, in excerpt 3)
Excerpt 3. From Shewhart book (1931), on page 294
Excerpt 3. From Shewhart book (1931), on page 294
Preprints 162230 sch003

2.2. Control Charts, as Sequential Testing, for Process Management

Statistical Process Management (SPM) entails Statistical Theory and tools used for monitoring any type of processes, industrial or not. The Control Charts (CCs) are the tool used for monitoring a process, to assess its two states: the first, when the process, named IC (In Control), operates under the common causes of variation (variation is always naturally present in any phenomenon) and the second, named OOC (Out Of Control), when the process operates under some assignable causes of variation. The CCs, using the observed data, allow us to decide if the process is IC or OOC. CCs are a statistical test of hypothesis for the process null hypothesis H0={IC} versus the alternative hypothesis H1={OOC}. Control Charts were very considered by Deming [9,10] and Juran [12] after Shewhart invention [19,20].
We start with Shewhart ideas (see the excerpts 3, 3a and 3b).
In the excerpts, X ¯ is the (experimental) “Grand Mean”, computed from D (we, on the contrary, use the symbol x ¯ ), σ is the (experimental) “estimated standard of each sample” (we, on the contrary, use the symbol s, with sample size n=20, in excerpts 3a, 3b), σ ¯ is the “estimated mean standard deviation of all the samples” (we, on the contrary, use the symbol s ¯ ).
Excerpt 3a. From Shewhart book (1931), on page 89
Excerpt 3a. From Shewhart book (1931), on page 89
Preprints 162230 sch013
Excerpt 3b. From Shewhart book (1931), on page 294
Excerpt 3b. From Shewhart book (1931), on page 294
Preprints 162230 sch014
So, we clearly see that Shewhart, the inventor of the CCs, used the data to compute the Control Limits, LCL (Lower Control Limit, which is the Lower Confidence Limit) and UCL (Upper Control Limit, the Upper Confidence Limit) both for the mean μ X (1st parameter of the Normal pdf) and for σ X (2nd parameter of the Normal pdf). They are considered the limits comprising 0.9973n of the observed data. Similar ideas can be found in [5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42] (with Rozanov, 1975, we see the idea that CCs can be viewed as a Stochastic Process).
We invite the readers to consider that if one assumes that the process is In Control (IC) and if he knows the parameters of the distribution he can test if the assumed known values of the parameters are confirmed or disproved by the data, then he does not need the Shewhart Control Charts; it is sufficient to use NHSTP or the Sequential Test Theory!
Remember the ideas in the previous section and compare Excerpts 3, 3a, 3b (where LCL, UCL depend on the data) with the following Excerpt 4 (where LCL, UCL depend on the Random Variables) and appreciate the profound “logic” difference: this is the cause of the many errors in the CCs for TBE [Time Between Events (see [4,5,24]).
Preprints 162230 i001
The same type of arguments are used in another paper [4] JQT, 2017 where the data are Erlang distributed with λ0 is the scale parameter and the Control Limits LCL and UCL are defined [copying Xie et al.] erroneously as
Excerpt 4. From a paper in the “Garden… [24]”. Notice that one of the authors wrote several papers…
Excerpt 4. From a paper in the “Garden… [24]”. Notice that one of the authors wrote several papers…
Preprints 162230 sch004
The formulae, in the excerpt 4, LCL1 and UCL1 are actually the Probability Limits (L and U) of the Probability Interval PI in the formula (9), when φ t ,   θ , n is the pdf of the Estimator T, related to the Normal model F(x; μ, σ2). Using (9), those authors do not extract the maximum information from the data in the Process Control. From the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36] we derive that the interval L=μY-3σY------μY+3σY=U is the PI such that the RV Y= X ¯
P [ μ Y 3 σ Y Y = X ¯ μ Y + 3 σ Y ] = 0.9973
and it is not the CI of the mean μ=μY [as wrongly said in the Excerpt 4, where actually (LCL1-----UCL1)=PI].
The same error is in other books and papers (not shown here but the reader can see in [21,22,23,24]).
The data plotted in the CCs [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33,34,35,36] (see the fig. 3) are the means x ¯ t i , determinations of the RVs X ¯ t i , i=1, 2, ..., n (n=number of the samples) computed from the sequentially collected data of the i-th sample Di={xij, j=1, 2, ..., k} (k=sample size)}, determinations of the RVs X t i j at very close instants tij, j=1, 2, ..., k. In other applications I-CC (see the fig. 3), the data plotted are the Individual Data x t i , determinations of the Individual Random Variables X t i , i=1, 2, ..., n (n=number of the collected data), modelling the measurement process (MP) of the “Quality Characteristic” of the product: this model is very general because it is able to consider every distribution of the Random Process X t , as we can see in the next section. From the excerpts 3, 3a, 3b and formula (10) it is clear that Shewhart was using the Normal distribution, as a consequence of the Central Limit Theorem (CLT) [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,26,27,28,29,30,31,32,33,34,35,36]. In fact, he wrote on page 289 of his book (1931) “… we saw that, no matter what the nature of the distribution function of the quality is, the distribution of the arithmetic mean approaches normality rapidly with increase in n (his n is our k), and in all cases the expected value of means of samples of n (our k) is the same as the expected value of the universe” (CLT in Excerpt 3, 3a, 3b).
Figure 3. Control Limits LCLX----UCLX=L----U (Probability interval), for Normal data (Individuals xij, sample size k) “sample means” x ¯ I and “grand mean” x ̿ .
Figure 3. Control Limits LCLX----UCLX=L----U (Probability interval), for Normal data (Individuals xij, sample size k) “sample means” x ¯ I and “grand mean” x ̿ .
Preprints 162230 g030
Figure 4. Individual Control Chart (sample size k=1). Control Limits LCL----UCL=L----U (Probability interval), for Normal data (Individuals xi) and “grand mean” x ¯ .
Figure 4. Individual Control Chart (sample size k=1). Control Limits LCL----UCL=L----U (Probability interval), for Normal data (Individuals xi) and “grand mean” x ¯ .
Preprints 162230 g004
Let k be the sample size; the RVs X ¯ t i are assumed to follow a normal distribution and uncorrelated; X ¯ t i [ith rational subgroup] is the mean of RVs IID X t i j j=1, 2, ..., k, (k data sampled, at very near times tij).
To show our way of dealing with CCs we consider the process as a “stand-by system whose transition times from a state to the subsequent one” are the collected data. The lifetime of “stand-by system” is the sum of the lifetimes of each unit. The process (modelled by a “stand-by …”) behaves as a Stochastic Process X t [25,26,27,28,29,30,31,32,33], that we can manage by the Reliability Integral Theory (RIT): see the next section; this method is very general because it is able to consider every distribution of X t .
If we assume that X t is distributed as f(x) [probability density function (pdf) of “transitions from a state to the subsequent state” of a stand-by subsystem] the pdf of the (RV) mean X ¯ t i is, due the CLT (page 289 of 1931 Shewhart book), X ¯ t i ~ N μ X ¯ t i ,   σ X ¯ t i 2 [experimental mean x ¯ t i ] with mean μ X ¯ t i and variance σ X ¯ t i 2 . X ̿ is the “grand” mean and σ X ̿ 2 is the “grand” variance: the pdf of the (RV) grand mean X ̿ ~ N μ X ̿ ,   σ X ̿ 2 [experimental “grand” mean x ̿ ]. In fig. 2 we show the determinations of the RVs X ¯ t i and of X ̿ .
When the process is Out Of Control (OOC, assignable causes of variation, some of the means μ X ¯ t i , estimated by the experimental means x ¯ i = x ¯ t i , are “statistically different)” from the others [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33,34,35,36]. We can assess the OOC state of the process via the Confidence Intervals (provided by the Control Limits) with CL=0.9973. Remember the trick valid only for the Normal Distribution ….; consider the PI, L=μY-3σY------μY+3σY=U; putting x ̿ in place of μ Y and s ¯ / k in place of σ Y we get the CI of μ X ̿ when the sample size k is considered for each X ¯ t i , with CL=0.9973. The quantity s ¯ is the mean of the standard deviations of each sample. This allows us to compare each (subsystem) mean μ X ¯ t q , q=1,2, …, n, to any other (subsystem) mean μ X ¯ t r     r=1,2, …, n, and to the (Stand-by system) grand mean μ X ̿ = μ . If two of them are different, the process is classified as OOC. The quantities L C L X = x ¯ ¯ 3 s ¯ / k and U C L X = x ¯ ¯ + 3 s ¯ / k are the Control Limits of the CC, which are the Confidence Limits. When the Ranges Ri=max(xij)-min(xij) are considered for each sample we have L C L X = x ¯ ¯ A 2 R ¯ , U C L X = x ¯ ¯ + A 2 R ¯ and L C L R = D 3 R ¯ , U C L R = D 4 R ¯ , where R ¯ is the “mean range” and the coefficients A2, D3, D4 are tabulated and depend on the sample size k [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,25,26,27,28,29,30,31,32,33,34,35,36].
We stress that the interval LCLX-------UCLX is the “Confidence Interval” with “Confidence Level” CL=1-α=0.9973 for the unknown mean μ X t of the Stochastic Process X(t) [25,26,27,28,29,30,31,32,33,34,35,36]. The interval LCLR----------UCLR is the “Confidence Interval” with “Confidence Level” CL=1-α=0.9973 for the unknown Range of the Stochastic Process X(t) [25,26,27,28,29,30,31,32,33,34,35,36].
Notice that, ONLY for normally distributed data, the length of the Control Interval (UCLX-LCLX, which is the Confidence Interval) equals the length of the Probability Interval, PI (U-L): UCLX-LCLX=U-L.
The error highlighted, i.e. the confusion between the Probability Interval and the Control Limits (Confidence Interval!) has no consequences for decisions when the data are Normally distributed, as considered by Shewhart. On the contrary, it has BIG consequences for decisions WHEN the data are Non-Normally distributed [4,5,24].
We think that the paper “Quality of Methods for Quality is important”, [1] appreciated and mentioned by J. Juran at the plenary session of the EOQC (European Organization for Quality Control) Conference (1989), should be considered and meditated.

2.3. Statistics and Reliability Integral Theory (RIT)

We are going to present the fundamental concepts about RIT (Reliability Integral Theory) that we use for computing the Control Limits (Confidence Limits) of CCs. RIT is the natural way for Sequential Tests, because the transitions happen sequentially; to provide the ideas, we use a “4 units Stand-by system”, depicted by 5 states (Figure 5): 0 is the state with all units not-failed; 1 is the state with the first unit failed; 2 is the state with the second unit failed; and so on, until the system enters the state 5 where all the 4 units are failed (down state, in yellow): any transition provides a datum to be used for the computations. RIT can be found in the author’s books…
RIT can be used for parameters estimation and Confidence Intervals (CI), (Galetto 1981, 1982, 1995, 2010, 2015, 2016), in particular for Control Charts (Deming, 1986, 1997, Shewhart 1931, 1936, Galetto 2004, 2006, 2015). In fact, any Statistical or Reliability Test can be depicted by an “Associated Stand-by System” [25,26,27,28,29,30,31,32,33,34,35,36] whose transitions are ruled by the kernels bk,j(s); we write the fundamental system of integral equations for the reliability tests, whose duration t is related to interval 0-----t; the collected data tj can be viewed as the times of the various failures (of the units comprising the System) [t0=0 is the start of the test, t is the end of the test and g is the number of the data (4 in the Figure 5)]
Firstly, we assume that the kernel b j , j + 1 s t j is the pdf of the exponential distribution f ( s t j | μ , σ 2 ) = λ e λ s t j , where λ is the failure rate of each unit and λ = 1 / θ : θ is the MTTF of each unit. We state that R j t t j is the probability that the stand-by system does not enter the state g (5 in Figure 5), at time t, when it starts in the state j (0, 1, …, 4) at time tj, W ¯ j t t j is the probability that the system does not leave the state j, b j , j + 1 s t j d s is the probability that the system makes the transition j→j+1, in the interval s-----s+ds.
The system reliability R 0 t is the solution of the mathematical system of the Integral Equations (13)
R j t t j = W ¯ j t t j + t j t b j , j + 1 t t j R j + 1 t s d s   f o r j = 0 , 1 , , , g 1 , R g t | t g = W ¯ g t t g
With λ e λ s t j we obtain the solution (see Figure 5, putting the Mean Time To Failure MTTF of each unit=θ, λ = 1 / θ ) (see the Figure 6)
R 0 t = e λ t 1 + λ t + λ t 2 2 ! + λ t 3 3 ! + λ t 4 4 !
The reliability system (13) can be written in matrix form,
R t r = W ¯ t r + r t B s r R s d s
At the end of the reliability test, at time t, we know the data (the times of the transitions tj) and the “observedempirical sample D={x1, x2, …, xg}, where xj=tj – tj-1 is the length between the transitions; the transition instants are tj = tj-1 + xj giving the “observedtransition sample D*={t1, t2, …, tg-1, tg, t=end of the test} (times of the transitions tj).
We consider now that we want to estimate the unknown MTTF=θ=1/λ of each item comprising the “associated” stand-by system [24,25,26,27,28,29,30]: each datum is a measurement from the exponential pdf; we compute the determinant det B s r ;   θ ,   D * = 1 / θ g exp T t of the integral system (14), where T t is the “Total Time on Test” T t = 1 g x i   [ t 0 in the figure 5]: the “Associated Stand-by System” [25,26,27,28,29,30,31,32,33] in the Statistics books provides the pdf of the sum of the RV Xi of the “observedempirical sample D={x1, x2, …, xg}. At the end time t of the test, the integral equations, constrained by the constraintD*, provide the equation
l n d e t B s r ;   θ ,   D * / θ = θ / g T t = 0
It is important to notice that, in the case of exponential distribution [11,12,13,14,15,16,25,26,27,28,29,30,31,32,33,34,35,36], it is exactly the same result as the one provided by the MLM Maximum Likelihood Method.
If the kernel b j , j + 1 s t j is the pdf f ( s t j | μ , σ 2 ) = 1 / 2 π σ e s t j μ 2 / ( 2 σ 2 ) the data are normally distributed, X ~ N μ X , σ X 2 = 1 / 2 π σ X e x μ X 2 / ( 2 σ X 2 ) , with sample size n, then we get the usual estimator X ¯ = X i / n such that E X ¯ = μ X .
The same happens with any other distribution (e.g. see the Table 2) provided that we write the kernel b i , i + 1 s .
The reliability function R 0 t | θ , [formula (13)], with the parameter θ , of the “Associated Stand-by System” provides the Operating Characteristic Curve (OC Curve, reliability of the system) [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36] and allows to find the Confidence Limits ( θ L Lower and θ U   Upper) of the “unknown” mean θ , to be estimated, for any type of distribution (Exponential, Weibull, Rayleigh, Normal, Gamma, Inverted Weibull, General Inverted Weibull, …); by solving, with (a general) unknown (indicated as) θ , the two equations R 0 ( t 0 | θ ) = 1 α / 2       a n d         R 0 t 0 θ = α / 2 ; we get the two values ( θ L , θ U ) such that
R 0 t o | θ L = α / 2     a n d         R 0 t o | θ U = 1 α / 2
where t o is the (computed) “total of the length of the transitions xi=tj - tj-1 data of the empirical sample D” and CL= 1 α is the Confidence Level. CI= θ L -------- θ U is the Confidence Interval: θ L = 1 / λ U and θ U = 1 / λ L .
For example, with figure 6, we can derive θ L = 62.5   d a y s = 1 / λ U and θ U = 200   d a y s = 1 / λ L , with CL=0.8. It is quite interesting that the book [14] Meeker et al., “Statistical Intervals: A Guide for Practitioners and Researchers”, John Wiley & Sons (2017) use the same ideas of FG (shown in the formula 16) for computing the CI; the only difference is that the author FG defined the procedure in 1982 [26], 35 years before Meeker et al.
As said before, we can use RIT for the Sequential Tests; we have only to consider the various transitions and the Total Time on Test to the last transition we want to consider.

2.4. Control Charts for TBE Data. Some Ideas for Phase I Analysis

Let’s consider now TBE (Time Between Event, time between transitions) data, exponentially or Weibull distributed. Quite a lot of authors (in the “Garden … [24]”) compute wrongly the Control Limits (which are the Confidence Limits) of these CCs.
The formulae, shown in the section “Control Charts for Process Management”, are based on the Normal distribution (thanks to the CLT; see the excerpts 3, 3a and 3b); unfortunately, they are used also for NON_normal data (e.g. see formulae (6)): for that, sometimes, the NON_normal data are transformed “with suitable transformations” in order to “produce Normal data” and to apply those formulae (above) [e.g. Montgomery in his book].
Sometimes we have few data and then we use the so called “Individual Control Charts” I-CC. The I-CCs are very much used for exponentially (or Weibull) distributed data: they are also named “rare events Control Charts for TBE (Time Between Events) data”, I-CC_TBE.
In the previous section, we computed the CI= θ L -------- θ U of the parameter θ , using the (subsample) “transition times durations”: t O =“total of the transition times durations (length of the transitions xi=tj - tj-1 data) in the empirical sample (subsample with n=4 only, as an example)” and Confidence Level CL= 1 α .
When we deal with a I-CC_TBE we compute the LCL and UCL of the mean θ through the empirical mean  t ¯ O = t O / n of each transition, for the… ; we solve the two following equations (17) for the two unknown values LCL and UCL, for R t ¯ O   θ   of each item in the sample, similar to (16)
R t ¯ O | L C L = α / 2 ,                           R t ¯ O | U C L = 1 α / 2
where now t ¯ O = t O /n is the “mean, to be attributed, to the single lengths of the single transitions xi=tj-tj-1 data in the empirical sample D with the Confidence Level CL= 1 α : L C L = 1 / λ U and U C L = 1 / λ L .
In the next sections we can see the Scientific Results found by a Scientific Theory (we anticipate them: the Control Limits are LCL=18.0 days and UCL=88039.3 days).

3. Results

In this section we provide the scientific analysis of the “remission time” data [3] and compare our result with those of the authors: the findings are completely different and the decisions, consequently, should be different, with different costs of wrong decisions.

3.1. Control Charts for TBE Data. Phase I Analysis

The “remission time of 128 bladder cancer patients” data are in the Table 3.
Table 3. Data “of remission time of 128 bladder cancer patients” from “Statistical Inference on the Shape Parameter of Inverse Generalized Weibull Distribution”, Mathematics (2024) [3] and Modified generalized Weibull distribution: theory and applications, Scientific Reports (2023):12828.
Table 3. Data “of remission time of 128 bladder cancer patients” from “Statistical Inference on the Shape Parameter of Inverse Generalized Weibull Distribution”, Mathematics (2024) [3] and Modified generalized Weibull distribution: theory and applications, Scientific Reports (2023):12828.
Preprints 162230 i003
Using all the 128 Cancer data the authors [3] write:
Excerpt 5. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Excerpt 5. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Preprints 162230 sch005
They add also:
Excerpt 6. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Excerpt 6. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Preprints 162230 sch006
So, the authors decided to “assume” (use) the GIW(x|β, η, ω): [ 1 e η / x β ] ω to analyse all the 128 data in table 3; their estimates are in Excerpt 5 and Excerpt 6. Looking at Q-Q Plot and Histogram (in Excerpt 7) the readers can have some doubts about the use. of GIW.
=61.38, β ^ =0.51, η ^ =8.19), from [3]
Excerpt 7. QQ plot of remission time of 128 bladder cancer patients with IGW histogram of the real data and probability density GIW (estimates of the parameters ω ^
Excerpt 7. QQ plot of remission time of 128 bladder cancer patients with IGW histogram of the real data and probability density GIW (estimates of the parameters ω ^
Preprints 162230 sch007
As a matter of fact, we can draw the Figure 7, TTOT (of the data xi) versus i/n (n=128); from the graph it is evident that the exponential distribution is suitable for the data analysis. Therefore, we will compare the models Exponential, Inverted W and GIW.
We divide the data in two sets: the first based on the first 32 data and the second considering the others.
Fitting the Weibull distribution, one finds β=1.17 and η=8.88, with -2lnL=198.58; since the 1∈CI of β, with CL=80% we are allowed to use the exponential distribution (as given in Figure 7). Seeing the Figure 8 we find that the data show an OOC.
The figure 8 shows that the first 32 data do not allow to assess if the “null hypothesis” H0={θ=10}, with α=0.025 is to be accepted or rejected in favour of the H1={θ=5.75} with β=0.025. The Wald Sequential Test is inefficient for the 32 data. Compare with Figure 9.
The last CI={6.08, 12.23}, Figure 10, shows that the first 32 data allows to assess that the “null hypothesis” H0={θ=10}, with α=0.025 is to be accepted but H1={θ=5.75} rejected with β>0.025: {5.75<6.08<10<12.23}. The Sequential CIs are not more efficient than Wald Test.
Figure 10. Sequential Confidence Intervals (α=β=0.025) for the Exponential distribution.
Figure 10. Sequential Confidence Intervals (α=β=0.025) for the Exponential distribution.
Preprints 162230 g010
Figure 11. Sequential Confidence Intervals (α=β=0.025) for the Inverse Weibull distribution.
Figure 11. Sequential Confidence Intervals (α=β=0.025) for the Inverse Weibull distribution.
Preprints 162230 g011
Fitting the Inverse Weibull distribution, on the first 32 data (1/xi), one finds β=1.0422 and η=0.281, with -2lnL=18.59; since the 1∈CI of β, with CL=80% we are allowed to use the exponential distribution, as we could do for the data xi. The conclusion, the first 32 data (1/xi), about the sequential CIs would be the same as for the data xi.
In the next session we consider all the 128 data and compare our results with the authors of [3].
The authors of the paper “Modified generalized Weibull distribution: theory and applications, Scientific Reports (2023):12828” use the Modified Power Generalised Weibull model MPGW(x|β,η,ω, θ ) =MPGW(x| 3.5251 , 2.905 , 0.1153 ,   19.231 ) , but they do not compute the Confidence Intervals; so, we cannot see if ω=0 is an acceptable estimation.
For exponentially distributed data (17) becomes (18) [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33], k=1, with CL= 1 α
e t ¯ O / L C L = 1 α / 2   and   e t ¯ O / U C L = α / 2
The endpoints of the CI= L C L -------- U C L are the Control Limits of the I-CC_TBE.
This is the right method to extract the “true” complete information contained in the sample (see the figs. 8, 9, 10). The figures are justified by the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33] and are related to the formulae [(12), (13) for k=1], for the I-CC_TBE charts.
Remember the book Meeker et al., “Statistical Intervals: A Guide for Practitioners and Researchers”, John Wiley & Sons (2017): the authors use the same ideas of FG; the only difference is that FG invented 30 years before, at least.
Compare the formulae [(18), for k=1], theoretically derived with a sound Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33], with the ones in the Excerpt [in the Appendix (a small sample from the “Garden … [24]”)] and notice that the two Minitab authors (Santiago&Smith) use the “empirical mean t ¯ O ” in place of the θ 0 in the figure 1: it is the same trick of replacing  x ̿ to the mean μ which is valid for the Normal distributed data only; e.g., see the formulae (1)!

3.2. Control Charts for TBE Data. Phase II Analysis

We saw in the previous section what usually it is done during the Phase I of the application of CCs: estimation of the mean and standard deviation; later, their values are assumed as “true known” parameters of the data distribution, in view of the Phase II.
We considered the first 32 (out of 128 remission times of bladder) Cancer data; using all the 128 data the authors found (Excerpts 5, 6) the distribution GIW(x|β, η, ω): 1 [ 1 e η / x β ] ω with estimated parameters β ^ =0.51, η ^ =8.19, ω ^ =61.38; on the contrary we found that the exponential distribution (after fitting the Weibull and the Inverse Weibull) was suitable: that allowed us to make many considerations about the use of sequential sampling.
Now we consider all the 128 data and see new considerations.
In particular, for TBE individual data the exponential distribution is assumed with a known parameter λ0 or θ0.
We consider now what it is done during the Phase II of the application of CCs for TBE data individual exponentially distributed.
As previously we find that the Exponential distribution is well fitting the data, Figure 12, opposite to the distribution GIW(x|β, η, ω): 1 [ 1 e η / x β ] ω .
The last CI={9.36, 11.22}, Figure 13, shows that the 128 data allows to assess that the “null hypothesis” H0={θ=10}, with α=0.025 is to be accepted but H1={θ=5.75} rejected with β>0.025: {5.75<9.36<10<11.22}. The Sequential CIs are less efficient than Wald Test.
Figure 13. Sequential Confidence Intervals (α=β=0.025) of the 128 Cancer data with hypotheses H0={θ=10} versus H1={θ=5.75}. Exponential distribution is suitable.
Figure 13. Sequential Confidence Intervals (α=β=0.025) of the 128 Cancer data with hypotheses H0={θ=10} versus H1={θ=5.75}. Exponential distribution is suitable.
Preprints 162230 g013
Figure 14. Sequential Test (Wald) of the 128 Cancer data; the decision to Accept H0={θ=10} happens at the 42nd point.
Figure 14. Sequential Test (Wald) of the 128 Cancer data; the decision to Accept H0={θ=10} happens at the 42nd point.
Preprints 162230 g014
Figure 14. Control Chart of the 128 Cancer data: the process is OOC.
Figure 14. Control Chart of the 128 Cancer data: the process is OOC.
Preprints 162230 g015
As it happened previously, we find that the CC provides much more information to the Manager to allow him to take sound decisions.
Since the CCs are “sequential tests”, as are the reliability flow graphs, we think that it is wise to use them.

3.3. Sequential Test by the Authors of [3]

Now we see what the authors of [3] did about their distribution GIW(x|β, η, ω): 1 [ 1 e η / x β ] ω (in the paper they use α in place of our ω; we introduced ω because α is the type I probability risk, associated to H0).
They found the MLE (Maximum Likelihood Estimate ω ^ n and Estimator Ω n ) of the parameter ω, with n the number of the data considered, H and B estimators of η and β
ω ^ n = n / 1 n l n 1 e η / x i β   and   Ω n = n / 1 n l n 1 e H / x i B = n / Y
and computed ω ^ 128 = 61.38 ; from that they computed the 95% Confidence Interval as (59.82, 63.07), which is defined as ω ^ n / d ,     d ω ^ n , where d is the “accuracy” of the CI. They proved a very interesting result: the distribution GIW(x|β, η, ω) depends only on n, the number of data considered, and not on the parameters of GIW(x|β, η, ω) of the Random Variable T = ω Y / n ~ G a m m a ( n , 1 / n ) .
Notice that the Confidence Interval (59.82, 63.07) is actually a Probability Interval, showing the same error mentioned in [24].
The authors state (in Excerpt 8)
Excerpt 8. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Excerpt 8. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Preprints 162230 sch008
We tried to draw a “TTOT (Total Time On Test transform) of the 128 Cancer data” (similar to Figure 12) with GIW(x|0.51, 8.19, 61.38): 1 [ 1 e 8.19 / x 0.51 ] 61.38 ; it is impossible to draw such a graph with data in table 3. To understand the reader can see the Figure 16:
It is evident from table 3 that only the smallest nine data 0.08, 0.20, 0.40, 0.50, 0.51 0.81, 0.90, 1.05, 1,19 could be shown in the figure 16; the other 119 data are all near the ordinate 1 (in the figure 16).
How could GIW(x|0.51, 8.19, 61.38) fit suitably the 128 Cancer data?
So, the Excerpt 8 is quite doubtful.
To understand we created the Figure 17 where we have shown, versus the number of data, the Inverse data (of those in Table 3), the sum of the inverse of the collected data named “Tot_inverse” and the “Tot_inverse_B 1 n l n 1 e η / x i β from the Distribution GIW(x|0.51, 8.19, 61.38), with their interpolating formulae (where x “actually” is the number of counts, 1, 2, 3 ,…, n-2, n-1, n).
We see that the “Tot_inverse_B”  1 n l n 1 e η / x i β , from the Distribution GIW(x|0.51, 8.19, 61.38), does not fit well the successive sums of the Inverse data (of those in Table 3).
What is the consequence? We leave it to the readers…
Computing the quantity “Tot_inverse_B”= 1 n l n 1 e 8.19 / x i 0.51 = 1.99 26 , we find the estimate ω ^ 128 _ F G = 64.2391 , that is different from the estimate of the authors ω ^ 128 = 61.38 ; so getting the Distribution GIW(x|0.51, 8.19, 61.38). Their 95% Confidence Interval, C I = ω ^ n / d ,     d ω ^ n , where d is the “accuracy” of the CI, was (59.82, 63.07): notice that the “named” CI_Zhuang=(59.82, 63.07) is actually a Probability Interval, showing the same error mentioned in [24].
It is important to notice that ω ^ 128 _ F G > 63.07 (the upper limit of the “named” CI_Zhuang=(59.82, 63.07). We leave to the readers to say what that means!
Let’s indicate as T the Random Variable Ω 1 g l n 1 e η / x i β / g ; we have that T   ~ G a m m a ( g , 1 / g ) , with density f ( t ; β , θ ) = t β 1 e t / θ / θ β Γ ( β ) where θ is the scale parameter and β is the shape parameter ( β = g = 1 / θ ).
We can write the Probability statement, for any value chosen g,
P G L = L < T = Ω 1 g l n 1 e η / x i β / g < U = G U = 1 α
where L------U is the interval that comprises the RV T with probability 1-α and G is the Cumulative Gamma Distribution.
From (15) can derive the “equivalent” Probability statement, for any value chosen g,
P G L / Ω = L / Ω < 1 g l n 1 e η / x i β / g < U / Ω = G U / Ω = 1 α
where L / Ω ------U / Ω is the random interval that comprises the parameter ω with probability 1-α.
After the estimation of ω ^ n we have the Confidence Interval, C I = G L / ω ^ n ,     G U / ω ^ n surely different from the “named” CI=(59.82, 63.07).
By taking advantage of the fact [3] that Y = 1 n l n 1 e H / x i B , follows a Gamma distribution with parameters n and 1/ω, we can compute directly the CI by computing the OC Curve (Operating Characteristic Curve) OC(%)=1 - Gamma(1.956|128, 1/ω); we show it in Figure 18:
It is clear that the intersections of the OC Curve (fig. 18) with the two horizontal lines y=0.025 and y=0.975, provide the limits of the CI, which are different from the “named” CI_Zhuang=(59.82, 63.07).
The Confidence limits are the values of the “unknown” variable ω satisfying next two equations, with D= Tot_inverse_B, computed with all the 128 data
0 D y 127 ω 128 e ω y d y / Γ ( 128 ) = 1 α / 2   ,                     0 D y 127 ω 128 e ω y d y / Γ ( 128 ) = α / 2
Putting Dg= 1 g l n 1 e η ^ / x i β ^ computed with g data, we can get the successive Confidence Intervals CIg; two of them can be seen in the Figure 19 with their OC Curves, for g=128and g=73.
Our CI73 (figure 19) is CI73≈(83, 131); notice the big difference with the one given in the Excerpt 9. Notice that the value 61.38, estimated by all the 128 data, is named as “true value” (which is unknown).
Excerpt 9. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Excerpt 9. Zhuang et al., Statistical Inference on … Generalized Weibull Distribution. 2024
Preprints 162230 sch009
Notice that ω128=64.24 while ω73=104.07, quite a big difference with the “true value” 61.38, as in the Excerpt 9.
Obviously the CIs are different from the ones in [3].

3.4. Other Cases

Another case we want to consider is in the paper [5] by Alshahrani et al., “On Designing of Bayesian Shewhart-Type Control Charts for Maxwell Distributed Processes with Application of Boring Machine. Mathematics 2023, 11, 1126”
The authors say:
Excerpt 10. Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”
Excerpt 10. Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”
Preprints 162230 sch010
If X is a RV having the Maxwell distribution with scale parameter σ 2 , then its pdf is as follows: f ( x | σ 2 ) = 2 / π σ 3 x 2 e x 2 / ( 2 σ 2 )   w i t h   x ,   σ 2 > 0 .
The ML Estimator (which is a RV) of the parameter σ 2 is Σ 2 = 1 n X j 2 / ( 3 n ) ; the transformation T = X 2 / ( 2 σ 2 ) shows that T ~ G ( 3 / 2,1 ) , the Gamma distribution, and U = 3 n Σ 2 / ( 2 σ 2 ) ~ G ( 3 n / 2,1 ) . It is interesting to note that if x is interpreted as the velocity of a particle of unit mass (m=1), the quantity w = m x 2 / 2 is the energy of the particle and the two components X1 and X2 can be considered as independent RVs normally distributed with mean E[X]=0 and variance Var[X]= σ 2 .
The authors consider correctly the Probability Limits, But, unfortunately, they wrote “Practically, the parameter σ 2 may be known or unknown then the probability control limits of the control chart are defined as follows:….”.
Notice that in Control Charts (CCs) we use the Control Limits, LCL and UCL, NOT the “probability control limits of the control chart”!.
The authors made a lot of simulation and eventually applied their ideas to
Excerpt 11. Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”
Excerpt 11. Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”
Preprints 162230 sch011
Using JMP for the Individual Control Chart on x2 data we found Figure 20
This can be compared to Figure 21
Figure 21. ICC by Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”; you see the Probability Limits (LPL, UPL) ….
Figure 21. ICC by Alshahrani et al., “On Designing … of Boring Machine. Mathematics 2023”; you see the Probability Limits (LPL, UPL) ….
Preprints 162230 g021
To understand the difference between the Control Limits (LCL and UCL) and the Probability Limits (L and U) you have to analyse the Figure 22.
The application of the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58] to the Boring data (Excerpt 11) is in the Figure 23.
The “scientific” Control Chart for the Boring data (Excerpt 11) is in the Figure 24: the process is OOC, contrary to the findings in the Figure 20; the cause is the use of the Probability Limits (LPL, UPL) instead of the Control Limits (LCL, UCL).
Figure 24. The “scientific” Control Limits (LCL and UCL) of the Boring data, according to the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], with CL=0.9973.
Figure 24. The “scientific” Control Limits (LCL and UCL) of the Boring data, according to the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], with CL=0.9973.
Preprints 162230 g024
The Figure 25 shows the sequential Confidence Intervals for the Boring data (Excerpt 11); 13 data are necessary for a Confidence Interval=(2333350, 5733177)
The Figure 26 shows the Sequential (Wald) Test Confidence for the Boring data (Excerpt 11); one sees that at the 13rd datum the “step-line” G ( x i 2 ) , number of failures versus the total of squared times, intersects the Acceptance line; the competing Hypotheses are H 0 = 3160782   w i t h   α = 0.025 versus H 1 = 1359136   w i t h   β = 0.025 .
It is important to remember that the CI=(2427479, 4286847) is computed from all the data with CL=0.95.
All the results are found via RIT (Reliability Integral Theory) [25,26,27,28,29,30,31,32,33].

4. Discussion

We decided to use the data from the papers [3,4,5] and the analysis by the authors.
We got different results from those authors: the cause is that they use the Probability Limits of the PI (Probability Interval) as they were the Confidence Limits (Control Limits of the Control Charts).
The proof of the confusion between the intervals L-------U (Probability Interval) and LCL-------UCL (Confidence Interval) in the domain of Control Charts (for Process Management) highlight the importance and novelty of these ideas in the Statistical Theory and in the applications.
For the “location” parameter in the CCs, from the Theory, we know that two mean μ X ¯ t q (parameter), q=1,2, …, n, and any other mean μ X ¯ t r     (parameter), r=1,2, …, n, are different, with risk α, if their estimates are not both included in their common Confidence Interval as the CI of the grand mean μ X ̿ = μ (parameter) is.
Let’s consider the formula (4) and apply it to a “Normal model” (due to CLT, and assuming known variance), sequentially we can write the “real” fixed interval L----U comprising the RV X ̿ (vertical interval) and the Random Interval comprising the unknown mean μ (horizontal interval) (fig. 14)
P L = μ σ z 1 α 2 k X ̿ μ + σ z 1 α 2 k = U = P X ̿ σ z 1 α 2 k μ X ̿ + σ z 1 α 2 k
When the RV X ̿ assume its determination (numerical value) x ̿ (grand mean) the Random Interval becomes the Confidence Interval for the parameter μ, with CL=1-α: risk α that the horizontal line does not comprise the “mean” μ.
This is particularly important for the Individual Control Charts for Exponential, Weibull, Inverted Weibull, General Inverted Weibull, Maxwell and Gamma distributed data: this is what Deming calls “Profound Knowledge (understanding variation)” [9,10]. In this case, the Figures 22, 23, 27 look like the figure 2, where you see the Confidence Interval, the realisation of the horizontal Random Interval.
The case we considered shows clearly that the analyses, in the Process Management, taken so far have been wrong and the decisions have been misleading, when the collected data follow a Non-Normal distribution [24].
Since a lot of papers (related to Exponential, Weibull, Inverted Weibull, General Inverted Weibull, Maxwell and Gamma distributions), with the same problem as that of “The garden of flowers” [24], are published in reputed Journals we think that the “alternative” title “History is written by the winners. Reflections on Control Charts for Process Control” is suitable for this paper: the authors of the wrong papers [24] are the winners.
Figure 27. Probability Interval L---U (vertical line) versus Random Intervals comprising the “mean” μ (horizontal random variable lines), for Normally distributed RVs X ¯ ~ N μ , σ 2
Figure 27. Probability Interval L---U (vertical line) versus Random Intervals comprising the “mean” μ (horizontal random variable lines), for Normally distributed RVs X ¯ ~ N μ , σ 2
Preprints 162230 g027
Further studies should consider other distributions which cannot be transformed into the above distributions considered before.

5. Conclusions

With our figures (and the Appendix, that is a short extract from the “Garden … [24]”) we humbly ask the readers to look at the references [1-58] and find how much the author has been fond of Quality and Scientificness in the Quality (Statistics, Mathematics, Thermodynamics, …) Fields.
The errors, in the “Garden … [24]”, are caused by the lack of knowledge of sound statistical concepts about the properties of the parameters of the parent distribution generating the data, and the related Confidence Intervals. For the I-CC_TBE the computed Control Limits (which are actually the Confidence Intervals), in the literature are wrong due to lack of knowledge of the difference between Probability Intervals (PI) and Confidence Intervals (CI); see the figures 22, 23, 26 and 1). Therefore, the consequent decisions about Process IC and OOC are wrong.
We saw that RIT is able to solve various problems in the estimation (and Confidence Interval evaluation) of the parameters of distributions. The basics of RIT have been given.
We could have shown many other cases (from papers not mentioned here, that you can find in [22,23,24]) where errors were present due to the lack of knowledge of RIT and sound statistical ideas.
Following the scientific ideas of Galileo Galilei, the author many times tried to compel several scholars to be scientific (Galetto 1981-2025). Only Juran appreciated the author’s ideas when he mentioned the paper “Quality of methods for quality is important” at the plenary session of EOQC Conference, Vienna. [1]
For the control charts, it came out that RIT proved that the T Charts, for rare events and TBE (Time Between Events), used in the software Minitab, SixPack, JMP or SAS are wrong [56,57,58]. So doing the author increased the h-index of the mentioned authors who published wrong papers.
RIT allows the scholars (managers, students, professors) to find sound methods also for the ideas shown by Wheeler in Quality Digest documents.
We informed the authors and the Journals who published wrong papers by writing various letters to the Editors…: no “Corrective Action”, a basic activity for Quality has been carried out by them so far. The same happened for Minitab Management. We attended a JMP forum in the JMP User Community and informed them that their “Control Charts for Rare Events” were wrong: they preferred to stop the discussion, instead to acknowledge the JMP faults [56,57,58].
So, dis-quality continues to be diffused people and people continue taking wrong decisions…
Deficiencies in products and methods generate huge cost of Dis-quality (poor quality) as highlighted by Deming and Juran. Any book and paper are products (providing methods): their wrong ideas and methods generate huge cost for the Companies using them. The methods given here provide the way to avoid such costs, especially when RIT gives the right way to deal with Preventive Maintenance (risks and costs), Spare Parts Management (cost of unavailability of systems and production losses), Inventory Management, cost of wrong analyses and decisions.
Figure 28. Probability Intervals L-----U versus Confidence Intervals LCL-----UCL in Control Charts.
Figure 28. Probability Intervals L-----U versus Confidence Intervals LCL-----UCL in Control Charts.
Preprints 162230 g028
We think that we provided the readers with the belief that Quality of Methods for Quality is important.
The reader should remember the Deming’s statements and the ideas in [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
Unfortunately, many authors do not know Scientifically the role (concept) of Confidence Intervals (Appendix B) for Hypothesis Testing.
Therefore, they do not extract the maximum information form the data in the Process Control.
Control Charts are a means to test the hypothesis about the process states, H0={Process In Control} versus H1={Process Out Of Control}, with stated risk α=0.0027.
We have a big problem about Knowledge: sound Education is needed.
We think that the Figure 29 conveys the fundamental ideas about the need of Theory for devising sound Methods, to be used in real applications in order to avoid the Dis-quality Vicious Circle.
Humbly, given our commitment to Quality and our long-life love for it [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], we would venture to quote Voltaire:
It is dangerous to be right in matters on which the established men are wrong.” because “Many are destined to reason wrongly; others, not to reason at all; and others, to persecute those who do reason.” So, “The more often a stupidity is repeated, the more it gets the appearance of wisdom.” and “It is difficult to free fools from the chains they revere.”
Let’s hope that Logic and Truth prevail and allow our message to be understood (figs. 27, 28).
The objective of collecting and analysing data is to take the right action. The computations are merely a means to characterize the process behaviour. However, it is important to use the right Control Limits take the right action about the process states, i.e., In Control versus Out Of Control.
On July-December 2024 we again verified (through several new downloaded papers, not shown here) that the Pandemic Disease about the (wrong) Control Limits, that are actually the Probability Limits of the PI is still present …
There will be any chance that the Pandemic Disease ends? See the Excerpt 12: notice the (ignorant) words “plugging into …”. The only way out is Knowledge… (fig. 28): Deming’s [7,8] Profound Knowledge, Metanoia, Theory.
Excerpt 12. From “Conditional analysis of Phase II exponential chart… an event”, Q. Tech. & Quantitative Mgt, ’19
Excerpt 12. From “Conditional analysis of Phase II exponential chart… an event”, Q. Tech. & Quantitative Mgt, ’19
Preprints 162230 sch012
We think that we provided the readers with several ideas and methods to be meditated in view of the applications, generating wealth for the companies using them.
The documents [56,57,58] are very important: ASSURE …
There is no “free lunch”: metanoia and study are needed and necessary.

Funding

This research received no external funding.

Data Availability Statement

“MDPI Research Data Policies” at https://www.mdpi.com/ethics.

Acknowledgments

In this section, you can acknowledge any support given which is not covered by the author contribution or funding sections. This may include administrative and technical support, or donations in kind (e.g., materials used for experiments).

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A (related to [24])

Excerpt A1. Typical statements in the “Garden …[24]” where the authors name LCL and UCL what actually are the Probability Limits L and U. See the figure 9 and the Excerpt 10.
Excerpt A1. Typical statements in the “Garden …[24]” where the authors name LCL and UCL what actually are the Probability Limits L and U. See the figure 9 and the Excerpt 10.
Preprints 162230 g0a1Preprints 162230 g0a2Preprints 162230 g0a3
Many other cases, with the same errors, can be found in the “Garden …[24], and the Conclusions” where the authors name LCL and UCL what actually are the Probability Limits L and U.
There is no “free lunch”: metanoia and study are needed and necessary.

References

  1. Galetto, F. , Quality of methods for quality is important. European Organisation for Quality Control Conference, Vienna. 1989. [Google Scholar]
  2. Galetto, F. , GIQA, the Golden Integral Quality Approach: from Management of Quality to Quality of Management. Total Quality Management (TQM), Vol. 10, No. 1999. [Google Scholar]
  3. Zhuang, Y. , Bapat, S.R.; Wang, W. Statistical Inference on the Shape Parameter of Inverse Generalized Weibull Distribution. Mathematics 2024, 12, 3906. [Google Scholar] [CrossRef]
  4. Hu, J. , Zheng, L.; Alanazi, I. Sequential Confidence Intervals for Comparing Two Proportions with Applications in A/B Testing. Mathematics 2025, 13, 161. [Google Scholar] [CrossRef]
  5. Alshahrani, F. , Almanjahie, I.M.; Khan, M.; Anwar, S.M.; Rasheed, Z.; Cheema, A.N. On Designing of Bayesian Shewhart-Type Control Charts for Maxwell Distributed Processes with Application of Boring Machine. Mathematics 2023, 11, 1126. [Google Scholar] [CrossRef]
  6. Belz, M. Statistical Methods in the Process Industry: McMillan; 1973.
  7. Casella, Berger, Statistical Inference, 2nd edition: Duxbury Advanced Series; 2002.
  8. Cramer, H. Mathematical Methods of Statistics: Princeton University Press; 1961.
  9. Deming W., E. , Out of the Crisis, Cambridge University Press; 1986.
  10. Deming W., E. , The new economics for industry, government, education: Cambridge University Press; 1997.
  11. Dore, P. , Introduzione al Calcolo delle Probabilità e alle sue applicazioni ingegneristiche, Casa Editrice Pàtron, Bologna; 1962.
  12. Juran, J. , Quality Control Handbook, 4th, 5th ed.: McGraw-Hill, New York: 1988-98.
  13. Kendall, Stuart, (1961) The advanced Theory of Statistics, Volume 2, Inference and Relationship:, Hafner Publishing Company; 1961.
  14. Meeker, W. Q. , Hahn, G. J., Escobar, L. A. Statistical Intervals: A Guide for Practitioners and Researchers. John Wiley & Sons. 2017. [Google Scholar]
  15. Mood, Graybill, Introduction to the Theory of Statistics, 2nd ed.: McGraw Hill; 1963.
  16. Rao, C. R. , Linear Statistical Inference and its Applications: Wiley & Sons; 1965.
  17. Rozanov, Y. , Processus Aleatoire, Editions MIR: Moscow, (traduit du russe); 1975.
  18. Ryan, T. P. , Statistical Methods for Quality Improvement: Wiley & Sons; 1989.
  19. Shewhart W., A. , Economic Control of Quality of Manufactured Products: D. Van Nostrand Company; 1931.
  20. Shewhart, W.A. , Statistical Method from the Viewpoint of Quality Control: Graduate School, Washington; 1936.
  21. D. J. Wheeler, Various posts, Online available from Quality Digest.
  22. Galetto, F. Papers, and Documents of FG, Research Gate. 2014. [Google Scholar]
  23. Galetto, F. , (2015-2024), Papers, and Documents of FG, Academia.
  24. Galetto, F. , (2024), The garden of flowers, Academia.
  25. Galetto, F. , Affidabilità Teoria e Metodi di calcolo: CLEUP editore, Padova (Italy); 1981-94.
  26. Galetto, F. Affidabilità Prove di affidabilità: distribuzione incognita, distribuzione esponenziale: CLEUP editore, Padova (Italy); 1982, 85, 94.
  27. Galetto, F. , Qualità. Alcuni metodi statistici da Manager: CUSL, Torino (Italy; 1995-99).
  28. Galetto, F. , Gestione Manageriale della Affidabilità: CLUT, Torino (Italy); 2010.
  29. Galetto, F. , Manutenzione e Affidabilità: CLUT, Torino (Italy); 2015.
  30. Galetto, F. Reliability and Maintenance, Scientific Methods, Practical Approach, Vol-1: 2016. Available online: www.morebooks.de.
  31. Galetto, F. Reliability and Maintenance, Scientific Methods, Practical Approach, Vol-2: 2016. Available online: www.morebooks.de.
  32. Galetto, F. Statistical Process Management, ELIVA press ISBN 9781636482897; 2019.
  33. Galetto, F. , Affidabilità per la manutenzione, Manutenzione per la disponibilità: tab edizioni, Roma (Italy), ISBN 978-88-92-95-435-9. 2022. Available online: www.tabedizioni.it.
  34. Galetto, F. , (2015) Hope for the Future: Overcoming the DEEP Ignorance on the CI (Confidence Intervals) and on the DOE (Design of Experiments), Science J. Applied Mathematics and Statistics, Vol. 3, No. 3, pp. 99-123. [CrossRef]
  35. Galetto, F. , (2015) Management Versus Science: Peer-Reviewers do not Know the Subject They Have to Analyse, Journal of Investment and Management. Vol. 4, No. 6, pp. 319-329. [CrossRef]
  36. Galetto, F. , (2015) The first step to Science Innovation: Down to the Basics., Journal of Investment and Management. Vol. 4, No. 6, pp. 319-329. [CrossRef]
  37. Galetto, F. , (2021) Minitab T charts and quality decisions, Journal of Statistics and Management Systems. [CrossRef]
  38. Galetto, F. (2012) Six Sigma: help or hoax for Quality?, 11th Conference on TQM for HEI, Israel.
  39. Galetto, F. , (2020) Six Sigma_Hoax against Quality_Professionals Ignorance and MINITAB WRONG T Charts, HAL Archives Ouvert, 2020.
  40. Galetto, F. (2021) Control Charts for TBE and Quality Decisions, Academia.edu.
  41. Galetto, F. (2021) ASSURE: Adopting Statistical Significance for Understanding Research and Engineering, Journal of Engineering and Applied Sciences Technology, ISSN: 2634 – 8853, 2021 SRC/JEAST-128. [CrossRef]
  42. Galetto, F. Control Charts, Scientific Derivation of Control Limits and Average Run Length, International Journal of Latest Engineering Research and Applications (IJLERA) ISSN: 2455-7137 Volume – 08, Issue – 01, January 2023, PP – 11-45. 2023. [Google Scholar]
  43. Galetto, F. (2006) Quality Education and quality papers, IPSI, Marbella (Spain). 2006. [Google Scholar]
  44. Galetto, F. Quality Education versus Peer Review, IPSI, Montenegro. 2006. [Google Scholar]
  45. Galetto, F. , (2006) Does Peer Review assure Quality of papers and Education?
  46. Galetto, F. Quality Education on Quality for Future Managers, 1st Conference on TQM for HEI (Higher Education Institutions), Toulon (France). 1998. [Google Scholar]
  47. Galetto, F. Quality Education for Professors teaching Quality to Future Managers, 3rd Conference on TQM for HEI, Derby (UK). 2000. [Google Scholar]
  48. Galetto, F. Looking for Quality in "quality books", 4th Conference on TQM for HEI, Mons (Belgium). 2001. [Google Scholar]
  49. Galetto, F. Quality and Control Carts: Managerial assessment during Product Development and Production Process, AT&T (Society of Automotive Engineers), Barcelona (Spain). 2001. [Google Scholar]
  50. Galetto, F. Quality QFD and control charts, Conference ATA, Florence (Italy). 2001. [Google Scholar]
  51. Galetto, F. Business excellence Quality and Control Charts”, 7th TQM Conference, Verona (Italy). 2002. [Google Scholar]
  52. Galetto, F. Fuzzy Logic and Control Charts, 3rd ICME Conference, Ischia (Italy). 2002. [Google Scholar]
  53. Galetto, F. Analysis of "new" control charts for Quality assessment, 5th Conference on TQM for HEI, Lisbon (Portugal). 2002. [Google Scholar]
  54. Galetto, F. The Pentalogy, VIPSI, Belgrade (Serbia). 2009. [Google Scholar]
  55. Galetto, F. The Pentalogy Beyond, 9th Conference on TQM for HEI, Verona (Italy). 2010. [Google Scholar]
  56. Galetto, F. , (2021) ASSURE: Adopting Statistical Significance for Understanding Research and Engineering. Journal of Engineering and Applied Sciences Technology. SRC/JEAST-128. [CrossRef]
  57. Galetto, F. , (2024), News on Control Charts for JMP, Academia.
  58. Galetto, F. , (2024), JMP and Minitab betray Quality, Academia.
Figure 1. The pictorial framework of a Statistical Hypothesis (based on a Probability Model).
Figure 1. The pictorial framework of a Statistical Hypothesis (based on a Probability Model).
Preprints 162230 g001
Figure 2. Theoretical and Practical Difference between L------U and LCL------UCL.
Figure 2. Theoretical and Practical Difference between L------U and LCL------UCL.
Preprints 162230 g002
Figure 5. A “4 units Stand-by system” and its states.
Figure 5. A “4 units Stand-by system” and its states.
Preprints 162230 g005
Figure 6. Example of Reliability R 0 λ t 0 of a “4 units Stand-by system” with MTTF=θ=123 days; t 0 is the total time on test of the 4 units. To compute the CI (with CL=0.8), find the abscissas of the intersections at R 0 λ L t 0 = 0.9 and R 0 λ U t 0 = 0.1 ….
Figure 6. Example of Reliability R 0 λ t 0 of a “4 units Stand-by system” with MTTF=θ=123 days; t 0 is the total time on test of the 4 units. To compute the CI (with CL=0.8), find the abscissas of the intersections at R 0 λ L t 0 = 0.9 and R 0 λ U t 0 = 0.1 ….
Preprints 162230 g006
Figure 7. TTOT (Total Time On Test transform) of the Cancer data. Exponential distribution is suitable.
Figure 7. TTOT (Total Time On Test transform) of the Cancer data. Exponential distribution is suitable.
Preprints 162230 g007
Figure 8. CC of the first 32 Cancer data (Exponential distribution): process OOC (1 point).
Figure 8. CC of the first 32 Cancer data (Exponential distribution): process OOC (1 point).
Preprints 162230 g008
Figure 9. Sequential Test (Wald) of the 32 first data.
Figure 9. Sequential Test (Wald) of the 32 first data.
Preprints 162230 g009
Figure 12. TTOT (Total Time On Test transform) of the 128 Cancer data. Exponential distribution is suitable.
Figure 12. TTOT (Total Time On Test transform) of the 128 Cancer data. Exponential distribution is suitable.
Preprints 162230 g012
Figure 16. Distribution GIW(x|0.51, 8.19, 61.38) for values from 0 to 1.2.
Figure 16. Distribution GIW(x|0.51, 8.19, 61.38) for values from 0 to 1.2.
Preprints 162230 g016
Figure 17. Inverse data (of those in Table 3), sum of the inverse of collected data named “Tot_inverse” and “Tot_inverse_B” 1 n l n 1 e η / x i β from the Distribution GIW(x|0.51, 8.19, 61.38). x in the interpolating formulae is the number of counts, not the collected data.
Figure 17. Inverse data (of those in Table 3), sum of the inverse of collected data named “Tot_inverse” and “Tot_inverse_B” 1 n l n 1 e η / x i β from the Distribution GIW(x|0.51, 8.19, 61.38). x in the interpolating formulae is the number of counts, not the collected data.
Preprints 162230 g017
Figure 18. OC Curve and pdf (multiplied by 40) versus ω (omega), given “Tot_inverse_B” 1 n l n 1 e η / x i β from the Distribution GIW(x|0.51, 8.19, 61.38). The intersections of the OC with the two horizontal lines y=0.025 and y=0.975, provide the limits of the CI.
Figure 18. OC Curve and pdf (multiplied by 40) versus ω (omega), given “Tot_inverse_B” 1 n l n 1 e η / x i β from the Distribution GIW(x|0.51, 8.19, 61.38). The intersections of the OC with the two horizontal lines y=0.025 and y=0.975, provide the limits of the CI.
Preprints 162230 g018
Figure 19. Two OC Curves versus ω (omega). The intersections of each OC with the two horizontal lines y=0.025 and y=0.975, provide the limits of the CIs. The curve “on the right” is for g=73 (the sample size “optimum” for decision, according to Excerpt 9); the other is for g=128.
Figure 19. Two OC Curves versus ω (omega). The intersections of each OC with the two horizontal lines y=0.025 and y=0.975, provide the limits of the CIs. The curve “on the right” is for g=73 (the sample size “optimum” for decision, according to Excerpt 9); the other is for g=128.
Preprints 162230 g019
Figure 20. ICC by JMP on x2 data; the LCL and UCL are not the Control Limits but the Probability Limits.
Figure 20. ICC by JMP on x2 data; the LCL and UCL are not the Control Limits but the Probability Limits.
Preprints 162230 g020
Figure 22. Difference between the Control Limits (LCL and UCL) and the Probability Limits (L and U) by the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
Figure 22. Difference between the Control Limits (LCL and UCL) and the Probability Limits (L and U) by the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58].
Preprints 162230 g022
Figure 23. Difference between the Control Limits (LCL and UCL) and the Probability Limits (L and U) of the Boring data, according to the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], with CL=0.95; the vertical lines (red, green) intersect the horizontal line at the points LCL and UCL.
Figure 23. Difference between the Control Limits (LCL and UCL) and the Probability Limits (L and U) of the Boring data, according to the Theory [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58], with CL=0.95; the vertical lines (red, green) intersect the horizontal line at the points LCL and UCL.
Preprints 162230 g023
Figure 25. The sequential Confidence Limits of the Boring data, with CL=0.95.
Figure 25. The sequential Confidence Limits of the Boring data, with CL=0.95.
Preprints 162230 g025
Figure 26. Sequential Test (Wald) for the Boring data, with α=0.025 and β=0.025.
Figure 26. Sequential Test (Wald) for the Boring data, with α=0.025 and β=0.025.
Preprints 162230 g026
Figure 29. Knowledge versus Ignorance, in Tools and Methods.
Figure 29. Knowledge versus Ignorance, in Tools and Methods.
Preprints 162230 g029
Table 1. Statistical Hypotheses and risks.
Table 1. Statistical Hypotheses and risks.
Preprints 162230 i002
Table 2. Some probability models for data analysis.
Table 2. Some probability models for data analysis.
Name F x | θ parameters Symbol
Exponential 1 e x p ( x / θ ) θ E(x|θ)
Weibull 1 e x p ( x / η ) β β η W(x|β,η)
Inverted Weibull 1 e η / x β β η IW(x|β,η)
General Inverted Weibull [ 1 e η / x β ] ω β η ω GIW(x|β,η,ω)
Modified Power Generalised Weibull 1 e 1 1 + x / η β ω e x / θ β η ω θ MPGW ( x | β , η , ω , θ )
Maxwell 2 / π σ 3 0 x t 2 e t 2 / ( 2 σ 2 ) d t σ2 MW(x|σ)
Normal 1 2 π σ 0 x e t μ 2 / ( 2 σ 2 ) d t μ σ2 N(x|μ, σ2)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated