Document 287747

Universal Journal of Industrial and Business Management 2(4): 103-110, 2014
DOI: 10.13189/ ujibm.2014.020402
http://www.hrpub.org
Efficient Determination of Sampling Rate and Sample Size
in Statistical Process Control
Alessio Gerardo Maugeri1,*, Gabriele Arcidiacono2
1
Leanprove – A&C srl, Via Alfonso La Marmora, 45, 50121 Florence, Italy
Università degliStudiGuglielmo Marconi, Dipartimento per le Tecnologie, e i Processi di Innovazione (DTPI)
Via Plinio, 44, 00193 Rome, Italy
*Corresponding Author: [email protected]
2
Copyright © 2014 Horizon Research Publishing All rights reserved.
Abstract We propose a simple method for the
determination of minimum efficacious sampling rate and
sample size in Shewhart-like control charts, a controversial
topic both in the academic and industrial fields dealing with
Statistical Process Control (SPC). By modeling the control
procedure as the sampling of a stochastic process and
analyzing data in the frequency realm, it is possible to
identify meaningful system time scales and apply the
well-known Nyquist–Shannon sampling theorem in order to
define conditions for an efficient quality control practice.
Such conditions express the minimal requirements for an
efficacious monitoring of quality indices, indicating the
minimal efficacious sampling rate and the minimal effective
size of rational subgroups in Xbar and R (or S) charts, which
are useful both in the set-up phase and in the on-line control
phase of the Shewhart’s control charts. Results can be
applied also to I-MR charts. No statistical assumptions are
made on the monitored data; in particular neither statistical
independence nor Gaussianity is assumed in the derivation of
the results. We focus on continuous processes like those
typical in, but not limited to, e.g. refining, chemical
processing and mining.
Keywords Sampling Rate, Sample Size, Control Charts,
Statistical Process Control (SPC), Nyquist-Shannon
Sampling Theorem, Correlation Time, Lean Six Sigma
(LSS)
1. Introduction
Improving quality by reducing variability is the
fundamental idea at the very heart of successful managerial
approaches like the well-known Lean Six Sigma (LSS) [1]
and the Total Quality Management (TQM) [2], which
emphasize the role of Statistical Process Control (SPC) in
process performance improvements and quality control.
New tools aimed at these purposes are continually
developed. Undoubtedly, however, industrial SPC still
heavily relies on the Shewhart’s pioneering work [3], whose
main tool, the Control Charts (CCs), is widespread and
ubiquitously utilized both in manufacturing and service
companies.
CCs allow monitoring the variability of strategic features,
e.g. numerical quality indices or Critical To Quality (CTQ)
variables, as they are referred to in LSS, of the system under
control, and allow comparing such variability with specific
control limits in order to asses if it is acceptable or not
(Shewhart’s approach).
Repeated observations of a CTQ are therefore needed in
the time domain, in order to assess if the process is in control
or not; control limits being determined from the CTQ
observations themselves.
1.1. SPC Means Sampling
Quality indices and CTQs routinely monitored in all kind
of businesses by CC-like procedures can be modeled as
sequences of random variables, i.e. as discrete-time
stochastic processes. Whenever such indices or CTQs may
be measured continuously in time, we might think of their
monitoring as the sampling of a continuous-time (or
piecewise continuous-time) stochastic quality process (or
simply quality process or quality signal). For general
definitions on stochastic processes we refer to [4-5].
Due to randomness (fluctuations of both controllable and
uncontrollable degrees of freedom involved in the process
dynamics), we cannot predict what the outcome of the
quality control will be, but the process will be under
statistical control (or stable as is customary to say in LSS) if
it may (at least) be modeled by a wide-sense-stationary
(WSS) stochastic process. Such processes are sometimes
also referred to as weak-sense stationary. Finally, in SPC
practices it is customary to assume that the sampled process
is stable and ergodic, so we use the temporal averages
obtained through the sampling to evaluate statistical
moments of interest for the quality monitoring purposes.
1.2. Limits of Sample-Size Formulas
A non-trivial problem often encountered when adopting
104
Efficient Determination of Sampling Rate and Sample Size in Statistical Process Control
CCs is how to choose a proper sampling rate and a proper
numerosity of the sample (i.e. sample size), especially if no
assumptions are made on the distribution function (usually
normal data are assumed) and on the independence (usually
statistical independence is assumed) of process data. Indeed,
real SPC activities very often need to deal with various forms
of correlation. Despite of this, the subgroup size of Xbar CCs
is often determined by formulas relying on the Central Limit
Theorem, which in turn relies on the statistical independence
of the sampling data. I-MR CCs instead heavily rely on
Gaussianity of the monitored variable.
Nonetheless, the Shewhart’s concept of stability relies
neither on statistical distribution nor on statistical
independence hypothesis of the sample observations, but
only on predictability of process variation [3]. This means
that in principle it is possible to control also processes
exhibiting correlation or processes not featuring Gaussianity.
Indeed Shewhart-type CCs accounting for different types of
correlation [6] and for non-normality [7] have been
developed.
Optimum sampling intervals, optimum sample sizes and
optimum control limits are the topics of many research
works in the Operational Excellence realm and in the context
of Optimization and Design of Industrial Processes [7-12].
Choosing a proper sampling rate and a proper sample size
is an essential task to correctly and efficiently assess
ordinary process variability, whose evaluation is in turn
essential to set up appropriate control limits in Shewhart CCs.
Control limits are subsequently used to assess if the future
process variability (i.e. during on-line control), and temporal
behavior in general, is acceptable or not.
1.3. A New Perspective from Spectral Analysis
Bearing in mind that CTQ monitoring is equivalent to the
sampling of a stochastic process, with this paper we would
like to open a new perspective in SPC, addressing in
principle data sets exhibiting generic correlation. We
propose to determine efficient sampling rates and sample
sizes in Shewhart CCs by addressing the quality process in
the spectral domain, without any constraints on the statistical
distribution function of sampled data nor assuming their
statistical independence. Such new perspective being
potentially useful also for Shewhart-type and non-Shewhart
CCs, in that it indicates a method to properly assess ordinary
process variability.
We will derive our results thanks to the stochastic
sampling theory as customary in many engineering fields
dealing with signal analysis. This will require identifying
both the bandwidth and the correlation time of the quality
signal.
2. Theoretical Framework
2.1. The Nyquist-Shannon Sampling Theorem
An effective sampling is a faithful sampling. This means
we must be able to exactly reconstruct the original signal at
each instant of time from the sampling.
According to the Nyquist–Shannon sampling theorem
[13], a result very well-known in Communications and
Information Theory, a sampled bandlimited (BL)
deterministic signal can be perfectly reconstructed from an
infinite sequence of samples, if the sampling rate fs exceeds
2B samples per unit of time (ut), where B is the fastest rate in
the original signal and 2B is the so called Nyquist rate. Such
a choice allows not incurring in the well-known aliasing
error.
A similar result also holds for BL WSS stochastic
processes [5], i.e. finite-power processes with a vanishing
power spectral density S(f) = 0, ∀ |f| > B. So, we can apply
the theorem also to the sampling of the quality process.
Moreover, being the quality process a random signal, it
can be approximated with arbitrary accuracy by a sum
involving only its past samples, provided that the Nyquist
criterion is satisfied (i.e. fs > 2B), and therefore we do not
need its future history [5,14].
The reconstruction of stochastic processes from their
samples is still a hot topic in many engineering fields and by
now many results have been presented in scientific literature.
For a general overview of stochastic-process sampling
theorems, we refer to [15-16].
2.2. Process Variability: Common and Special Causes
Each industrial process is supposed to have its own
standard working conditions. Stability of the process is
assured by a proper set up of internal degrees of freedom, i.e.
what we usually call best practices. Operational or
managerial best practices concern - but they are not limited
to - persons involved into the process, machines used into the
process, state of life or degree of wear of such machines,
materials used, operating conditions (e.g. temperature,
pressure, or even type of organizational structure), methods
and procedures adopted.
Within the boundaries defined by operational or
managerial best practices, the process will only suffer from
an ordinary variability. These intrinsic and unavoidable
fluctuations of process internal degrees of freedom are due to
so called common causes, as they are usually indicated in the
LSS approach [1]. Common causes cannot be annihilated
(given the process we are monitoring), but only reduced.
Exit from or violation of the best practice boundaries
determines non-standard behaviors of the process dynamics,
due to so called special causes as they are referred to in LSS
methodology [1], and therefore unusual dynamics of internal
degrees of freedom.
Both common and special causes are triggered by external
(with respect to the monitored process) degrees of freedom
that are coupled to process dynamics. Provided such external
degrees of freedom stay in a stability region of the parameter
space (stability being defined with respect to the
best-practice prescriptions and to process robustness to
Universal Journal of Industrial and Business Management 2(4): 103-110, 2014
external fluctuations), no special causes can occur.
Monitoring of the process is aimed at identifying and
eliminating special causes, and reducing common variability
through external degrees decoupling (or reduction of the
strength of their coupling to internal dynamics) or removal.
We may also refer to the ensemble of all the external
degrees of freedom as the environment, which is responsible
of noisy (both common and special) fluctuations of the
quality signal, as shown in fig. 1.
105
c(t). Such variability can only be reduced, but not eliminated
from the given system, and in general is not easily
attributable to specific and identifiable external degrees of
freedom. Being a WSS process, the average value of c(t) is
constant and its autocorrelation function depends only on
time differences [4-5].
No hypothesis are a priori made on s(t), which accounts
for the non-stationary behavior of system dynamics (due to
special causes of variability). It describes not-ordinary
features of system dynamics, chargeable to specific and
addressable external degrees of freedom.
For sake of clarity, without affecting the overall value of
the study, from now on expression (1) will indicate a generic
CTQ, measured on process outputs.
2.4. Continuous-Time Vs Piecewise Continuous-Time
CTQs
Figure 1. System-environment coupling: the system/process we are
monitoring interacts with external degrees of freedom, which are coupled to
system/process dynamics. Red wavy lines represent such coupling. The time
evolution of a CTQ with a natural variability is naively depicted. Interaction
with the environment determines aleatory modifications (dashed-line circle)
of the isolated system/process dynamics (full-line circle), and, consequently,
randomizes the CTQ signal (full-line deterministic signal vs dashed-line
stochastic signal).
2.3. Modeling the Quality Signal
The CTQ to be monitored could be a process input or
output variable, or even a generic process variable, but in any
case we write the CTQ dynamics as a random signal made of
three different components
Q(t)=q(t)+c(t)+s(t)
(1)
where q(t) is the quality function in absence of
environmental noise sources, i.e. in absence of external
degrees of freedom coupled to system dynamics; c(t) is a BL
WSS stochastic process responsible for common variability;
and s(t) is the process accountable for out-of-ordinary
variability and lack of statistical control.
The deterministic function q(t) accounts for the dynamics
of the process degrees of freedom in absence of any kind of
perturbation stemming from external degrees of freedom
coupled to the process ones. Such isolated (or natural, i.e. not
due to any environmental noise sources) system dynamics is
a pure abstraction, representing a regulative ideal, which
cannot be met in practice, due to the unavoidable interaction
with noisy degrees of freedom.
The stochastic process c(t) accounts for the effect of all the
unavoidable fluctuations of the internal degrees of freedom
(triggered by coupling to external ones), which are
responsible for mean-reverting shifts of the system
deterministic quality signal. Normal, i.e. ordinary, process
variability due to common causes is therefore explained by
The stochastic process described by (1) may properly
represent different control practices in both service and
manufacturing industries.
We now specify (1) distinguishing between
continuous-time and piecewise continuous-time CTQs [4-5].
Such continuous processes are frequent in e.g. refining,
chemical processing and mining.
We first address an ideal case, where (1) represents an
infinite, analog (i.e. continuous-time) signal related to a CTQ
measurable on an infinite and continuous output
(continuous-time processes). Although ideal, this situation is
a good starting point to understand the modelling of more
realistic occurrences; furthermore it may well approximate
such occurrences when the discrete nature of the process
outputs do not hinder the sampling procedure, as will be
more clear later on.
We secondly address the case of signals related to CTQs
that are only piecewise continuously measurable (piecewise
continuous-time processes, related to discrete outputs). This
is the case of all real control practices. Due to discreteness,
periodic time intervals exist where it is intrinsically not
possible to measure any CTQ.
In this paper we do not address quality signals related to
CTQs which are uniquely measurable at specified time
instants (i.e. on a countable set of the time domain): this is
for example the case of measures made once per single
output (as e.g. in destructive testing or in some
conformance-to-requirement measures involving the output
as a whole), giving rise to a sequence of random variables
each one related to a unique output (discrete-time processes).
2.4.1. Continuous-Time Processes
Let’s first suppose that the process we are monitoring
generates an output featuring an infinite and continuous
dynamics with respect to time. In this case, monitoring of
process quality can be modeled as the sampling of an infinite,
analog BL stochastic signal.
A continuous-time process describes a CTQ related to an
output whose features vary continuously in time. For
106
Efficient Determination of Sampling Rate and Sample Size in Statistical Process Control
example this could be the case of an infinite extruded tube, of
an infinite liquid flow or of the maximum amount of
information that can be reliably transmitted over a
communication channel, respectively featuring a
continuously variable thickness, viscosity and channel
capacity (which are all continuous CTQs with respect to
some parameters). A continuous-time process may also
describe conformance to some requirements or number of
non-compliant features per reference unit (which are instead
discrete CTQs).
Process stability and dispersion around the target are
assessed by monitoring one or more of such CTQs related to
the output.
Finally, both uncountable, i.e. continuous (e.g. inches of
thickness, centipoise of viscosity or bits conveyed per
second), and countable, i.e. discrete (e.g. good/no-good
indices or number of defects per reference unit), CTQs may
be modeled by a continuous-time process.
2.4.2. Piecewise Continuous-Time Processes
In practice, we never deal with continuous-time processes,
because our outputs are discrete, e.g. parts produced or
documents written up. This means that the quality signal (1),
although (in principle) not time-bounded, is only piecewise
defined, for both continuous and discrete CTQs.
Whatever CTQ variable is chosen to be monitored, it will
be at best piecewise continuously varying in time, due to the
discrete nature of the outputs realized. In comparison with
the previous case, here a limitation is imposed on the number
of subsequent observations we can draw from the process at
a given sampling rate as better outlined in section 3.
2.5. Rational Subgroups: Fast Noise Vs Slow Noise
In the Shewhart’s approach, appropriate evaluation of
system ordinary (i.e. only due to common causes and,
eventually, to natural system dynamics) variability is
fundamental to set up an effective CC featuring proper
control limits.
Shewhart explains that if observations can be grouped in
rational subgroups, then appropriate control limits can be
specified. Rational subgroups are defined as groups of
observations whose variability is only due to common causes
[3].
In other words, the underlying argument for rational
subgroups is that special causes of variability have a slow
dynamics compared to ordinary process dynamics: we might
therefore refer to the effect of special causes as slow noise.
On the contrary, the effect of common causes might be
referred to as fast noise.
Bearing in mind results in section 2.1, set up of proper
rational subgroups could therefore be achieved by analyzing
the spectral content of the monitored system dynamics. In
particular, by identifying the bandwidth upper limit of the
monitored CTQ, it is possible to determine a proper sampling
rate; by identifying the time scale of the CTQ dynamics (as
will be clearer in section 2.6), it is possible to set a proper
sample size for the control procedure. Identifying the
bandwidth upper limit would therefore be useful for both the
phase of set up and on-line control of the CC.
Of course this might require a thorough knowledge of the
underlying monitored process. However we do not need to
know anything about slow noise in order to properly set up
our Shewhart’s sampling rate and sample size. In fact, when
the process is stable (by definition we assess control limits in
condition of statistical control and from data not affected by
special causes) the CTQ dynamics is not affected by s(t),
which operates at frequencies much slower than the CTQ
slowest ordinary dynamics. Setting up of control limits in
Shewhart-like CCs will therefore be based on a spectral
analysis of the BL WSS stochastic process accounting for
common system variability
Qfast(t)=q(t)+c(t)
(2)
Although bandlimitedness assumption is often invoked in
this paper concerning industrial applications (both
manufacturing and service), it is worth saying that sampling
theorems have been developed also for non-BL signals [17].
Furthermore it may also be possible filtering out stray
external (i.e. from sources other than the system, e.g. the
quality measurement system) high frequencies. Finally,
almost sure convergence of the Shannon sampling series has
been demonstrated also for generalized BL processes [18].
2.6. Time Scale of the CTQ Dynamics
The ordinary evolution of each CTQ dynamics is
characterized by a typical time scale accounting for its
common variability, due to both fast noise and, if present, to
system natural variability. In case of a system with e.g. a
natural periodic time evolution, as in fig. 1, such time scale
would be the period of the natural dynamical fluctuations.
Instead, in case of a naturally static CTQ (that is a CTQ stuck
to a target value in absence of any environmental noise
sources), the CTQ time scale would be determined by the
fast-noise component into (2). Fast noise is typically
responsible for fluctuations tending rapidly to zero and, in
this case, we trivially identify the CTQ time scale with the
width of the relative autocorrelation function.
As usual in Signal Theory, the time scale of the CTQ
dynamics is the correlation time τc, which can be defined also
for arbitrary stochastic processes. A formal standard
definition for the correlation time can for example be found
in [4-5]
A practical identification of the correlation time can be
achieved through the observation of the power spectrum of
(2), as customary in Signal Analysis and in many physical
problems [19-21]. Depending on the shape of the power
spectrum, identification of the correlation time may vary,
reflecting the possibly different nature of the dynamical
regimes of the various CTQs. In most cases (e.g. a static
CTQ suffering from fast noise) the width of the
autocorrelation function of (2) is inversely proportional to
the width of the power spectrum (in this case the power
Universal Journal of Industrial and Business Management 2(4): 103-110, 2014
spectrum would have a bell shape). In some cases the power
spectrum may have a peak structure (if e.g. the CTQ has a
natural oscillatory frequency) and the correlation time is
drawn from the inverse of the peak frequency [21].
In any case, the correlation time of (2) defines the time
needed to gain complete knowledge of the monitored CTQ
ordinary variability, and, finally, together with the fastest
rate in the power spectrum of (2), the number of observations
we need to gather to realize proper rational subgroups, as
explained in section 3.
3. Results and Discussions
3.1. Nyquist-Shannon Theorem Applied To Quality
Control
If we monitored our system continuously, we would know
exactly a realization of the random quality signal (1).
Consequently, under the ergodicity assumption, we would be
able to appraise quality dynamics instant by instant.
Unfortunately, due to many practical reasons, it is not
possible or it is not economic a continuous monitoring of the
system, and indeed in practice we just measure (1) at
specified time points (sampling). However we should accept
the discrete quality index stemming from such sampling only
provided it is a faithful approximation of the original signal.
As we have seen, we can borrow results from
stochastic-process sampling theory [14-20].
3.2. Sampling Rate
According to section 2.1, the sampling rate is given by the
fastest frequency (where appropriate, filtering can prevent
aliasing due to external high frequencies) in the power
spectrum of (2).
Mathematically [5], if Sfast(f) is the spectral density of (2)
and Sfast(f) = 0, ∀ |f| > B, then must be
fs > 2B
(3)
B being the fastest rate in the bandwidth of the quality signal
(2).
B being the fastest rate in the bandwidth of the quality
signal (2). Expression (3) holds for both continuous-time
processes and piecewise continuous-time processes
3.3. Sample Size
By virtue of the Nyquist-Shannon theorem, knowing B we
are able to determine the minimal efficacious sampling rate,
which is just above the Nyquist rate as expressed by
condition (3). Acquiring also knowledge of τc allows
determining the number of single observations we need to
gather, because this time scale fully accounts for common
variability of the monitored CTQ, thus setting the minimal
time required to acquire a complete knowledge of process
ordinary dynamics. We can then easily estimate the minimal
107
meaningful sample size as the product of the minimum
efficacious sampling rate and the correlation time, as better
outlined in subsections 3.4.1 and 3.4.2 for Shewhart's CCs.
This information (sampling rate and sample size) will then
be used both to set up proper control limits and to properly
manage the on-line control of the monitored process.
In the case of piecewise continuous-time processes, due to
the fact that we are likely monitoring some CTQ variable on
a discrete output, we may not have enough time to gather
clusters of subsequent observations. Condition (3) will also
apply to such processes, but differently from
continuous-time processes, here sampling of the quality
signal will need to take into account the ratios between τc, tav
(total available time for taking measures of a given CTQ on a
single output, i.e. time interval from the first measurable part
of an output and its last measurable part) and Δt (the time
interval between the last measured part of an output and the
first measurable part of the subsequent output); measure
ability being defined in relation to the specific CTQ to be
monitored.
We will pursue our analysis in section 3.4, highlighting
such relations.
3.4. Shewhart-Type Ccs: Set Up Of Control Limits and
On-Line Control
In order to specify the concepts previously illustrated, let’s
focus on Shewhart-type CCs routinely used in SPC.
Shewhart-type CCs only rely on current observations to
assess the process control state, i.e. each point on the chart
only relies on the sample gathered at the moment of
monitoring (other well known charts, e.g. CUSUM or
EWMA, also utilize previous observations). When adopting
Shewart-type CCs, particular care should therefore be
devoted to the time scales of the monitored CTQ dynamics,
in order to properly estimate sampling rate and sample size,
and, finally, control limits [3,6]. It is worth noticing that
usual formulas to estimate e.g. the required sample size for
subgroups of an Xbar chart may not be riskless, if used when
data presents a certain degree of correlation.
For sake of clarity, let’s adopt the set of definitions in table
1, to be utilized in subsections 3.4.1 and 3.4.2, where we will
indicate how to compute efficient sampling rate and sample
size both in the case of individual (I-MR charts) and
gathered-in-subgroup observations (Xbar-R or Xbar-S
charts). In the following we will assume to know the CTQ
correlation time and its upper rate B, whose empirical
determination will be outlined by the practical procedure in
section 3.5.
3.4.1. I-MR Charts
In order for a sampling to be faithful and to obtain an
unbiased estimate of the variability of the CTQ normal
dynamics, the Overall Sampling Period must be at least as
long as the correlation time, that is T ≳ τc.
According to the Nyquist-Shannon theorem and relation
(3) in particular, for both continuous-time and piecewise
108
Efficient Determination of Sampling Rate and Sample Size in Statistical Process Control
continuous-time processes, a necessary condition for a
faithful sampling is
(4)
fs,b> 2B
which sets the more efficient sampling rate for the I-MR
chart of a continuously (in time) measurable CTQ.
Once we are provided with the sampling rate, we can
estimate the minimal meaningful sample size for the I-MR
chart, made of subsequent measures taken at a rate given by
(4) for a sampling time T~ τc.
For continuous-time processes, the more efficient sample
size, aimed at properly assess control limits, would be the
higher nearest integer to
n = 2B ∙ τc
(5)
For piecewise continuous-time processes, the discrete
nature of the output can be an obstacle to the sampling
procedure, because, given a rate set by (4), we run the risk to
miss subsequent measures within the τc time scale. That is, if
tav≤τc, we cannot reach the value set by (5) via subsequent
measures on a single output (even if the equality holds it is
not possible to take the higher nearest integer to (5) because
the last measure is not feasible). In other words, we are not
allowed monitoring slow components of the CTQ ordinary
variability.
Furthermore, if it also holds Δt≥1/2B, we miss
information even if we take measures from the subsequent
output. This happens due to the discrete nature of the process
outcomes, which acts as a high-pass filter of the CTQ
ordinary dynamics. In this case, in order to avoid
underestimating the process variability and incurring in the
aliasing error, sampling should be extended for time periods
longer than τc, which are to be evaluated taking also into
account the length of the time interval Δt.
From the gathered observations it will then be possible to
determine the CTQ ordinary variability and therefore
establishing proper control limits. On-line control will be
carried out at a rate set by (4).
3.4.2. Xbar-R (S) Charts
We make the reasonable assumptions TS ≥ τc(choosing a
shorter Subgroup Sampling Period would undermine the
validity of the results, as it would not be possible to gather
enough observations) and TW ≠ 0 (i.e. sampling within the
subgroup is not instantaneous).
Within the subgroups we have operating conditions
analogous to the one described in subsection 3.4.1 for I-MR
CCs. According to the Nyquist-Shannon theorem, in order to
properly assess the CTQ ordinary dynamics (including
eventual short- or long-term correlations), a faithful
sampling of a continuous-time process would at least require
TW< 1/2B, i.e.
fs,w > 2B
(6)
and a subgroup size equal to the higher nearest integer to (5).
The rate set by (6) can then be used to evaluate proper
control limits and, subsequently, to carry out on-line control.
Relation (6) also holds for piecewise continuous-time
process, however we may miss slow components of our CTQ
ordinary variability, if tav ≤ τc and sampling is not extended
for time periods longer than τc, analogously to the case in
subsection 3.4.1. Missing slow components of ordinary
dynamics would in general result in a systematic
underestimate of the process variability, which would be
incorrectly transferred to variability between subgroups. The
variability usually observed between subgroups in absence
of special causes may therefore be attributed also to the
intrinsic process filtering determined by the output
discreteness when tav ≤ τc.
As far as the Between Sampling Period is concerned, it
should be short enough to detect the eventual occurrence of
special causes, but not too short in order to reduce
unnecessary sampling effort. If there is a clue of special
cause dynamics, once again the Nyquist-Shannon theorem
suggests how to choose Tb. If for example slow noise is
known to affect the natural dynamics at most with frequency
fslow<< 1 / τc, then we will choose fs,b> 2 fslow. In general,
however, slow noise dynamics is unknown, therefore it is not
possible to determine a priori a smart Between Sampling
Period. A prudential choice could be setting Tb ~ 10 τc,
eventually relaxing this conditions if appropriate from
observations gradually accumulated. In any case a trade-off
need to be found between the risk of missing special cause
detection and the sampling effort.
Table 1. Relevant time scales and parameters: definition of sampling periods, sampling rates and sample size for I-MR CCs, Xbar-R CCs, Xbar-S CCs
Between Sampling
Period
Tb
Time interval between two single observations (I-MR charts) or between two rational subgroups (Xbar-R or
Xbar-S charts)
Between Sampling Rate
fs,b
Number of single observations (I-MR charts) or rational subgroups (Xbar-R or Xbar-S charts) in the ut
Within Sampling Period
Tw
Time interval between two observations within a rational subgroup
Within Sampling Rate
fs,w
Number of observations within the subgroup in the ut
n
Overall numerosity of the sample (I-MR charts) or number of observations in a single rational subgroup (Xbar-R
or Xbar-S charts)
Ts
Duration of the sampling within a rational subgroup
T
Overall duration of the sampling procedure
Sample Size
Subgroup Sampling
Period
Overall Sampling
Period
Universal Journal of Industrial and Business Management 2(4): 103-110, 2014
3.5. Practical Determination of Correlation Time and
Nyquist Rate
We outline a practical procedure in order to determine the
stable process power spectral density, and finally the CTQ
correlation time and the Nyquist rate.
The procedure encompasses the following steps:
Qualitatively estimating the order of magnitude of the
fastest time scale of the ordinary (which includes fast noise)
dynamics of the CTQ to be monitored;
Sampling at a reasonable rate based on the previous
estimation (e.g. three times faster) and compute the
corresponding power spectral density until a valid result is
obtained and rapid changes in the spectra stop (stabilization
in the frequency domain);
Sampling again increasing the sampling rate and
re-computing power spectral density, till no observable
differences appear in the stabilized spectra.
Once the stable spectrum has been computed, it is possible
to observe its shape and bandwidth in order to properly
identify the correlation time (see section 2.6) and the upper
rate.
3.6. Future Research
109
the CTQ common variability allows determining lower
bounds for both the sampling rate and the sample size of
usual ShewhartCCs, for both countable and uncountable
CTQs. The sampling rate lower bound is the Nyquist rate,
while the sample size lower bound is the product of the
Nyquist rate times the correlation time.
The illustrated approach is potentially extensible to other
CCs and provides a general method to properly assess
ordinary process variability.
A spectral characterization of common and special causes
of process variability has also been illustrated in terms of fast
and slow noise.
Acknowledgements
We would like to thank dr. Giuseppe Mangano for very
fruitful and stimulating discussions and Alessandro Costa for
proofreading. Dr. Maugeri acknowledges financial support
from Leanprove – A&C srl.
REFERENCES
[1]
G. Arcidiacono, C. Calabrese, K. Yang. Leading processes to
lead companies: Lean Six Sigma, Springer, 2012.
[2]
B. O. Ehigie, E. B. McAndrew. Innovation, diffusion and
adoption of total quality management (TQM), Management
Decision, Vol. 43 Iss: 6, 925 – 940, 2005.
[3]
W. A. Shewart. Economic Control of Quality of
Manufactured Product, D. van Nostrand Company, Inc:
Toronto, 1931.
[4]
Y. Dodge. The Oxford Dictionary of Statistical Terms,
International Statistical Institute, Oxford University Press,
6th ed., USA, 2006.
[5]
A. Papoulis, P. S. Unnikrishna. Probability, Random
Variables and Stochastic Processes,4th ed., McGraw-Hill,
2002.
4. Conclusions
[6]
H. J. Lenz, P. TH. Wilrich. Frontiers in Statistical Quality
Control (Book 8), Physica-Verlag, Heidelberg, 2006.
Neglecting correlations or assuming Gaussianity when it
is not the case might undermine many SPC practices. Special
CCs have been developed to address some forms of
correlation or non-normal data, but a general method to deal
with whatever set of data might be missing. Here a method to
determine the minimum efficacious sampling rate and
sample size in Shewhart CCs and Shewhart-type CCs has
been presented, with no assumptions concerning data
independence or their statistical distribution. The method
models the monitoring of a control variable as the sampling
of a stochastic process and approaches the problem of
process control in the frequency realm. Results rely on the
Nyquist–Shannon sampling theorem and on knowledge of
the correlation time of the quality signal. In particular, the
Nyquist criterion together with the time scale accounting for
[7]
H. Chen, Y. Pao. The joint economic-statistical design of X
and R charts for nonnormal data, Quality and Reliability
Engineering International, Vol. 27, No.3, 269-280, 2010. DOI:
10.1002/qre.1116.
[8]
C. A. Carolan, J. F. Kros, S. E. Said. Economic design of
Xbar control charts with continuously variable sampling
intervals, Quality and Reliability Engineering International,
Vol. 26, No. 3, 235-245, 2009. DOI: 10.1002/qre.1050.
[9]
V. Carot, J. M. Jabaloyes, T. Carot. Combined double
sampling and variable sampling interval X chart,
International Journal of Production Research, Vol. 40, No. 9,
2175-2186, 2002. DOI: 10.1080/00207540210128260
Future research should be devoted to develop specific case
studies and to compare the sensitivity of traditional CCs with
that of CCs coming from spectral analysis.
Generating data sets featuring known correlation, it would
be possible comparing the results obtained with modified
Shewhart CCs or Shewhart-type CCs (if available for the
type of correlation generated) with results coming from our
approach, which should prove to be valid for whatever form
of correlation. In particular it would be beneficial comparing
results from our approach with those from modified
Shewhart CCs, from residual charts and from modified
residual charts in case of serially correlated data. These
charts in fact proved to be effective for serial correlation [6].
[10] Y. C. Lam, M. Shamsuzzaman, S. Zhang, Z. Wu. Integrated
control chart system—optimization of sample sizes, sampling
intervals and control limits, International Journal of
Production Research, Vol. 43, No. 3, 563-582, 2005. DOI:
110
Efficient Determination of Sampling Rate and Sample Size in Statistical Process Control
10.1080/00207540512331311840
[11] Y.C. Lin, C.Y. Chou. On the design of variable sample size
and sampling intervals Xbar charts under non-normality,
International Journal of Production Economics, Vol. 96, No.
2, 249-261, 2005.
[12] Z. Wu, M. Yanga, M. B. C. Khoob, P. Castagliola. What are
the best sample sizes for the Xbar and CUSUM charts?
International Journal of Production Economics, Vol. 131,
No.2, 650-662, 2011.
[13] C. E. Shannon. Communication in the presence of noise,
Proceedings of the Institute of Radio Engineers, Vol. 37, No.1,
10–21, 1949. Reprint as classic paper in Proceedings of the
IEEE, Vol. 86, No.2, 1998.
[14] L. A. Wainstein, V. Zubakov. Extraction of Signals from
Noise, Prentice-Hall: Englewood Cliffs, NJ, 1962.
[15] A. J. Jerri. The Shannon sampling theorem—Its various
extensions and applications: A tutorial review, Proceedings of
the IEEE, Vol. 65, No.11, 1565–1596, 1977.
[16] T. Pogány. Almost sure sampling restoration of band-limited
stochastic signals. In: J. R. Higgins, R. L. Stens. Sampling
Theory in Fourier and Signal Analysis—Advanced Topics,
Oxford University Press: Oxford, U.K., 1999.
[17] Q. Chen, Y. Wang, Y. Wang. A sampling theorem for
non-bandlimited signals using generalized Sinc functions,
Computers and Mathematics with Applications, Vol. 56, No.
6, 1650-1661, 2008. DOI: 10.1016/j.camwa.2008.03.021.
[18] M. Zakai. Band-limited functions and the sampling theorem,
Information and Control, Vol. 8, No.2, 143–158, 1965.
[19] R. L. Allen, D. Mills. Signal Analysis: Time, Frequency,
Scale, and Structure, Wiley-IEEE Press, 1st ed., 2004.
[20] R. A. Meyers. Mathematics of Complexity and Dynamical
Systems, Springer, 2011.
[21] E. Paladino, A. G. Maugeri, M. Sassetti, G. Falci, U. Weiss.
Structured environments in solid state systems: Crossover
from Gaussian to non-Gaussian behaviour, Physica E:
Low-dimensional Systems and Nanostructures, Vol. 40, No. 1,
198–205, 2007.