*That’s*fake news.

*Real*news COSTS. Please turn off your ad blocker for our web site.

Our PROMISE: Our ads will never cover up content.

Six Sigma

Published: Monday, October 31, 2022 - 11:03

The cumulative sum (or Cusum) technique is occasionally offered as an alternative to process behavior charts, even though they have completely different objectives. Process behavior charts characterize whether a process has been operated predictably. Cusums *assume* that the process is already being operated predictably and look for deviations from the target value. Thus, by replacing process characterization with parameter estimation, Cusums *beg* the very question process behavior charts were created to address.

To illustrate the Cusum approach and compare it with an average chart, I’ll use the example from page 20 of Shewhart’s first book, *Economic Control of Quality of Manufactured Product* (Martino Fine books, 2015 reprint)*.*These data consist of 204 measurements of electrical resistivity for an insulator. Shewhart organized them into 51 subgroups of size four, based upon the time order in which the measurements were obtained. Figure 1 gives the averages and ranges for these 51 subgroups.

The grand average is 4,498 megohms, and the average range is 659 megohms. The target value for this material was 4,400 megohms. If we place the central line for the average chart at 4,400 megohms, our average chart will detect deviations from the target value. Since the objective of Cusum is to detect deviations from target, then using 4,400 as the central line for the average chart will give us a direct comparison between the two techniques.

The average chart shows nine excursions where the measurements were detectably different from 4,400. Including points in the same runs, these excursions involved 27 of the 51 subgroups. The numbers between the two charts denote those subgroups where the average chart first detects each of these nine excursions. Seven of these signals are points outside the three sigma limits, and two signals (at points 16 and 39) are runs beyond two sigma. Thus, the story told by the average chart is one of a process that is off target more than half the time.

Now we turn to see what story a Cusum will tell about these data.

Although the original version of the cumulative sum technique was based on a graph, today it is an algorithmic technique. To create your Cusum algorithm, certain information is needed from the data, and certain choices have to be made by the user.

**Step one: Select a target value for the time series**

The Cusum technique is essentially a sequential test of hypotheses for detecting deviations from a target value. (The null hypothesis being that the process is operating on target.) In the case of the resistivity data, the target value is 4,400 megohms.

**Step two: Obtain a measure of dispersion for the time series**

Under the assumption that the process has been operated predictably, the usual estimate for dispersion would be the standard deviation statistic computed using the 51 subgroup averages. Here, this statistic is *s* = 352 megohms. However, since we already know the assumption of a predictable process is not true, we’ll use the average range of 659 megohms to estimate:

When a process is operated predictably, these two different estimates of dispersion will be quite similar. By using the second value, we substantially increase the sensitivity of our Cusum technique.

**Step three: Compute the standardized deviations from target**

We subtract the target value from each subgroup average and divide by *Sigma (Averages) *to get the standardized deviations from target:

**Step four: Determine the size of shift to be detected**

From 1954 to the present, Cusum has been presented as being *more* sensitive to process shifts than process behavior charts. One reason for this is that Cusums are traditionally set up to detect small shifts. The Cusum shift constant, *k*, is one-half the size of the minimum shift that you wish to detect. This shift is expressed as a multiple of the standard deviation of the statistic being tracked by the Cusum. By far, the most common value used for the Cusum shift constant has been *k* = 0.5. With subgroups of size *n* = 4, this shift constant will result in a Cusum that should detect shifts greater than:

While it is logical to assume that a Cusum that detects shifts greater than 0.5 *SD(X)* will be more sensitive than the average chart in figure 2, we’ll see this isn't the case.

**Step five: Determine the critical value for the cumulative sums**

The critical value for the cumulative sums, *h*, will depend upon the Cusum shift constant, *k, *and the advertised alpha level for the Cusum technique:

Thus, in addition to choosing the size of shift to be detected, the user also has to choose a risk of a false alarm, *α*. Here, since we are seeking to compare techniques, we will use the nominal alpha level of the average chart, which is *α* = 0.003. With* k* = 0.5, the formula above gives:

*h* = 5.81

Thus, in addition to an estimate of dispersion for the process and a known target value, the Cusum technique depends upon three parameters—*k, h*, and the subgroup size *n*.

**Step six: Compute the cumulative sums**

The Cusum algorithm for the standardized deviations from target computes two partial sums for each value in the time series. These will be the upper cumulative sum, *U*_{i}*, *and the lower cumulative sum, *L*_{i}, where:

*U _{i}* = MAX{ (

*L _{i}* = MIN{ 0 and (

The initial values of *U*_{0} and *L*_{0} are taken to be zero.

Whenever *U*_{i} exceeds the critical sum *h*, or whenever *L*_{i} falls below the negative of the critical sum,* –h*, the Cusum is said to have detected a shift away from the target value.

The operation of the Cusum algorithm is illustrated in the table in figure 3. There, the Cusum identifies three periods involving 21 subgroups where the average is said to be detectably different from the target value of 4,400. These departures were first detected at subgroups 8, 21, and 40.

The first Cusum signal identifies eight subgroups (subgroups 8 through 15) as being detectably above 4,400. The second Cusum signal identifies subgroup 21 as detectably above 4,400. This subgroup is the fifth in a run of five subgroups that average just over 1 standard error above 4,400 (which is just the size of shift we asked the Cusum to find). Finally, the Cusum identifies the last 12 subgroups (subgroups 40 through 51) as being detectably greater than 4,400.

How these two techniques worked with this data set is summarized in figure 4. There, we find that these two techniques characterize 18 of the 51 subgroups differently—they disagree on more than one-third of the subgroups!

Cusum

So, is the Cusum more sensitive than the average chart? It did detect a one-sigma shift for subgroups 17 through 21 that the average chart did not detect. However, this Cusum missed six three-sigma signals (subgroups 3, 4, 5, 22, 31, and 36), and it also missed two two-sigma signals (subgroups 15 and 16 and subgroups 38 and 39). Moreover, this Cusum reported six subgroups with averages *below* 4,400 as being detectably *greater* than 4,400 (subgroups 10, 13, 14, 15, 49, and 50).

So while this Cusum did belatedly detect the two long excursions above 4,400, the inertia inherent in the cumulative sums made the Cusum slow to detect when the deviations began, and slow to detect when those deviations ended. Thus, with eight missed signals and six spurious signals, we can only conclude that the “increased sensitivity” of the Cusum is more of a myth than a reality.

The following is what Shewhart wrote about figure 2.

“Several of the [averages] lie outside these limits. This was taken as an indication of the existence of causes of variability which could be found and eliminated.

“Further research was instituted at this point to find these causes of variability. Several were found, and after these had been eliminated another series of observed values gave the results [on the right in Figure 5]. Here we see that all of the points lie within the limits. We assumed therefore, upon the basis of this [chart], that it was not feasible for research to go much further in eliminating causes of variability. Because of the importance of this particular experiment, however, considerably more work was done but it failed to reveal causes of variability. Here, then, is a typical case where the [process behavior chart] indicates when variability should be left to chance.”

Notice that they not only removed the outlying averages, but they also ended up with much tighter limits centered near 4,400. By identifying the assignable causes of exceptional variation and making them part of the controlled inputs for the process, we not only eliminate outliers, but we also remove major sources of variation from our processes.

Because of the many choices required, many different Cusums can be created to meet various criteria. In fact many different Cusums can be created to meet a single criterion. These choices may allow a Cusum to be more sensitive to small shifts than a process behavior chart; however, by the very nature of the inertia inherent in the cumulative sums, they will always be slower than a process behavior chart in detecting large, sudden, or ephemeral changes.

This fact of life has been well documented in the literature. It is the reason several authors have suggested using Shewhart’s average chart along with a Cusum, the idea being that the process behavior chart will detect the large changes while the Cusum will detect the small changes.

Of course, using the Western Electric zone tests with a process behavior chart will accomplish the same thing with much less effort. But when you have invested all the time and effort required to figure out your Cusum algorithm, you are probably going to insist on using it come hell or high water!

Shewhart created process behavior charts to detect large changes because those are the changes where it will be worthwhile to find the cause and fix it. He was not concerned with small shifts because he was looking for ways to improve a process, and small shifts have little to offer.

Cusum was created to maintain the status quo by detecting small shifts in the average of a predictable process. Recent articles show that statisticians are currently trying to tweak the Cusum technique to make it less sensitive to small changes and more sensitive to large changes. This means they are trying to get Cusum to operate more like a process behavior chart.

Why would they try to do this? Because there is a mathematical theorem that tells them that when the process is operated predictably, the Cusum is an optimal technique for detecting *sustained* shifts in location. Blinded by this limited “optimality,” they look no further. However, when the assumption of a predictable process is not satisfied, the Cusum technique is no longer optimal (as was shown by the example above).

Predictable operation is not a natural state for a production process. And this is the fundamental problem with the Cusum approach. You cannot simply assume a process is being operated predictably. Predictable operation is an accomplishment that can only be maintained by the continued use of a process behavior chart. A process behavior chart is simpler to set up than Cusum; it is easier to use than Cusum; it is more sensitive to shifts that are economically important than is Cusum; and a process behavior chart is necessary to maintain a predictable process. So why go anywhere else?

## Comments

## Thank you for the article in understandable human language.

Thank you for the article, dear Danald

## Process Behavior Chart vs Control Chart

I never miss reading a Dr. Wheeler article in

Quality Digest. They are always insightful and challenging.I understood that Dr. Wheeler uses the term process behavior chart instead of Dr. Shewhart's term control chart. Am I incorrect?

If I am correct, I am thrown off by the use of the material target value of 4,000 for the center line of average chart rather than the grand mean of 4,498. Page 20 of

Economic Control of Quality of Manufactured Productused 4,498.I would appreciate any insight into my misunderstanding.

Thanks

## Reply for Jonathan Boyer

Normally we would use the average of 4498 as the central line. However, since Cusum is a technique for detecting deviations from a TARGET value, I used the target value of 4,400 on the Average chart to make the comparison between the two twchniques more equitable

## Process Behavior Chart vs Control Chart

Dr. Wheeler, Thanks for clearing that up for me. Your article did say that, but I did not grasp it when I read the article.

Thanks again,

Jonathan

## The Simplest Solution is Often the Best Solution

It's hard enough to get people to use and understand simple X and R charts. Complicating the calculations with other charts makes prospective users run for the hills.

Dr. Wheeler continues to demystify SPC and point us in useful directions. Thanks!

## Wonderful articles

These are just wonderful articles. I used Dr. Wheeler's simple, easy to understand books for some 35 years to bring order out of chaos. Thanks.