ITS - Intelligent Transportation Systems Report ITS Home Page

Appendix A:  Sampling Theory for Voice Routing Field Test

The theory behind the staged statistical sampling design for the conduct of the acceptance test of the voice routing system is discussed in this section.

 

The acceptance test developed for this application begins with the following considerations.

 

  1. System failure rate (p0) and probability of error in concluding acceptance test failure (α) – It is initially proposed that a system failure rate of less than one percent is desired.  For this acceptance test procedure, we will collect a data sample and count the number of failures found in this sample.  We will then compare the proportion of failures in the sample against the desired system failure rate.  If the estimate is much greater than one percent, we will conclude that the true system failure rate is likely greater than one percent.  However, there is a possibility that we could make this conclusion in error (i.e., even though the sample failure rate was high, the true system failure rate could be one percent).  The probability of this error is denoted α, and we desire that it be a low value. This value is commonly set at approximately five percent for acceptance sampling.
  2. Tolerable upper limit on system failure rate (p1) and corresponding probability of error in failing to reject initial assumption (β) – If the true system failure rate exceeds the desired one percent threshold, the sample failure rate may not be high enough to recognize this. For any value over one percent, there is a probability β that we will fail to conclude the system failure rate is over one percent.  We desire that β be a small value (as with α, the value is commonly set at approximately five percent for acceptance sampling) at some true system failure rate greater than the desired system failure rate.  This point is often referred to as the tolerable upper limit.  We initially propose a tolerable upper limit on system failure rate of five percent.
  3. For sampling efficiency, the smallest sample size (n) and acceptable number of failures (c) that meet the criteria above should be selected.

Noting that the sample response measurements for this acceptance test fall into only two categories (success or failure), the acceptance test quantities can be related by the following equations, based on the binomial probability distribution:

 

Two Equations with Assumptions. Equation 1 is for the probability, a, of error in concluding system test failure.  It is the sum of probabilities for the terms in a binomial probability distribution (of probability p0, the assumed failure rate) with number of failures greater than the failure limit, c, and continuing up to the total number of samples, n.  
Equation 2 is for the probability, ß, of error in failing to reject the initial assumption (i.e., concluding system test acceptance). It is the sum of probabilities for the terms in a binomial probability distribution (of probability p1, the tolerable upper limit on failure rate) with number of failures ranging from zero to the failure limit, c.  
The equations are valid so long as n and c are integers, p0  is greater than 0 and less than p1, p1 is greater than p0 and less than 1 and c is greater than or equal to zero but less than n.

 

 

From our initial requirements of the acceptance testing, we have established target values for p0 (0.01), α (0.05), p1 (0.05), and β (0.05).  Since these equations produce discrete values, it is not necessarily possible to solve the two equations exactly for desired n and c.  Instead, we develop a solution by fixing p0 and p1 and then experimenting with different values of n and c, searching for corresponding α and β reasonably close to the desired values.

 

For this particular acceptance test, an appropriate solution is found at n=147 and c=3 with p0=0.01, p1=0.05, α=0.06 and β=0.06.  This leads to the following test:

 

  1. Select a sample of 147 items and count the number of failures.
  2. If the number of failures is 3 or less, there is not sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test passes acceptance.  It is possible this conclusion is incorrect (i.e., that the true system failure rate is greater than one percent).  However, if the true system failure rate is five percent more, there is only a six percent chance of this error.  Above a five percent system failure rate, this probability is even lower.
  3. If the number of failures is 3 or more, there is sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test fails acceptance.  It is possible this conclusion is incorrect (i.e., that the true system failure rate is one percent or less).  However, if the true system failure rate is one percent, there is only a six percent chance of this error.  Below a one percent system failure rate, this probability is even lower.

The practical interpretation of this test is:

 

Reducing Tolerable Upper Limit

From the test above, true system failure rates between the p0 and p1 values are a problem because there is a definite risk in these cases that an acceptance sample will pass for these values when it should have failed.  Therefore, it is desirable to make p1 as close as possible to p0.  However, for a fixed p0, α, and β, this can only be done by increasing the sample size n.  Modifying the acceptance sample plan above, p1 can be reduced to four percent for n=294 and c=6 with p0=0.01, p1=0.04, α=0.03 and β=0.05.

 

Though this acceptance test has slightly better α and β and is not exactly comparable to the original test, it does show that we essentially need to double the sample size to get a one percent reduction in the tolerable upper limit.

Staged Sequential Sampling

If acceptance samples are not collected all at one time, it may be possible to evaluate the sample results as they are obtained.  The acceptance test criteria can be modified in this case to allow consideration of these interim sample results as a means of coming to an early conclusion of whether an acceptance test passes or fails.  The general procedure would be to collect a predetermined number of samples and then evaluate the number of failures.  If at or below a certain number, c1A, conclude the acceptance test passes.  If at or above another number, c1R , which is greater than c1A, the acceptance test fails.  If the number of failures falls between c1A and c1R then complete another stage of sampling.  At the end of this second stage of sampling, compare the cumulative failures to a new set of acceptance limits to determine acceptance, failure, or continuation.  This process can continue for as many stages as desired before reaching a final stage, where the cumulative number of failures from all stages is compared to a single acceptance limit, cT.  If the cumulative total of failures from all stages is at or below the limit, the acceptance test passes.  Otherwise, it fails.

 

If the desired acceptance plan has two stages of sampling, the previous equations can be modified to account for a potential additional stage of sampling by

 

Two Equations which are are expanded versions of Equations 1 and 2.  They include additional terms to account for a two stage sample design where the conclusions of the first stage may be acceptance, failure, or continuation to a second stage.. The first equation is for the probability, a, of error in concluding system test failure.  The second equation is for the probability, ß, of error in failing to reject the initial assumption (i.e., concluding system test acceptance). 

 

 

A final solution that satisfies the equations above is n1=147, n2=147, c1A=1, c1R=5, cT=6 with p0=0.01, p1=0.04, α=0.06 and β=0.06.  The test is completed as:

  1. Select a sample of 147 items and count the number of failures.
  2. If the number of failures is 1 or less, there is not sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test passes acceptance.
  3. If the number of failures is 5 or more, there is sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test fails acceptance.
  4. If the number of failures is 2, 3, or 4, no immediate conclusion is reached about acceptance or failure.  Instead, another sample of 147 items is taken, the failures are counted, and these failures are added to the original stage failures.
    1. If the aggregate number of failures is 6 or less, there is not sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test passes acceptance.
    2. If the aggregate number of failures is 7 or more, there is sufficient evidence to disprove an initial hypothesis that the true system failure rate is less than one percent and the test fails acceptance.
  1. It is possible that an overall passing acceptance result from this test (regardless of whether it happens in the first or second stage of sampling) is incorrect (i.e., that the true system failure rate is greater than one percent).  However, if the true system failure rate is four percent or more, there is only a six percent chance of this error.  Above a four percent system failure rate, this probability is even lower.
  2. It is possible that an overall failing acceptance result from this test (regardless of whether it happens in the first or second stage of sampling) is incorrect (i.e., that the true system failure rate is one percent or less).  However, if the true system failure rate is one percent, there is only a six percent chance of this error.  Below a one percent system failure rate, this probability is even lower.

Relative to the single stage sample plan of 147 samples, this final sample plan provides the desired protection against the lower tolerable upper limit at four percent instead of five percent.  However, it could require two stages, or 294 samples to achieve this.  Relative to the single stage sample plan of 294 samples, this final sample plan provides the possibility of finishing sampling in half the total number of samples.  However, its α and β are not quite as good as the single stage sample plan.  Balancing these two considerations, the two-stage sample is recommended for this particular application.

 

 

Appendix B: Data Collected for Acceptance Test of FOT Voice Routing Functionality

Appendix B. Acceptance data from QA review of sample calls from August 10, 2005 through August 19, 2005.

 

 

Appendix C:Data Summaries for FOT Data Routing Performance

Week

Dates

Successful Call Deliveries

Total Failures

Average Delivery Time (Seconds)

AACN

CAN

SOS

19

2004.09.27-to-2004.10.03

0

4

21

0

0.76

20

2004.10.04-to-2004.10.10

0

4

31

0

0.91

21

2004.10.11-to-2004.10.17

0

1

32

0

1.00

22

2004.10.18-to-2004.10.24

1

2

37

0

0.80

23

2004.10.25-to-2004.10.31

0

11

27

0

0.79

24

2004.11.01-to-2004.11.07

0

5

19

0

0.71

25

2004.11.08-to-2004.11.14

0

6

40

0

0.85

26

2004.11.15-to-2004.11.21

1

2

40

0

0.77

27

2004.11.22-to-2004.11.28

0

7

21

0

23.25

28

2004.11.29-to-2004.12.05

1

0

21

0

0.64

29

2004.12.06-to-2004.12.12

 

 

 

 

 

30

2004.12.13-to-2004.12.19

0

0

36

0

0.89

31

2004.12.20-to-2004.12.26

0

5

33

0

0.84

32

2004.12.27-to-2005.01.02

1

6

32

0

0.82

33

2005.01.03-to-2005.01.09

0

7

26

0

0.91

34

2005.01.10-to-2005.01.16

0

2

36

0

0.71

35

2005.01.17-to-2005.01.23

0

5

36

0

0.80

36

2005.01.24-to-2005.01.30

1

4

22

0

0.93

37

2005.01.31-to-2005.02.06

0

0

37

0

0.81

38

2005.02.07-to-2005.02.13

0

3

21

0

0.96

39

2005.02.14-to-2005.02.20

0

3

26

0

0.86

40

2005.02.21-to-2005.02.27

0

1

29

0

0.77

41

2005.02.28-to-2005.03.06

0

0

25

0

0.84

42

2005.03.07-to-2005.03.13

1

0

30

0

0.81

43

2005.03.14-to-2005.03.20

0

5

22

0

0.96

44

2005.03.21-to-2005.03.27

1

2

19

0

0.91

45

2005.03.28-to-2005.04.03

0

1

27

0

0.75

46

2005.04.04-to-2005.04.10

0

1

30

0

0.84

47

2005.04.11-to-2005.04.17

1

2

25

0

0.79

48

2005.04.18-to-2005.04.24

0

0

15

14

0.80

49

2005.04.25-to-2005.05.01

0

0

0

36

0.00

50

2005.05.02-to-2005.05.08

 

 

 

 

 

51

2005.05.09-to-2005.05.15

 

 

 

 

 

52

2005.05.16-to-2005.05.22

1

2

36

0

0.82

53

2005.05.23-to-2005.05.29

0

1

21

0

0.91

54

2005.05.30-to-2005.06.05

3

5

40

0

0.90

55

2005.06.06-to-2005.06.12

 

 

 

 

 

56

2005.06.13-to-2005.06.19

 

 

 

 

 

57

2005.06.20-to-2005.06.26

0

3

38

0

0.83

58

2005.06.27-to-2005.07.03

0

4

32

0

1.64

59

2005.07.04-to-2005.07.10

0

4

41

0

0.76

60

2005.07.11-to-2005.07.17

1

2

22

0

0.84

61

2005.07.18-to-2005.07.24

1

15

23

0

0.79

62

2005.07.25-to-2005.07.31

0

11

30

0

0.85

63

2005.08.01-to-2005.08.07

3

1

46

0

0.78

64

2005.08.08-to-2005.08.14

 

 

 

 

 

65

2005.08.15-to-2005.08.21

0

2

0

0

1.00

66

2005.08.22-to-2005.08.28

0

1

0

0

1.00

67

2005.08.29-to-2005.09.04

0

5

0

0

0.60

 

Previous  | Table of Contents