TEST UNCERTAINTY & TEST UNCERTAINTY RATIO (TUR) Samira A Khanam1, Edward Morse1 Department of Mechanical Engineering and Engineering Science The University of North Carolina at Charlotte Charlotte, NC 28223, USA Measurement is a process to determine the value or physical quantity of a measurand. The measurand is the particular quantity subject to measurement [1]. Each test of dimensional measurement contains one or more measurement steps. These steps contain inaccuracies or errors. To describe these errors the term “Uncertainty” is used. The ISO International Vocabulary of Basic and General Terms in Metrology (VIM) define uncertainties as “a parameter, associated with the result of a measurement that characterizes the dispersion of the values that could be reasonably attributed to the measurand (VIM 3.9).” In general, different measurements may have different uncertainty statements, even if they are measured with the same instrument [2]. Elimination of uncertainty is not the focus for measurement only the reduction of it is possible. The most common dimensional measurand is associated with the work piece. The example can be the measurand when measuring the diameter of the hole by using micrometer or any other instrument. The uncertainty associated with this kind of measurand is known as “Task specific uncertainty”. The uncertainty associated when testing a piece of equipment is a new concept in this field. As a result this discussion focuses on this new type of uncertainty which is known as “Test uncertainty”. The example can be to calibrate micrometer by using a gauge block or to calibrate CMM by using a step gauge. The uncertainty associated with this kind of measurand which is the “error of the instruments” is known as “Test Uncertainty”. This is mentioned in ISO 23165 in which test uncertainty is characterized for CMM. As stated above error is associated with the measurement process. The term error and the uncertainty are not exactly same. Error is the difference between the result and the true value. In practice, no one can know the true value. So the measured value is compared with the reference value. The reference value gives a point of reference with an uncertainty which is commonly accepted to compare an indicator. If calibrated at NIST the true value is inside that reference value. In [2] S.D Phillips et al states, “The relationship between the measured or indicated values and those of the reference values is a key issue with regards to calibration…all calibration must include the statement about the accuracy of the instrument or artifact as required by traceability.” We interpret this to mean test uncertainty in this discussion. Any calibration result showed the error or the deviation of the measurand with respect to the reference value and uncertainty associated with this error. There are different kinds of measurand for different kinds of measurement process. Consequently the uncertainty will be different also. A clear description of different kinds of uncertainty associated with different kind of measurand is explained in table-1. TABLE 1. Measurand/ Quantity Activity Uncertainty (U) Some characteristics of the work piece The length of the artifact Measuring a work piece to conform to specific value or tolerance Task specific U Calibrating the material standard size. ( Diameter of a sphere) Calibrating an instrument Calibration U The error of the instrument (E-value) Test U Table 2 is showing the uncertainty contributors for different kinds of Uncertainty. Example of “Task specific uncertainty” & “Test uncertainty” TABLE 2. Contributor Environment 1)Temperature Task Specific Uncertainty Test Uncertainty 1) If instrument does not compensate with temperature it has effect on TSU 1) Any error introduced by the instrument is not the part of TU 2) The temp. difference between W/P and room temp. 2) If artifact is compensate with temperature of instrument, then it is part of instrument, not include in TU Does not apply ( Part of the instrument) 2)Work Piece Reference ele ment of measurement equipment Resolution of the main scale ( analogue or digital) Measurement setup ( Probe selection, tip size etc.) Form deviation of tip, Offset, extension Usually specified by std. Poor setup may influence TU. Software and calculations Rounding/Quantific ation Sampling Algorithms Experience, training, knowledge ( precision) Surface roughness, form deviation of the w/p Well defined in the standard, may be eliminated Reproducibility Alignment, Clamping fixturing, Number of measurement etc. example: Due to the measurement process of w/p like repeatability Alignment, Clamping fixturing , number of measurement etc. example: Instrument repeatability error is not the part of TU, but the repeatability error for tester is the part of TU. Metrologist Measurement object , work piece or measuring instrument characteristics Measuring procedure CMM (Test equipment) Task: To measure true position of the hole FIGURE 1. CMM is measuring the true position of the hole of a block (Task specific uncertainty) CMM (Instrument) Step Gauge (Test equipment) Form error of artifact, uncertainty for the length of artifact FIGURE 2. CMM is calibrating by using a step gauge (Test uncertainty) Test uncertainty depends on significantly test equipment, test procedure (which should be well recognized) and the influence of the human operator. When calculating the test uncertainty all these factors should be reasonably estimated to ensure that the result does not give any wrong impression of uncertainty. Test uncertainty is only the indication of the quality of the test, it is not the machine performance. By defining the test, how well test is performed, the influence of the operator, selection and placement of test instrument test uncertainty can be decreased and consequently increasing the precision and usefulness of the test. Another topic, Test Uncertainty Ratio (TUR) is a measure of the ability of a particular measurement instrument and/or process to evaluate conformance to specification. TUR is the ratio between the tolerance or specification and the uncertainty present in the test of this tolerance or specification. Historically, the rule of thumb for an appropriate ratio was that the TUR must be at least 10:1. The higher the ratio, the better the performance of the test. Currently, a ratio of 4:1 or even 3:1 is considered acceptable. This is due mostly to the better performance of manufacturing equipment, and the tighter and tighter specifications on manufactured components. In many cases, test equipment with an uncertainty small enough for a 10:1 TUR does not exist, or is prohibitively expensive for the application. There are two main applications of the TUR: the first is in the calibration of measuring instruments and equipment, the second is in the inspection of manufactured components. The focus of this document is the second application – how do we determine the Test Uncertainty Ratio for a part that we need to measure using a particular gage? The list below gives some of the important things to consider. 1. The uncertainty statement in the gage's product literature might not be the uncertainty needed to calculate TUR. 2. The result of the gage's most recent calibration is almost certainly not the uncertainty needed to calculate TUR. 3. The tolerance value on the part drawing is – if interpreted correctly – going to be needed to calculate TUR. 4. There will be more than one TUR calculation for a part if there is more than one tolerance that must be inspected. To preview what will be covered in this paper, the task-specific measurement uncertainty must be estimated for the measurand in question for each tolerance. This is the "1" value. The range of allowable values for the measurand in question (this is usually the tolerance) must be known, and is compared to "4". TUR = Tolerance for the measurand Task ⋅ specific. uncertaint y As is easily inferred from the equation above, the TUR ratio compares the allowable variation for the measurand (the numerator) with the variability associated with finding the measurand (the denominator). LSL USL Specification zone Out of Specification Out of Specification Measur ement U increas Non confor mance zone U Conform ance Zone U Non confor mance zone FIGURE 3. Relationship between Specification Zone and Conformance Zone One function of industrial measurement is to determine whether a particular part measurand (length, form, location, etc.) conforms to the specification given on the part drawing or model. The difference between the specification zone and the conformance zone for a measurand is explained with the aid of FIGURE 3, taken from ISO 14253-1:1998(E). The upper horizontal line of the figure shows the specification zone, which is bounded by the lower specification limit (LSL) and the upper specification limit (USL). If the "true value" of the measurand is within the specification zone, then the specification is satisfied, otherwise the measurand is out of specification. However, we can never know the "true value" of the measurand. In order to state whether we believe that the measurand is in or out of specification, we have to acknowledge the existence of uncertainty in the measurement process. This is shown in the lower horizontal line in FIGURE 3. If the measurand is in conformance zone, we are suitably confident that the true value is in specification. Similarly, if the measurand is in the non-conformance zone, we are confident that the true value is out of specification. For the uncertainty region shown between conformance and non-conformance, we need to apply the "Decision rules." A decision rule is a method – agreed on by two parties – to decide whether to accept or reject a part when the measurement value lies in this uncertainty region. One common decision rule used in industry is "Simple Acceptance and Rejection Using an N:1 Decision Rule" as defined in ASME B89.7.3.1-2001. Using this decision rule, the measured value is compared directly to the specification zone (i.e. the conformance zone and the specification zone are identical). This rule has the effect of dividing the risk for accepting bad parts and rejecting good parts between the supplier and customer. In order to limit this risk, the requirement of N:1 is placed on the test uncertainty ratio (N is often 4, as mentioned already). In FIGURE 4 below, taken from the B89.7.3.1 document, the acceptance and rejection zones are shown to be identical to the specification zones. FIGURE 4. Schematic view of Simple acceptance/rejection with a 4:1 uncertainty requirement The measurement uncertainty interval is of width 2U, where U is expanded uncertainty, and the uncertainty interval is no larger than one-fourth the product’s specification zone for N = 4. [B 89.7.3.1-2001] For the simple acceptance rule with a 4:1 ratio means that the tolerance range is at least 4 times the uncertainty interval. REFERENCES [1] http://www.measurementuncertainty.org/mu /guide/uncertainty.html [2] S.D Phillips et al. A Careful Consideration of the Calibration Concept. Journal of Research of the National Institute of Standards and Technology. 2001; 106:371379. [3] Adams M. A2LA Guide for the Estimation of the Measurement Uncertainty in Testing. American Association of Laboratory Accreditation Manual, 2002. [4] http://en.wikipedia.org/wiki/Measurements. [5] Guidance Uncertainty of Measurement concept in European Standards.
© Copyright 2026 Paperzz