Instrument manufacturers usually supply specifications for their equipment that define its accuracy, precision, resolution and sensitivity. Unfortunately, not all of these specifications are uniform from one to another or expressed in the same terms. Moreover, even when they are given, do you know how they apply to your system and to the variables you are measuring? Some specifications are given as worst-case values, while others take into consideration your actual measurements.

**Accuracy** can be defined as the amount of uncertainty in a measurement with respect to an absolute standard. Accuracy specifications usually contain the effect of errors due to gain and offset parameters. Offset errors can be given as a unit of measurement such as volts or ohms and are independent of the magnitude of the input signal being measured. An example might be given as ±1.0 millivolt (mV) offset error, regardless of the range or gain settings. In contrast, gain errors do depend on the magnitude of the input signal and are expressed as a percentage of the reading, such as ±0.1%. Total accuracy is therefore equal to the sum of the two: ±(0.1% of input +1.0 mV). An example of this is illustrated in Table 1.

Table 1. Readings as a function of accuracy

**Input Voltage ** | **Range of Readings within the Accuracy Specification ** |

0V | -1 mV to +1 mV |

5V | 4.994V to 5.006V (±6 mV) |

10V | 9.989V to 10.011V (±11 mV) |

conditions: input 0-10V, Accuracy = ±(0.1% of input + 1mV)

**Precision **describes the reproducibility of the measurement. For example, measure a steady state signal many times. In this case if the values are close together then it has a high degree of precision or repeatability. The values do not have to be the true values just grouped together. Take the average of the measurements and the difference is between it and the true value is accuracy.

**Resolution** can be expressed in two ways:

1. It is the ratio between the maximum signal measured to the smallest part that can be resolved - usually with an analog-to-digital (A/D) converter.

2. It is the degree to which a change can be theoretically detected, usually expressed as a number of bits. This relates the number of bits of resolution to the actual voltage measurements.

In order to determine the resolution of a system in terms of voltage, we have to make a few calculations. First, assume a measurement system capable of making measurements across a ±10V range (20V span) using a 16-bits A/D converter. Next, determine the smallest possible increment we can detect at 16 bits. That is, 2^{16} = 65,536, or 1 part in 65,536, so 20V÷65536 = 305 microvolt (uV) per A/D count. Therefore, the smallest theoretical change we can detect is 305 uV.

Unfortunately, other factors enter the equation to diminish the theoretical number of bits that can be used, such as noise. A data acquisition system specified to have a 16-bit resolution may also contain 16 counts of noise. Considering this noise, the 16 counts equal 4 bits (2^{4} = 16); therefore the 16 bits of resolution specified for the measurement system is diminished by four bits, so the A/D converter actually resolves only 12 bits, not 16 bits.

A technique called averaging can improve the resolution, but it sacrifices speed. Averaging reduces the noise by the square root of the number of samples, therefore it requires multiple readings to be added together and then divided by the total number of samples. For example, in a system with three bits of noise, 2^{3} = 8 , that is, eight counts of noise averaging 64 samples would reduce the noise contribution to one count, √64 = 8: 8÷8 = 1. However, this technique cannot reduce the affects of non-linearity, and the noise must have a Gaussian distribution.

**Sensitivity **is an absolute quantity, the smallest absolute amount of change that can be detected by a measurement. Consider a measurement device that has a ±1.0 volt input range and ±4 counts of noise, if the A/D converter resolution is 2^{12} the peak-to-peak sensitivity will be ±4 counts x (2 ÷ 4096) or ±1.9mV p-p. This will dictate how the sensor responds. For example, take a sensor that is rated for 1000 units with an output voltage of 0-1 volts (V). This means that at 1 volt the equivalent measurement is 1000 units or 1mV equals one unit. However the sensitivity is 1.9mV p-p so it will take two units before the input detects a change.

**Measurement Computing's USB-1608G Series** **Example**

Let’s use the USB-1608G and determine its resolution, accuracy, and sensitivity. (Refer to Table 2 and Table 3, below, for its specifications.) Consider a sensor that outputs a signal between 0 and 3 volts and is connected to the USB-1608G's analog input. We will determine the accuracy at two conditions: Condition No. 1 when the sensor output is 200 mV and Condition No. 2 when it is 3.0 volts.

**Accuracy: The USB-1608G uses a 16 bit A/D converter**

**Condition No. 1**: 200 mV measurement on a ±1 volt single-ended range

- Temperature = 25º C
- Resolution = 2V ÷ 2
^{16} = 30.5 uV - Sensitivity = 30.5 uV × 1.36 LSB rms = 41.5 uV rms
- Gain Error: 0.024% × 200mV = ±48uV
- Offset Error = ±245uV
- Linearity Error = 0.0076% of range = 760uV

- Total Error = 48uV + 245uV + 760uV = 1053uV

Therefore a 200 mV reading could fall within a range of 198.947 mV to 201.053 mV.

**Condition No. 2:** 3.0 V measurement on a ±5 volt single-ended range

- Temperature = 25º C
- Resolution =10 volts ÷ 2
^{16} = 152.6uV - Sensitivity = 152.6 uV × 0.91 LSB rms= 138.8 uV rms

- Gain Error: 0.024% × 3.0V = 720uV
- Offset Error = 686uV
- Linearity error = 0.0076% of range = 380uV
- Total Error = 720uV + 686uV + 380uV = 1.786mV

Therefore, a 3.0V reading could fall within a range of 2.9982 mV to 3.0018 mV.

**Summary Analysis:**

Accuracy: Consider Condition No. 1. The total accuracy is 369 uV ÷ 2 V × 100 = 0.0184%

Accuracy: Consider Condition No. 2. The total accuracy is 1.786 mV ÷ 10 V × 100 = 0.0177%

Effective Resolution: The USB-1608G has a specification of 16 bits of theoretical resolution. However the effective resolution is the ratio between the maximum signal being measured and the smallest voltage that can be resolved, i.e. the sensitivity. For example...if we consider Condition No. 2, divide the sensitivity value by the measured signal value or (138.5uV ÷ 3.0 V) = 46.5e^{-6} and then converting to the equivalent bit value produces (1V ÷ 46.5e^{-6}) = 21660 or 214.4 bits of effective resolution. To further improve on the effective resolution, consider averaging the values as previously discussed.

Sensitivity: The most sensitive measurement is made on the ±1 volt range where the noise is only 41.5uV rms whereas the sensitivity of the 5 volt range is 138.8uV rms. In general, when selecting a sensor, set the equipment to capture the highest output with the best sensitivity. For example, if the output signal is 0-3 volts select the 5 volt range instead of the 10 volt.

**Table 2. **

**Table 3. **