Errors and Uncertainties

Reliability estimates of measurements greatly enhance their value. Thus, saying that the average diameter of a cylinder is 10.00$ \pm$0.02 mm tells much more than the statement that the cylinder is a centimeter in diameter.

To physicists the term ``error" is interchangeable with ``uncertainty" and does not have the same meaning as ``mistake". Mistakes, such as ``errors" in calculations, should be corrected before estimating the experimental error. In estimating the reliability of a single quantity (such as the diameter of a cylinder) we recognize several different kinds and sources of error:

FIRST, are actual variations of the quantity being measured, e.g. the diameter of a cylinder may actually be different in different places. You must then specify where the measurement was made; or if one wants the diameter in order to calculate the volume, first find the average diameter by means of a number of measurements at carefully selected places. Then the scatter of the measurements will give a first estimate of the reliability of the average diameter.

SECOND, the micrometer caliper used may itself be in error. The errors thus introduced will of course not lie equally on both sides of the true value so that averaging a large number of readings is no help. To eliminate (or at least reduce) such errors, we calibrate the measuring instrument: in the case of the micrometer caliper by taking the zero error (the reading when the jaws are closed) and the readings on selected precision gauges of dimensions approximately equal to those of the cylinder to be measured. We call such errors systematic, and these cause errors on accuracy.

THIRD, Another type of systematic error can occur in the measurement of a cylinder: The micrometer will always measure the largest diameter between its jaws; hence if there are small bumps or depressions on the cylinder, the average of a large number of measurements will not give the true average diameter but a quantity somewhat larger. (This error can of course be reduced by making the jaws of the caliper smaller in cross section.)

FINALLY, if one measures something of definite size with a calibrated instrument, errors of measurement still exist which (one hopes) are as often positive as negative and hence will average out in a large number of trials. Such errors are called random, and result in less precision. For example, the reading of the micrometer caliper may vary because one can't close it with the same force every time. Also the observer's estimate of the fraction of the smallest division varies from trial to trial. Hence the average of a number of these measurements should be closer to the true value than any one measurement. Also the deviations of the individual measurements from the average give an indication of the reliability of that average value. The typical value of this deviation is a measure of the precision. This average deviation has to be calculated from the absolute values of the deviations, since otherwise the fact that there are both positive and negative deviations means that they will cancel. If one finds the average of the absolute values of the deviations, this ``average deviation from the mean" may serve as a measure of reliability. For example, let column 1 represent 10 readings of the diameter of a cylinder taken at one place so that variations in the cylinder do not come into consideration, then column 2 gives the magnitude (absolute) of each reading's deviation from the mean.

Measurements Deviation from Ave.
      9.943 mm 0.000
9.942 0.001
9.944 0.001
9.941 0.002
9.943 0.000
9.943 0.000
9.945 0.002 Diameter =
9.943 0.000
9.941 0.002 9.943$ \pm$0.001 mm
$ \underline{{~9.942~}}$ $ \underline{{~0.001~}}$
Ave = 9.943 mm                     Ave = 0.0009 mm$ \approx$0.001 mm

Expressed algebraically, the average deviation from the mean is = ($ \sum$| xi - $ \bar{{x}}$|)/n), where xi is the ith measurement of n taken, and $ \bar{x}$ is the mean or arithmetic average of the readings.

Standard Deviation:

The average deviation shown above is a measure of the spread in a set of measurements. A more easily calculated version of this is the standard deviation $ \sigma$ (or root mean square deviation). You calculate $ \sigma$ by evaluating

$\displaystyle \sigma$ = $\displaystyle \sqrt{{\frac{1}{n}\sum_{i=1}^{n}(x_{i}-\overline{x})^{2}}}$

where $ \overline{{x}}$ is the mean or arithmetical average of the set of n measurements and xi is the ith measurement.

Because of the square, the standard deviation $ \sigma$ weights large deviations more heavily than the average deviation and thus gives a less optimistic estimate of the reliability. In fact, for subtle reasons involving degrees of freedom, $ \sigma$ is really

$\displaystyle \sigma$$\displaystyle \sqrt{{\frac{1}{(n-1)}\sum_{i=1}^n (x_i-\bar{x})^2}}$

$ \sigma$ tells you the typical deviation from the mean you will find for an individual measurement. The mean $ \bar{{x}}$ itself should be more reliable. That is, if you did several sets of n measurements, the typical means from different sets will be closer to each other than the individual measurements within a set. In other words, the uncertainty in the mean should be less than $ \sigma$. It turns out to reduce like 1/$ \sqrt{{n}}$, and is called the error in the mean $ \bf\sigma_\mu$:

$\displaystyle \sigma_{\mu}^{}$ = error in mean = $\displaystyle {\frac{{\sigma}}{{\sqrt{n}}}}$ = $\displaystyle {\frac{{1}}{{\sqrt{n}}}}$ $\displaystyle \sqrt{{\frac{1}{n-1}\sum_{i=1}^{n} (x_i-\bar{x})^2}}$

For an explanation of the (n - 1) factor and a clear discussion of erros, see P.R. Bevington and D.K Robinson, Data Reduction and Error Analysis for the Physical Sciences, McGraw Hill 1992, p. 11.

If the error distribution is ``normal" (i.e. the errors, $ \epsilon$ have a Gaussian distribution, e-$\scriptstyle \epsilon^{{2}}$, about zero), then on average 68% of a large number of measurements will lie closer than $ \sigma$ to the true value. While few measurement sets have precisely a ``normal" distribution, the main differences tend to be in the tails of the distributions. If the set of trial measurements are generally bell shaped in the central regions, the ``normal" approximation generally suffices.

Relative error and percentage error:
Let $ \epsilon$ be the error in a measurement whose value is a. Then ($ {\frac{{\epsilon}}{{a}}}$) is the relative error of the measurement, and 100 ($ {\frac{{\epsilon}}{{a}}}$)% is the percentage error. These terms are useful in laboratory work.

SYSTEMATIC ERRORS IN THE LABORATORY STANDARDS OF
LENGTH, TIME AND MASS
For the experiments in this manual these systematic errors are usually negligible compared to other uncertainties. An exception sometimes occurs for the larger masses especially the 100 gram, the 500 gram, and 1 kg masses. Some contain drilled holes into which lead shot and a plug have been added to adjust the mass to within tolerance (typically 1.000$ \pm$0.003 kg). Occasionally a plug works loose and the calibration lead shot is lost. You can check the assigned mass values by weighing them on the triple beam balances. Report any deviations greater than 0.4% to the instructor.

UNCERTAINTY ESTIMATE FOR A RESULT INVOLVING MEASUREMENTS OF
SEVERAL INDEPENDENT QUANTITIES

A.) If the desired result is the sum or difference of two measurements, the ABSOLUTE uncertainties ADD:
Let $ \Delta$x and $ \Delta$y be the errors in x and y respectively. For the sum we have z = x + $ \Delta$x + y + $ \Delta$y = x + y + $ \Delta$x + $ \Delta$y and the relative error is $ {\frac{{\Delta x + \Delta y}}{{x + y}}}$. Since the signs of $ \Delta$x and $ \Delta$y can be opposite, adding the absolute values gives a pessimistic estimate of the uncertainty. If errors have a normal or Gaussian distribution and are independent, they combine in quadrature, i.e. the square root of the sum of the squares, i.e.,

$\displaystyle \Delta$z = $\displaystyle \sqrt{{\Delta x^{2}+\Delta y^{2}}}$

For the difference of two measurements we obtain a relative error of $ {\frac{{\Delta x + \Delta y}}{{x - y}}}$. which becomes very large if x is nearly equal to y. Hence avoid, if possible, designing an experiment where one measures two large quantities and takes their difference to obtain the desired quantity.

B.) If the desired result involves multiplying (or dividing) measured quantities, then the RELATIVE uncertainty of the result is the SUM of the RELATIVE errors in each of the measured quantities.

Proof:

Let   z = $\displaystyle {\frac{{x_1~x_2~x_3 .....}}{{y_1~y_2~y_3 .....}}}$   and hence *2.1in

lnz = lnx1 + lnx2 + lnx3 + ..... - lny1 - lny2 - lny3 - .....

Then find the differential, d (lnz):

d (lnz) = $\displaystyle {\frac{{dz}}{{z}}}$ = $\displaystyle {\frac{{dx_1}}{{x_1}}}$ + $\displaystyle {\frac{{dx_2}}{{x_2}}}$ + $\displaystyle {\frac{{dx_3}}{{x_3}}}$ + ..... - $\displaystyle {\frac{{dy_1}}{{y_1}}}$ + $\displaystyle {\frac{{dy_2}}{{y_2}}}$ + $\displaystyle {\frac{{dy_3}}{{y_3}}}$ - .....

Consider finite differentials, $ \Delta$z, etc. and note that the most pessimistic case corresponds to adding the absolute value of each term since $ \Delta$xi and $ \Delta$yi can be of either sign. Thus

$\displaystyle {\frac{{\Delta z}}{{z}}}$ = $\displaystyle \sum_{{i}}^{}$($\displaystyle {\frac{{\Delta x}}{{x}}}$) + $\displaystyle \sum_{{i}}^{}$($\displaystyle {\frac{{\Delta y}}{{y}}}$)

Again, if the measurement errors are independent and have a Gaussian distribution, the relative errors will add in quadrature:

$\displaystyle {\frac{{\Delta z}}{{z}}}$ = $\displaystyle \sqrt{{\sum_{i}(\frac{\Delta x}{x})^{2}+\sum_{i}(\frac{\Delta y}{y})^{2}}}$

C.) Corollary: If the desired result is a POWER of the measured quantity, the RELATIVE ERROR in the result is the relative error in the measured quantity MULTIPLIED by the POWER: Thus z = xn and

$\displaystyle {\frac{{\Delta z}}{{z}}}$ = n$\displaystyle {\frac{{\Delta x}}{{x}}}$.

The above results also follow in more general form: Let R = f (x, y, z) be the functional relationship between three measurements and the desired result. If one differentiates R, then

dR = $\displaystyle {\frac{{\partial{f}}}{{\partial{x}}}}$dx + $\displaystyle {\frac{{\partial{f}}}{{\partial{y}}}}$dy + $\displaystyle {\frac{{\partial{f}}}{{\partial{z}}}}$dz

gives the uncertainty in R when the uncertainties dx, dy and dz are known.

For example, consider the density of a solid (Exp. M1). The relation is

$\displaystyle \rho$  = $\displaystyle {\frac{{m}}{{\pi r^2 L}}}$

where m = mass, r = radius, L = length, are the three measured quantities and $ \rho$ = density. Hence

$\displaystyle {\frac{{\partial{\rho}}}{{\partial{m}}}}$ = $\displaystyle {\frac{{1}}{{\pi r^2L}}}$           $\displaystyle {\frac{{\partial{\rho}}}{{\partial{r}}}}$ = $\displaystyle {\frac{{-2m}}{{\pi r^3L}}}$           $\displaystyle {\frac{{\partial{\rho}}}{{\partial{L}}}}$ = $\displaystyle {\frac{{-m}}{{\pi r^2L^2}}}$

and so

d$\displaystyle \rho$ = $\displaystyle {\frac{{1}}{{\pi r^2L}}}$ dm$\displaystyle {\frac{{-2m}}{{\pi r^3L}}}$ dr$\displaystyle {\frac{{-m}}{{\pi r^2L^2}}}$ dL.

To get the relative error divide by $ \rho$ = m/$ \pi$r2L. The result, if one drops the negative signs, is

$\displaystyle {\frac{{d\rho}}{{\rho}}}$ = $\displaystyle {\frac{{dm}}{{m}}}$ +2$\displaystyle {\frac{{dr}}{{r}}}$ + $\displaystyle {\frac{{dL}}{{L}}}$

and represents a worst possible combination of errors. For small increments:

$\displaystyle {\frac{{\Delta\rho}}{{\rho}}}$ = $\displaystyle {\frac{{\Delta m}}{{m}}}$ +2$\displaystyle {\frac{{\Delta r}}{{r}}}$ + $\displaystyle {\frac{{\Delta L}}{{L}}}$

and

$\displaystyle \Delta$$\displaystyle \rho$ = $\displaystyle \rho$$\displaystyle \left[\vphantom{ \frac{\Delta m}{m} + 2 \frac{\Delta r}{r}
+ \frac{\Delta L}{L} }\right.$$\displaystyle {\frac{{\Delta m}}{{m}}}$ +2$\displaystyle {\frac{{\Delta r}}{{r}}}$ + $\displaystyle {\frac{{\Delta L}}{{L}}}$$\displaystyle \left.\vphantom{ \frac{\Delta m}{m} + 2 \frac{\Delta r}{r}
+ \frac{\Delta L}{L} }\right]$

Again if the errors have normal distribution, then

$\displaystyle {\frac{{\Delta\rho}}{{\rho}}}$ = $\displaystyle \sqrt{{\left(\frac{\Delta m}{m}\right)^2
+ \left(2\frac{\Delta r}{r}\right)^2 + \left(\frac{\Delta L}{L}\right)^2}}$

SIGNIFICANT FIGURES

Suppose you have measured the diameter of a circular disc and wish to compute its area A = $ \pi$d2/4 = $ \pi$r2. Let the average value of the diameter be 24.326$ \pm$0.003 mm ; dividing d by 2 to get r we obtain 12.163$ \pm$0.0015 mm with a relative error $ {\frac{{\Delta r}}{{r}}}$ of $ {\frac{{0.0015}}{{12}}}$ = 0.00012. Squaring r (using a calculator) we have r2 = 147.938569, with a relative error 2$ \Delta$r/r = 0.00024, or an absolute error in r2 of 0.00024 x 147.93 ... = 0.036 $ \approx$ 0.04. Thus we can write r2 = 147.94$ \pm$0.04, any additional places in r2 being unreliable. Hence for this example the first five figures are called significant.

Now in computing the area A = $ \pi$r2 how many digits of $ \pi$ must be used? A pocket calculator with $ \pi$ = 3.141592654 gives

A = $\displaystyle \pi$r2 = $\displaystyle \pi$ x (147.94$\displaystyle \pm$0.04) = 464.77$\displaystyle \pm$0.11 mm2

Note that $ {\frac{{\Delta A}}{{A}}}$ = 2$ {\frac{{\Delta r}}{{r}}}$ = 0.00024. Note also that the same answer results from $ \pi$ = 3.1416, but that $ \pi$ = 3.142 gives A = 464.83$ \pm$0.11 mm2 which differs from the correct value by 0.06 mm2, an amount comparable to the estimated uncertainty.

A good rule is to use one more digit in constants than is available in your measurements, and to save one more digit in computations than the number of significant figures in the data. When you use a calculator you usually get many more digits than you need. Therefore at the end, be sure to round off the final answer to display the correct number of significant figures.

SAMPLE QUESTIONS

  1. How many significant figures are there in the following number?
    1. 976.45
    2. 4.000
    3. 10
  2. Round off each of the following numbers to three significant figures.
    1. 4.455
    2. 4.6675
    3. 2.045

  3. A function has the relationship Z(A, B) = A + B3 where A and B are found to have uncertainties of $ \pm$$ \Delta$A and $ \pm$$ \Delta$B respectively. Find $ \Delta$Z in term of A, B and the respective uncertainties assuming the errors are uncorrelated.

  4. What happens to $ \sigma$, the standard deviation, as you make more and more measurements? what happens to $ \overline{{\sigma}}$, the standard deviation of the mean?
    1. They both remain same
    2. They both decrease
    3. $ \sigma$ increases and $ \overline{{\sigma}}$ decreases
    4. $ \sigma$ approachs a constant and $ \overline{{\sigma}}$ decreases


Working Demo

The htmlscript Quiz System working demo is now available.


Suggestions on Form for Lab Notebooks:

NUMBER AND TITLE (e.g. E1. ELECTROSTATICS)

Date performed:____________                               Partner:_________________________

Subdivisions: If appropriate, name and number each section as in the manual.

DATA:

Label numbers and give units. In a few words, state what quantities you measured. If appropriate, record the data in tabular form. Label the tables and give units.
CALCULATIONS:
State the equations used and present a sample calculation. (Inclusion of the arithmetic is not necessary.)
CONCLUSIONS:
If any important conclusions follow from the experiments, state them and show by a brief statement how they follow. Compare your results with accepted values if the experiment has involved the measurements of a physical constant.

Errors:

Some of your experiments will be qualitative while others will involve quantitative measurements of physical constants. Where it is appropriate, estimate the uncertainty of each measurement used in a calculation and compute the uncertainty of the result. Does your estimate of uncertainty indicate satisfactory agreement between your result and the accepted one (or between your several values if you have several)? Intelligent discussion is welcomed, but don't make this section a burden on you.


Michael Winokur 2005-08-30