Quick links on this page: | Introduction | Micrometer | vernier | Toolbar | Density |
OBJECTIVES:
To learn about systematic and random errors; to understand
significant figures; to estimate the reliability of one's measurements;
and to calculate the reliability of the final result.
Note: This experiment illustrates the earlier sections on Errors
and Significant Figures. The actual density of the metal is incidental.
However, the accuracy of your estimate of reliability will
show whether you have mastered the material in the earlier sections.
APPARATUS:
Metal cylinders of varying sizes, micrometer and vernier
calipers, precision gauge blocks, precision balance.
You should be able to complete this brief quiz before proceding.
PRECAUTIONS:
Avoid dropping or deforming in any other way the metal
cylinders. Avoid damage to the precision screw of the micrometer by
turning only the friction head to open or close the caliper jaws?
Be sure
to disengage the caliper lock before using. (The caliper lock lets you
preserve a reading.)
Improper weighing procedures may damage the precision balance.
Consult
your instructor if in doubt. In handling the gauge
blocks avoid touching the polished surfaces since body acids are corrosive.
INTRODUCTION:
First read the material on Errors and
Significant Figures, p. 8-12. Since density is the mass per unit
volume, you must measure the mass (on a balance) and
compute the volume ()
from measurements of the cylinder's dimensions where d is the diameter and
h is the height.
Any one of three length measuring devices may be used. These include a
micrometer, a vernier caliper and/or a simple metric ruler. The micrometer
will permit the highest
precision measurments but using one can be cumbersome, especially when reading the
vernier scale. All methods will demonstrate the aforementioned objectives.
Your instructor will give you guidance in choosing an appropriate measurement
device.
SUGGESTIONS FOR THE MICROMETER:
Record serial number of your micrometer caliper. Then
familiarize yourself with the operation of the caliper and the reading of
the scales. If you have trouble with the principle and operation of the
micrometer caliper, ask your instructor for help or refer to the
appendices.
Note that use of the ``friction'' head in closing the jaws insures
the same pressure on the measured object each time. Always estimate
tenths of the smallest division. Some micrometers have verniers to assist
the estimation.
SUGGESTIONS FOR THE VERNIER CALIPER:
If you are not familiar with the principle of the
vernier
(invented 1631 by
Pierre Vernier), ask the instructor to explain it or try this
java vernier
.
Experiment with one of the
large verniers in the lab until you are sure you understand it.
Note that
verniers need not be decimal: for many inch scales the vernier estimates
1/8's of the 1/16 inch division, i.e. 1/128's of an inch. However
vernier calipers divide the inch into 50 divisions and the vernier estimates
1/25 of the 1/50 inch divisions, i.e. 1/1000 inch or 1 mil.
Precautions on use of the calipers:
Open caliper to within a few mm of the dimension being measured.
Close right thumbscrew to lock position of lower horizontal knurled
cylinder which executes fine motions of caliper jaw.
Never over tighten!
CALIBRATION OF THE MICROMETER (or VERNIER, ETC.):
Measure one or four (OPTIONAL) calibration gauge blocks (6, 12, 18, and 24 mm): Set the gauge blocks on end, well-in from the edge of the table, and thus freeing both hands to handle the caliper. Record the actual (uncorrected) readings and take several readings (3 to 5) of each gauge block used. If measuring only a single block choose one with a thickness intermediate to the diameter and height of the metal cylinder.
If using a single precision block obtain a correction factor you will use to minimize systematic measurement errors. Otherwise, plot a correction curve for your micrometer, i.e. plot errors as ordinates and nominal blocks sizes (0, 6, 12, 18, and 24 mm) as abscissa. Normally the correction will not vary from block to block by more than 0.003 mm (for the micrometer). If it is larger, consult your instructor.
DENSITY DETERMINATION:
Calculate the average length and average diameter reading, and then use your correction value (curve) to correct these average readings. If you were to use the uncorrected values, how much relative error would this introduce?
Calculate the standard deviation for your length and diameter averages.
Weigh the cylinder a few times on the electronic balance; estimate to 0.1 mg.
From the average dimensions and the mass, calculate the volume and density. Make a quantitative estimate of the uncertainty in the density. You should use, as your starting point:
*.3in
Compare the density with the tabulated value. Tabulated values are averages over samples whose densities vary slightly depending upon how the material was cast and worked; also on impurity concentrations.
In your notebook or lab form summarize the data and results. Also record your result on the blackboard. Is the distribution of blackboard values reasonable, i.e. ``normal'' distribution?
OPTIONAL: To test how accurately you can estimate a fraction of a
division, estimate the fractions on the vernier caliper before reading the
vernier. Record both your estimate and the vernier reading.
Related facts and URL links:
Question: Why are there are exactly 25.4 mm in 1 inch?
Answer(not verified): The Treaty of the Meter (Convention du Metre) in the
late 19th century
established the first centralized international system of metrology. This
defined the meter.
In 1959, the countries of the world that were using Imperial units defined them
uniformly based on the metric units. The inch was simply defined that way and
agreed to by all. Before 1959, different countries related inches to meters in
other ways. Among them was the United States. The Metric Act of 1866 defined
the meter in terms of inches (i.e., before the Treaty of the Meter), and that
relationship had continued to been used even after the Treaty fixed the length
of the meter.
Changing the definition in the U.S. in 1959 caused very little problem,
except for the U.S. Geological Survey. When you deal with things 105
meters big (like the sizes of the states), even 1 part in 105 changes
affect the specifications of boundary lines in significant ways. So, to
this day the U.S. has two different systems of inches, feet, and yards (in
the ratio of 36:3:1, for both). There's the usual inch, foot, and yard;
and there's the survey foot (and inch and yard), which is based on the
pre-1959 definition. (The ambiguity goes away, of course, when metric
specification is used. Newer USGS maps are metric.)
Links to metrology site(s):
National Institute of Science &
Technology