DATA PROCESSING
Precision and Accuracy
There is some degree of uncertainty associated
with every physical measurement. In order to express the reliability of
measurements, the terms accuracy and precision are used.
Accuracy is a measure of the agreement of a measured value with
the true or accepted value. The true value is seldom known, except by definition
or agreement. Hence, precision is often a better measure of reliability.
Precision is the agreement of several measurements with each other
and is concerned with the reproducibility of measurements. Generally, it
can be assumed that better precision indicates a higher probability of
the date being accurate.
Significant Figures
A rough estimate of precision an be shown with
significant figures. These are all the figures which are known with
certainty. For example, suppose you are measuring temperature with a thermometer
calibrated in degrees centigrade. If the temperature is indicated to 21o,
we know that it is 21o, not 20o or 22o.
The uncertainty is generally assumed to be one-half of the last indicated
place. Thus 21o is actually 21 + 0.5o. With
a thermometer calibrated in hundredths of degrees the temperature might
be recorded as 21.37 + .005o.
Measurements may be expressed in different units.
For example, a length might be 0.026m or 2.6cm or 26mm.
When using different units, it is customary to express the number in exponential
or scientific notation. The digits express the number of significant figures,
while the exponent serves to locate the decimal point. We write 2.6cm
as 2.6x107 nm,
not as 26,000,000 nm.
Zeros may or may not be significant. In the number
0.026 the zeros are not significant. They only locate the decimal point.
But in the number 0.0260 the last zero is significant. In 260mm
the zero may or may not be significant. Again to avoid confusion, 260 should
be written as 2.6 x 102 or 2.60 x 102.
Errors
Errors associated with measurements are usually
divided into two types: systematic errors and random errors.
In principle, systematic errors can be corrected or eliminated, since they
involve errors in technique or method. Common systematic errors include
instrument errors (calibration or adjustment), reagent error (impurities
or change in concentration), and calculation errors (failure to use correct
equation or theory; these are not mathematical mistakes). In many cases
systematic errors are not detected and, thus, not corrected. Since most
systematic errors are reproducible, they effect the accuracy of the measurement,
but not the precision.
Random errors are those which can be reduced
with proper care, but which cannot be eliminated. Since random errors effect
reproducibility, they influence both precision and accuracy. Common random
errors include titration errors, proper draining of buret, removing last
drop from tip, and judging the correct color at the end point, weighing
errors, judging the level of liquid in volumetric glassware, and parallax
error in reading instruments and volumetric glassware.
Error can be expressed as an absolute error or
relative error. The absolute error can be defined as
Where O is the observed value and A
is the accepted value.
The relative error may be defined as
Relative error may be expressed as per cent,
100 Er, or parts per thousand, 1000 Er.
Statistical Analysis of Data
Measures of central tendency (mean, median, and
mode) serve as reference points for interpreting date. The purpose of measures
of central tendency is to show where the typical or central values lies
within a group.
The most common measure of central tendency are
the following:
1. Arithmetic mean - also referred to
as simple the mean or the average.
The mean, ,
of N values, Xj
is given by
2. Median - the midpoint of a distribution.
3. Mode - the most frequent value in a
distribution.
It is generally recognized that the mean is the
best measure of central tendency. However, if there are some extremely
high or low values in a distribution, it may be advisable to use the median.
In addition to knowing the central tendency of
a distribution, it is also necessary to know the spreading out or dispersion
of data. The most commonly used measures of dispersion are average deviation
and standard deviation.
The average deviation, a, is the average
of the deviations from the mean.
where dj
is the deviation,
The standard deviation,
s,
is the square root of the average of the squares of the deviation,
When N is small, less than 30, it is better
to use N-1 in computing the standard deviation. In large samples,
it makes little difference which one is used.
Computer Analysis
Statistical programs are available on
the computers in the DSU ITS computer labs. You
are encouraged to use these. However, if
the computer does your calculations, be sure you know what the computer
is doing.
|