Appendix 5

 

ERROR PROPAGATION & COMPARISON

 

            A typical pattern for a sophomore physics experiment is to calculate a final result from measurements of several interacting physical quantities.  This result is then compared to what is expected from theory (or another experimental method).  If the two results differ, the theory is not necessarily faulty.  Remember, there is no real disagreement if the difference between the two results is no larger than what might be expected from the probable random or systematic errors in the measurements.

 

            The question at hand then is "How do you combine the uncertainties to give the uncertainty in a final result?"  In science combining uncertainties is called propagating errors. (That's incredibly unscientific, it should be called "propagating uncertainties).   The goal of error propagation is to predict how large an uncertainty can be expected in a result calculated from measurements which possess intrinsic uncertainties.

 

            You might be tempted to guess that uncertainties add algebraically every time you combine numbers that have uncertainties.  For example, if you wanted to add two lengths, l1 = 50.3 + 0.2 cm and l2 = 22.6 + 0.1 cm, you might well guess that the total length is: 1 = l1 + l2 = 72.9 + 0.3 cm.  This is NOT correct.  However it is a good approximation of a more rigorous method of propagating uncertainties.

 

I.  Sketchy Derivation of the Error in a Final Result

 

            This more rigorous method is derived by assuming that the uncertainty in the final result is just the sum of each individual quantity's uncertainty weighted by the amount each individual quantity affects the final result.  The amount each quantity affects the final result is the partial derivative of the final result with respect to that given quantity.  Using the above example of two lengths this translates into:

 

In general, consider the final result as a quantity, f, and assume f depends upon two independent, measured quantities, x and y.  You may then write f as a function of these two quantities: f = f(x,y).  In these more general terms the above propagated uncertainty becomes;

 

However, a second simplifying assumption can be made.  To understand this, first square the above equation to get:

 

   sf2 = [¶f/¶x]2sx2 + [¶f/¶y]2sy2 + 2(¶f/¶x)(¶f/¶y)sxsy

 

The last term in the above equation is often called the correlation term because it represents the degree to which any fluctuations in both x and y affect sf.  For example, this correlation term will increase if x and y vary so that sx, sy and f all increase simultaneously.  These sort of correlations between two independent, measured quantities (x and y) are rare.  Hence this correlation term will be neglected.  This yields a useful equation for calculating the propagated error in a function f.

 

sf2 = [¶f/¶x]2sx2 + [¶f/¶y]2sy2                          (5.1)

 

When more than two measured quantities go into a calculation to produce a final result, then the above equation is expanded to take into account the other terms.  For example, if f = f(x,y,z), then the above equation becomes:

 

sf2 = [¶f/¶x]2sx2 + [¶f/¶y]2sy2 + [¶f/¶z]2sz2           (5.2)

 

Equation (5.l) is useful, only if the individual uncertainties are known.  In practice, these sx's and sy's are either estimated uncertainties from a few measurements of x and y OR they are the standard deviations of the means of x and y.

 

            You are expected to use this relation (or the consequent relations that follow) in your Data Analysis section in all experiments in which you cannot make a good case for a less sophisticated comparison of results.  However, sometimes a full scale error propagation is NOT needed.  Only experience will allow you to determine which is the correct path to follow.

 

            It is usually more convenient to use the equations in the following table which have been derived from equation (5.l) for several of the most frequently occurring functions. To show you how the following table of error propagation formulas have been derived, several examples are first presented.

 

II.  Error Propagation For Some Common Functions

 

  1. Addition and Subtraction.

If f = x ± y, then ¶f/¶x = 1 and ¶f/¶y = ±1.  Since these partial derivatives are squared, the ± sign has no effect, so equation (5.1) yields:

sf2 = sx2 + sy2

 

The rule for addition and subtraction is therefore that uncertainties add "in quadrature", (which is a fancy word for the square root of the sum of the squares of the individual uncertainties), instead of simply adding uncertainties.

 

2.  Multiplication and Division.

 

            If f = xy, then ¶f/¶x = y, ¶f/¶y = x, so equation (5.1) yields

 

      sf2 = y2sx2 + x2sy2.

 

This is usually put into a more convenient form by dividing through by f2 = x2y2 to obtain

 

[sf/f]2 = [sx/x]2 + [sy/y]2,                   (5.3)

 

whose square root gives the uncertainty as a fractional part of f, a most useful form.  For f = x/y, exactly the same equation is found for sf/f  !

 

3.  Powers.

 

Let f = axn, where a is a known constant and x is the variable.  Then [¶f/¶x]2 = [naxn-1]2 = [nf/x]2, so equation (5.1)  yields

 

sf/f = n[sx/x].

 


III  A Table of Error Propagation Formulas

 

      Given an arbitrary function (f), two independent variables (x and y), two constants (a and b) and an integer n, the following formulas may be used to compute the propagated error (uncertainty) in the function f.  The uncertainties for x and y (sx and sy) can be either standard deviations or estimated uncertainties.  If they are standard deviations, then x and y must be those corresponding

means.

 

f(x,y)                           sf

-------------------------------------------------------------

x ± y                                              1.         [sx2+ sy2]1/2

 

xy  or x/y                           2.         xy[(sx/x)2 + (sy/y)2]1/2

 

x2                                       3.         2x2[sx/x] = 2x[sx]

 

axn                                     4.         naxn[sx/x] = u figger it.

 

a(x/y2)                                           5.         a(x/y2)[(sx/x)2 + (sy/y)2]1/2

 

aebx                                                6.         abebx[sx]

 

aln(bx)                                           7.         [a/x]sx

 

IV.  Words To The Wise

 

     Calculating standard deviations is generally the best way to evaluate uncertainties, however, you must exercise judgement as to when data warrant such calculations.  Sometimes you may be able to decide upon a valid uncertainty by "eyeballing" the data or by plotting the data or by closely scrutinizing the scale from which the data was read.  By considering how the uncertainties in several measured quantities will propagate, you can often conclude that the uncertainties in all but one or two of the quantities can be neglected in calculating the propagated error.  Such a conclusion should, of course, be defended in your report.

 

In this course you are responsible for evaluating all systematic and random errors in all measured quantities as rigorously and quantitatively as practicable and to propagate them mathematically when necessary.  Then, most importantly, you must compare the disagreements in your results with these uncertainties.

 

For a very readable treatment of error propagation, you may be interested in reading P. R. Bevington's Data Reduction and Error Analysis for the Physical Sciences, McGraw Hill, New York, 1969.  This book is in your CACC library at this very moment!

 

V  Error Comparison

 

      Suppose you have just completed an experiment in which you measured several quantities, calculated their uncertainties, used these quantities to compute an experimental result (called xe) and propagated these uncertainties to an uncertainty in the final result (called sxe).  Further, suppose that some theory predicts that you should have obtained a value of xt+sxt as your final result.  There exists a well defined method of determining whether these two x's agree within the stated experimental and theoretical error bars.  To do this you must first compute the disagreement between the theoretical and experimental result (called d below)

 

d = |xe - xt|       (don't forget xe and xt are mean values).

 

Next compute the propagated uncertainty in this disagreement

 

sd = [sxe2 + sxt2]l/2

If the disagreement (d) is larger than this propagated uncertainty (sd) then xe and xt do not agree within the limits of experimental error.  But if d < sd then the two x's do agree and the experiment can be termed a success.  This procedure is the best quantitative method for judging the success or failure of an experiment.  You will sometimes be required to determine whether the difference between your final result and a theoretical predication is less than the propagated uncertainty in this difference.

            Do not interpret "failure" here to mean that everything you've done is worthless.  Many factors have gone into this sequence of calculations.  For example, if you estimated some uncertainties in the initial measurements, then perhaps you overestimated  the accuracy  you could obtain.  Also there might be systematic errors present in the experiment.  It was stated at the outset that this quantitative method of calculating and propagating errors applied only to random errors.  Therefore if systematic errors are present, this analysis has neglected them.

 

 

            If sd is slightly larger than d, the next best thing is to see if d < (sxe + sxt) since sxe + sxt is the maximum difference you might expect to obtain.  If the disagreement is smaller than this upper bound then the experiment can still be called marginally successful.  However, if the disagreement is larger than

3(sxe + sxt) then you probably made some procedural mistake which has invalidated your results.  You must make an attempt to uncover such mistakes or postulate some plausible experimental errors which could cause this failure.  Without this sort of critical thought, any preceding pages of calculations are worthless and your report grade will reflect this.

 

Term 

Definition

Symbol or Eqn.

 

the quantity

the universal unknown to be measured

x

mean

the average of a set of measurements.

µ

uncertainty

the intrinsic difference between a measured quantity and the "true" or  "perfect" value of that quantity.

sx

fractional uncertainty

the uncertainty of a quantity divided  by that quantity.

sx/x

percent uncertainty

the fractional uncertainty multiplied by 100.

100(sx/x)

discrepancy

the absolute value of the difference between a theoretical value and an experimental value.

d = |xt - xe|

percent error

the difference between a theoretical  value and an experimental value, expressed as a percent of the  theoretical value.

percent difference

the difference between two experimental expressed as a percent of their average.

 

***********************************************