PRECISION AND ERROR

Given a number of measured values x1, x2, ... of a quantity X we can construct the frequency distribution by subdividing the range of X in intervals (bins) and counting the number of xk that fall within each bin,

Ni = number of xk in [xi, xi+1)

The frequency distribution if fi=Ni / N and it satisfies

i fi = 1

When the bins are very small we have a continuous distribution f(x), which satisfies ∫ f(x) dx=1.

Given a distribution we can compute several statistical estimators of the "true" value. The mean is Xmean=∑ xi fi. The mode is the value of x for which f(x) is maximum. It makes sense for unimodal distributions, ie, distribution with a single maximum. The median is the value of x such that X has equal probability to fall before or after it. We usually take the mean as the best estimate of the true value,

XN = N-1 ∑ xk

It is also important to have an estimate of the accuracy of the apparatus. The mean absolute deviation is Sabs=∫ |x-Xmean| f(x) dx. The mean square deviation is Ssqr=∫ (x-Xmean)2f(x) dx The standard deviation is its square root, S = Ssqr1/2. The estimate of the presision of the measurement is not quite the standard deviation, because it must be compensated for number of independent values, thus there is N-1 instead of N,

S'N2 = (N-1)-1 ∑ (xk - XN)2

The accuracy of the estimate of the "true" value is the standard error s(XN) = S/√(N), or the adjusted standard error s'(XN) = SN/√(N).

The accuracy of the standard deviation gives the numer of significant figures. The relative accuracy is usually s(SN)/SN = (N-2)-1/2.


Variable function of X

Given Z=F(X), the statistical estimators of Z can be computed from those of X. In general

ZN = F(XN)
SN(Z) = F'(XN) SN(X)
S'N(Z) = F'(XN) S'N(X)

For example for a linear varaiable Z = a + b X + c Y, we have

ZN,M = a + b XN + c YM
S(Z)2 = a2 S(X)2 + b2 S(Y)2


Least squares

The Least Squares is the procedure of minimization of the square errors, E = ∑ (xk - X)2. It leads to the mean, XN, as value for X. It can be used to find the best straight line that interpolates a set of combined (X,Y) measures. Supposing that the variables are linearly related, Y = a X + b, the coefficients turn out

a = ( [XY] -[X][Y] ) / ( [X2] - [Y2] )
b = ( [X2][Y] - [X][XY] ) / ( [X2] - [Y2] )

If the errors in xk are negligeable, and the errors in a and b arise only from errors in yk we have

SN(a) = S(Y) (N ([X2 - [X]2))-1/2
SN(b) = S(Y) [X2]1/2 (N ([X2 - [X]2 ))-1/2

where S(Y)=[Y2] -[Y]2 - ([XY]-[X][Y])2/([X2 - [X]2) The adjusted error have N-2 instead of N.

When we combine the estimates of two different experiments, X'+S' and X"+S", the means are weighted with the inverse of the square errors,

X = (X' S'-2 + X" S"-2) / (S'-2 + S"-2)


Systematic and random errors Systematic errors are due to effects that have not been taken into account, like the effect of the temperature on the length of a metal rod. They are often constant or vary slowly over time. Random errors arise from ambiguities or uncertainties in the measurement process, or irregular fluctuations. The random errors combine among themselves, therefore it is important to make them of the same amplitude.

The distribution of random errors is binomial. The porbability of having an error E=m e out of n errors of equal amplitude e is

Q(m,n) = n! /( 2n ([n+m]/2)! ([n-m]/2)! )

As m and n go to infinity (while E stays finite) this probability tends to exp(-E2/[2 S2] )/√(2 π S2). Therefore the probability of obtaining an error smaller than E is ψ(E/S) where ψ(y)=√(2/π) ∫0y exp(-y2/2) dy. The followin table contains some values of ψ.

   y   ψ(y)
 0.0   0.399
 0.5   0.352
 1.0   0.242
 1.5   0.129
 2.0   0.054
 2.5   0.017
 3.0   0.004


Poisson distribution

If the probability of having a certain event is p, the probability of having r events in n trials is the Poisson distribution (neglecting multiple occurrences of the event in a single trial)

P(r,n) = pr (1-p)n-r n!/(r! [n-r]!)

Letting u=np fixed while n tends to infinity, we have P(r,u)=e-u ur/r! This has mode {u}, the greatest integer less than or equal to u, and mean u. The standard deviation is √(u). When n is large the Poisson distribution becomes a gaussian with mean np and standard deviation √(np).

Likelyhood and confidence

Is a process has distribution f(x) than the likelyhood of a set of results xk is

P = f(x1) f(x2) ...

When f(x) is gaussian the maximum likelyhood is obtained for X = (1/N) ∑ xk, and S2 = (1/N) ∑ (xk - X)2. If f(x) is Poisson u=(1/N) ∑ rk and the standard deviation should be √(u): this is a check for the assumption of Poisson distribution.

The confidence is the probability that x lies in a certain interval. For the gaussian the confidence of the 1-sigma interval [X-S,X+S] is 0.683. That of the 3-sigma interval is 0.997 (0.99 confidence is achieved at 2.58 sigma). The central limit theorem assures that the mean XN of N measures follows a gaussian distribution with standard deviation the standard error SN.

We can approach the combination of two different measurement sets (assumed both gaussian with same mean X but different standard deviations) using the principle of maximum likelyhood. It results a cubic equation for X

(X-X'N) (X-X"M) { N(X-X"M) + M(X-X'N) } + N S"2 (X-X'N) + M S'2 (X-X"M) = 0

where S'2=SN2 + (X'N-X)2 and a similar formula for S". This equation could be solved in general, but it is already useful in special limiting cases where simplifying assumptions can be made.


Backing and chi-square

x is worse than x' if f(x) < f(x'). The backing of x is the probability of getting a value worse than X. Therefore B(x) = ∫ f(x) dx where the integral is extended to the values that are worse than x. If the backing of x is 1 than x is the mode. The backing of x tends to 0 as x tends to infinity. For a gaussian distribution B(x)=ψ([x-X]/S).

Given two measures x' and x" (each with its own gaussian distribution) we can define when a pair (x',x") is worse than another, and define the backing B(x',x") which turns out 1/(2 π) ∫ exp( - [x'2+x"2]/2 ) where the integral is over the region outside the circle of radius X. This generalizes to 2n measurements,

P2n(>X2) = 1 / (2n-1 G(n-1)) ∫ x2n-1 exp(-x2/2) dx

where G is the Gamma function and the integral ranges from X to infinity.

Now suppose to test a distribution and fix x0 ... xm with probability pk of having x in the interval [xk, xk+1). If the outcome of a measure gives nk in this intervals, and the total number of N is large enough that each Npk>10, so that each nk can be thought to belong to a gaussian distribution (Npk, √(Npk)), letting the chi-square

X2 = ∑ (nk - Npk)2/ Npk

the backing is Pm-1(>X2). Here there is m-1 because there is one constraint among the x. Any other constraint that might come from particular situation of the measurements reduces the index of P by 1.

The chi-square test is a check to decide whether a set of measurements supports a certain distribution. It gives the confidence level for it. The following table has some values of Pn(>X2).

n      1.0   3.0   5.0  10.0 
X2
 1.0  0.32  0.80  0.96  1.00  
 5.0  0.02  0.17  0.41  0.89
10.0  0.00  0.02  0.07  0.44
15.0        0.00  0.01  0.13
20.0              0.00  0.03



Marco Corvi - Page hosted by geocities.com.