P. Vanicek, Introduction To Adjustment Calculus
P. Vanicek, Introduction To Adjustment Calculus
P. Vanicek, Introduction To Adjustment Calculus
ADJUSTMENT
CALCULUS
P. VANICEK
September 1973
TECHNICAL REPORT
LECTURE NOTES
NO.
35217
INTRODUCTION TO
ADJUSTMENT CALCULUS
(Third Corrected Edition)
Petr Vanfcek
February, 1980
Latest Reprinting October 1995
PREFACE
In order to make our extensive series of lecture notes more readily available, we have
scanned the old master copies and produced electronic versions in Portable Document
Format. The quality of the images varies depending on the quality of the originals. The
images have not been converted to searchable text.
FOREWORD
existing courses tend to.slide over the fundamentals and treat the
as a discipline with its own rights, with a firm basis and internal
mistakes. It is hoped that the student will hence display some patience
and understanding.
These notes have evolved from the first rushed edition - termed
comments and remarks to the author. To all these, the author is heavily
ematics Dept. UNB, made very extensive reviews that helped in clarifying
any errors and misconception still present. Any comment or critism com-
P. Van!~ek
October 7, 1974
CONTENTS
Introduction • . . .. . 1
objects around us. The parameters can be usually evaluated only on the
This can be measured directly. Here the only parameter we are trying to
case, and we have to deduce its value from the results of observations of
connecting the observed quantities and the wanted parameters. The more
complicated- the problems get, of course, the more complex is the system
the surface of the earth would be an example of one such still more
sophisticated task.
able to deal with such problems systematically the adjustment has to use a
language suitable for this purpose, the obvious choice being mathematics.
l
2
for one, two or m quantities) the most general form of the mathematical
F (X, L) = 0 •
mathematical model has to be done taking into account all the physical
trivial:
X= L
R, = R, 0 (1 + at)
interest.
This can be the case, say, with our first example, if the length is measured
X = R. 1
X = R,
2
X = R.n
where t 1 , t 2 , .•• , tn are all encompassed by the symbol L. Or, in the
t1 = t0 (1 + at 1 )
R- 2 = R- 0 (1 + at 2 )
4
••• I t
n
) = L.
As we can easily see, these overdetermined models may or may
not have a unique solution. They usually have not. Therefore, in order
observations were not quite correct, that there were errors in their
determinations.
This leads us into the area of the theory of errors with its
solution.
The first four sections describe the relevant parts of the individual
fields that are necessary to understand what adjustment is all about. They,
student will supplement his reading from other sources, such as those
is used in adjustment.
5
Finally, the last section deals with the basics of the adjust-
ment proper. Here again, only the introductory parts of the adjustment
calculus could be treated with the understanding that only the subsequent
While random samples are the quantities we deal with in every-day practice,
the mathematical tools used are predominently from the continuous domain.
for anyone who wants to be able to use the adjustment calculus properly.
guished one from another. The set is defined when all its elements are
defined.
A2 - { 1, 8,
/I I'-
15, <r
"
0,
'I I,
.::0~
/p-
4}
A3 - { 0, 1}
in New Brunswick}
The text within the brackets{ ••• } is known as the list of the-
set. If an element a exists in the list of a set A, we say that the element
a e: A
a t A
6
7
written as
HC G
H¢ G
. {2 35 118} C::: I
A set which does not contain any element is known as the empty
we find that A5 =~ .
The sets are called egual if they contain the same and none but
. {1 ' 2 ' 3} ' {3 ' 1 ' 2} ' {2 ' 3 ' 1} ' •••
arranged such that each of its elements has a specific position) ensemble
of objects (elements) that may not all be distinguishable one from another.
The definition set D of a progression ~ is the set composed from all the
belongs to ~.
Example 1. 6: ~ = tl , 2 , W', 2 , 1 , 8 , C/ ) , is a
progression, and its definition set D is given by
D :: {1 , 2 , u, 8 , CV' }•
At this point, the difference between a progression and a set
ion than the one given in Example 1.6. However, the sets
{U, 8 ' 2 ' 1} ' . {2 ' 1 ' 8 ' cfC/' ' " }' ...
are all the same as the definition set D in Example 1.6.
product set and denoted by Ax B (reads A cross B), whose elements are all
n sets. In such a case, the elements of the product set will be all the
define the Cartesian n-power Anlo'r An.if no danger of confusi.on with indexed
set exists·I of. a set A as ·t:ne Cartes-ian p;rC!>duat ot tl:i.e same ·s:e-t A with itself
n-times.
Example 1.7: If A:: {3 , 1 , 5} and B = {2 , 4} , then the product set
Ax B is AX:B- {(3, 2), (3, 4), (1, 2), (1, 4), (5, 2),
( 5 , 4)} . Referring to Example 1.1, we can easily see that:
' ',
' :0: )
I I '
g Al X A2 ' ( 4 .18 ' <r
9
(1 , <:::::> ) t A1 x A2 but (- 1 , ~ ) e: A2 x A1 ,
set which is a subset of both A and B and does not contain any elements
other than the common elements to A and B. The intersection of two sets
can be represented by the shaded area in Figure 1.1· Diagrams of this kind
are called "Venn diagrams" .
Figure 1.1
Figure·l.2
If A() B = r/J, then the sets A and B are said to be disjoint sets.
n
() A. , where
i=l ~
n
('\ A. ::: A1 () A2 tl A3 ... () A •
i=l ~ n
·Figure 1.3
section, the union of the two sets is represented by the shaded area in
11
Figure 1.4
Figure 1.4
0' 4 } '
and I U R ::: R
of two sets is known as the complement of one into the other. Referring
to Figure 1.5, and considering the two sets A C:: B, the set of all the
of A in B, and is dimoted by B - A. --
12
13-A
Figure 1.5
The complement of A3 in A2 is
-\I,~
A 2 - A3 : { 8, 15 , (t , :0: , 4} , and
I I \
element from B to each element of A. This means that for each element
Figure·l.6
Note here that the one-to-one relationship (i.e each b £ B has got one
and only one argument. a-~ A) is not required. We shall denote any such mapping
by
and read it as "f is an element of the set of all the mappings of A into B",
13
f is called a.n onto ~a.pping, or simply we say that "f maps A onto B".
Example 1.11: Given the set A = {a1 , a 2 , a. 3 } = {2, -1, 3} and the mapping
= 27.
Generally, f is an into function, hence we write
1. 7. Exercise 1
(v) is Ac:H .
1 (vi) is (AU D) C: H ?
(iii) AUB .
) (iv) U- A j •
following sets:
( i) DUH j
(ii) H.n c )
(iii) en B . I
(iv) A- C )
(v) BUc .
J
(viii) A - (c U B) ·
6. Given the set X:: {-2, -1, 0, 1, 2} , with f E {X~ Y}. If for each
onto function.
2. FUNDAMENTALS OF THE MATHEMATICAL THEORY OF PROBABILITY
real numbers "b" satisfying the inequalities (0 .:::_ b < 1)) that has the
(1) If D'C D, then P(D') = 1- P(D- D'), (note that D- D') is the
of the probability function P (takes any value from [0, 1]) is called
the probability. Note that the difference between the function and
following consequences:
{_1} P(_D) = 1,
(2) P(~) = 0,
15
16
the conditional ;probability. The right hand side, that is P(A/13),is read
notice that:
(2) If B CA; then A ()B =B (see section 1.4), and then P(A/B) = 1, ·
(3) If A(\ B = C/J; i.e. A and Bare disjoint sets ; then P(A/B) = 0.
such a case we say that A and 13 are independent. Using the definition of
probabilities, i.e.
p (D.' Dj)
l.
= p (D.)
l.
P(Dj) i :! j
p (D.
l.
I
Dj' Dk) = p .(D.)
l.
P(Dj) P(Dk) i :! j, j :! k, i :! k,
n
IT p (D.)
l.
i=l
Example 2.1: Suppose we have decomposed the probability space D into seven
Figure 2.1.
+------....-··--···------·-----......-
O,'?J"'c.
0. '2,-11-----...----·---r------ --------------1--
0.\-t------ ·----
of their independence.
2.4. Exercise 2.
bility of appearing when the die is tossed, proportional to the number itself.
19
ordered progression of elements (see section 1.2) s = (sl' s2' ... , sn)
such that:
~1 s2 s3 s4 s5 s6 s7
~ !! { 1, <:>
' cfCJ': 1, 1, <I. ' r:::; }
dl d2 d3 d4
D= {1 , ~ , &Cr: CL } , which
20
21
are:
From now on we shall deal with DCR (recall that R is the set of
all real numbers}, i.e. with numerical sets and progressions only. Also,
is not standard in the sense that it admits much larger family of objects to
be called random samples than the standard definition. More will be said
about it in 3.2.4.
22
appeared:
number d.J.. 1 2 3 4 5 6
count c.J. 14 17 20 18 15 16
•
Find the probability that:
(i) a 3 appears ;
(ii) a 5 appears;
(iii) an even number appears j
Solution:
20
(i) P(3) = - - = 0.20 )
100
(ii) P(5) ··--
15
100 = 0.15 J
(iii} P(2 ,4 ,6) = P(2) + P(4) + P(6)
17 18 16 51
= 10'0 + 10'0 + IOo = IOo = 0. 51 ~
(iv) P(2,3,5)'= P(2) + P(3) + P(5)
17 20 15 52
= I05' + I05' + laO= laO= 0.52 •
of course, its definition set Dis a numerical set), which we shall from
now on always assume, then Pis a discrete function mapping D into [0,1].
23
relative frequencies.
~ :.; ( 1 , 2 , 4 , 1 , 1 , 2 , 1 , 1 , 2 ) , n = g_.
c1 = 5 , c2 = 3 and c3 = 1.
2
Figure 3.1
24
is denoted by P(D' 2 or P(~ .::_ d .::_ dj). To answer this question, we use
(3.1)
The above expression (equation 3.1) must be understood as giving the actual
probability of d€D 1 s {dk' ••• , dj~D rather than d€[~, dj] (i.e. the
~' dk+l' ••• , dj-l' dj rather than the probability that d will be
anywhere in the continuous interval I~, dj]}. This is not always properly
understood in practice.
C(d.l
~
= E P(dj) €[0,1] (3.2)
j<i
Example 3.4:
-
abbreviated by CDF.
sample 1;.
Figure 3.2
CDF:
experiment as follows:
to Figure 3.2.
n
1
M=- E ~- E [d1 , d ], (3.3)
n i=l ~ m
[ M =
m
i=l
E d.~ P(di)
.1 (3.4)
27
m 1m
R.H.S. = ~ d.
~
(c./n)
~
=- ~ d.c. =
i=l n i=l ~ ~
1 n
= l: t;, =M •
n i=l ~
written as:
(i) E(k) = k,
(iv) E(E ~j) = E E(~j), where ~j, j = 1, 2, ••. , s, ares random samples
j j
with the same number of elements m in their corresponding definition
sets Dj (Do not confuse~. with ~j~ the former is an element in the
J
latter. In other words, ~. is a single element in a sample, but
J
~j is one sample in a class of samples) ,
(v) If ~ = (~ 1 ), then E~) = ~l'
(vi) E (E(~)) = E (~) •
28
(3.3) as follows:
1 n 1 15 2
M=E( ~ )= - r ~. = -(1+2+4+1+1+2+1+1+2 }= - = 1-
n i=l 1 9 · 9 3
Also, we can use equation (3-4), from which we
get:
M=E(~)=
m
.: djP(dj)= 1.
J-l 1
+ 2 • ~ + 4 •
15 2
t ~
=- (5+6+4'=- = 1-
9 '- 9 3 .
Obviously, both formulae (3.3) and (3.4) give
identical answers.
about any point , e.g. the origin 0 (see Figure 3. 3 which uses the data
I
I f (cw-s) .l.. (or 3) }(or1)
~ ·9~
•
I
~0
(
1 4
'
2 4
I I
~ M ~
I
Figure 3.3
29
with mechanics, the mean M computed from equation {3.4) is also called the
is called the variance (dispersion) the actual sample. The square root
n m
1 E 2
E d. P(d.)
n
i=l j=l J J
m
= E P (d.) (d. -M) 2 (3.7)
j=l J J
30
of the operator Eon (s-E(s)) 2 meaning really P(d.) (d.-M) 2 and is often
J J
written as
we obtain:
From equation (3.5) we have E(S) = M, then by substituting for E(s) we get
(3.8b)
m
2 2
E d. P (d.) -M (3.9)
j=l J J
(as discussed in the previous section) we can regard the variance of the
follows:
n
E(E; 2 )= 1. ~ E;~= 1.
9 (1+4+16+1+1+4+1+1+4) =J1_ •
n i=l ~ 9
33 9 .3;3..(15) 2-
8 2-- var (c)-
"' - 9 - (15)2.-
9 - 81
- 297-225 - 72 - ~ ~ 0 89
- 81 - 81 - 9 - - · - .
2 12 = 2.828 ;. 0.943.
3 _3
= 1. 33
9 (5+12+16) =- 9 '
82 33
= 9- (159)2 = _98 ;. 0.89.
It should be noted here that the same value for the sample
variance can be obtained from equations (3.6) and (3,7). The verifica-
32
tion is left to the student (e.g. using the data from the above example).
differently for n odd and for n even. For n odd, Med (~) equals the ~
Med(F;) = 1 •
as the difference between the largest (t;9;l and the smallest (F;s) elements
of t; that is:
(3.12a)
Note that the range of the sample can be also determined from its
number).
to make the intervals equidistant. The difference between the upper and
scale). Relative
Frequency (c; )
3
2 ··- - - - -- -..
Figure 3.4
Note in the above figure that a rectangle is drawn over each interval
with constant height equal to the corresponding class-count.
It is usually required that the area of,ol: under,the histogram
given by:
k
= ~(c 1+c 2 + •.. +ck) = ~ t c ..
1
= ~n.
i=l
This means that the area under the histogram equals the class-width
simply have to divide each ordinate c. by the quantity n~. The new
1
-c = s·· = 0.08,
2 100
37
right-hand scale).
ness of computing - c .•
J.
gram by .Q. and s, respectively (e.g. in Figure 3.4, g.= 17 and s=l).
Notice that for any subinterval D' = [a,b] of the interval 6. = [s, R-],
we can compute the area a (D') under the histogram. This a(D') will be
histogram. On the other. b.anao., the ordinates of the histogram do not represent
probabilities (again compare the histogram with the bar diagram given in
section 3.1.2).
= 0.27 .
- = c2- = 102.5
cl
9
=. 0.0878,
- 5 . 0.0488, and
c3 = 102.5
c4 =
2
102.5 =. 0.0195.
Relative
frequency
-'
C·
9 o. 0~78
2
0+-----~----~~~~~~~--~~---+--~~
0-7 ./f..8 1 /3.0
I
6
Figure 3.5
= 2.9.0.0878 + 1.1.0.0488 ;
. 0.2546 + 0.0537
.
= 0.3083 = 0.31,
.
which gives a better agreement with the actual
example 3.11.
41
In order to make the total area under the polygon egual to one
we have to add one more class interval on each side (tail) of the
corresponding histogram. The midpoints s' and i' of these two, lower and
Therefore, it can be easily seen that the area~· under the poly-
gon has again the properties of probability. This means that ~· is one
of the possible PDF's of the sample. Hence~~ can be used for determining
0.12 ........- - - -
3 IS
Figure 3.6
42
a = '4' (1.2
' • 0.12 + ~
1 ( 0.12 + 0.08 ) ~
= 2 (0.50) = 1.00.
Let us compute the probability P(6~x~l0)
construct the histogram and the polygon corresponding to the CDF of the
sample which will be respectively called the cumulative histogram and the
(3.2), namely
this section.
zero}.
cumulative histogram.
44
C(;_)
o. s6--···--- ,--
I
1
0.8'3
I
c,oo~------~~------~._--~r-~---4-------+----------~--
51 9 : 13 1"1
6 10
Figure 3.7
C(-1) = O,
C(3) = ~(4.0.12)= ~(0.48)= 0.24,
= p[:..l,3]
3.4).
45
CC;)
J.oo
o. 9b
o.f16
0.64
0,54--
0.24
-I 3 -, I
I
fI 15 \9
6 10
Figure 3.8
By examining Figures 3.7 and 3.8, we can see that the cumulative
polygon uses tbe central point of each class-interval along with its
Because of the nature of the CDF, we can see that the cumulative
using:
respectively. Therefore:
•
P[6,10]= P(6~x~l0)= 0.805-0.54~0.27 ,
To close this section, we should point out that both the histo-
val probabilities (see the histograms and polygons of section 3.1.6) more
easily and readily, the science of statistics has developed a more con-
xe: {U + R}
be used.
When a large enough number of values x(u) have been recorded, we can
plot a histogram showing the relative count of the x{u) values within
any interval [.x0 , ~]. In this abstraction we can imagine that we have
Denoting the limit of the relative count d~vided by the width dx o£ the
Going now back to the realm of mathematics, we. see that the
literature to refer just to the values x(u) as random variable with the
We note that the function $ is thus defined over the whole set
of real numbers R and has either positive or zero values, i.e. cjl is non-
piece-wise continuous on R.
j t/J(x) dx =1 ( 3 .15)
-oo
since the area under the "smooth histogram" must again equal to 1 (see
non-negativeness being the first two. We note that eq. 3.15 is also
the integral (3.15) is illustrated by the shaded area under the rp.
Figure 3.9
l
The definite integral of the PDF, tP, over an interval D c D is
I tfl(x) dx = P(x -
> X )
0
€ [0 1 1] 1 (3 .1Gb)
€ [0, 1] • (3.16c)
Consequently,
P (x > x )
- 0
=1 - P (x _< Xo> • (3.17)
The integrals (3.16a), (3.16b) and (3.16c) are represented by
(a) ~)
~--T-----~~~~~~~~x
..... 00
(c:)
._.. ,._,__.....,___,'-'-""""~"""'1~---1---:11• )(
)f' )(2.. -to oO
Figure 3.10
ity spaces, should be again-born • in.··mind. In the discrete space, the value
of the PDF at any point, which is an element of the discrete definition set
in the compact space, it is only the area under the PDF, that has
got the properties of probability.* We have already met this problem when
* The whole development for the discrete and the compact spaces could be
made identical using either Dirac's functions or a more general definition
of the integral.
so
X
0
P (x = x0 ) = f ~(x) dx = 0*).
X
0
I----------------------------------
X
~(x) =f~(y) dy, e:[x+ [0, 1]} (3.18)
-<»
\t'(x) \f(x):: I.
l.o - - - - - - -
L---~~------------~~--~~x
--+00
a..
Figure 3.11
* This may not be the case for a more general definition of the integral,
or for ~being the Dirac's function.
51
Figure 3.12
d 'l:'(x)
<P (x) =
dx
In addition, we can see that q,(x) has to disappear in the infinities in order
! <P (x) dx = 1.
-co
Hence, we have:
if we do not know (or assume) its PDF. On the other hand, we do not have the
one-to-one relation between the random variable and its PDF as we had with
52
'
the random samples (section 3.1.1 and 3.1.2). The random variable acts
function called PDF, that runs from minus infinity to plus infinity.
Therefore, strictly speaking, we cannot talk about the "mean" and the
about the "mean" and the "var-iance" of a random sample. On the other
hand, we can talk about the value of the argument of the centre of gravity
of the area under the PDF. Similarly, we can define the variance related
to talk about the mean and the variance of the random variable; and
we write
00
E* (x) =~ =f x ~ (x) dx
t) (3.21)
We can see that the argument in the operator E* is x·~(x) rather than x,
again use the customary notation to conform with the existing literature.
. . . r'
Note the analogy of (3.22) with equation (3.8), section 3.1.4. The
random variable.
00 00 00
ttin order to prove this equation, one has to again use the Dirac's
function as the PDF of E*(x).
54
term equals~ (equation (3.20)), and the integral in the last term equals
00
(3.23)
-oo
(3.23a)
mind that E* is not operating on the argument, but on the product of the
The expression
(3.24)
-oo
is usually called the r-th moment of the PDF (random variable); more
precisely; the r-th moment of the PDF about zero. On the other hand,
00
r
m'r = ! (x-~) ~ (x) dx • (3.25)
-oo
55
l.l = ml (3.26a)
Compare the above result (3.26 a, b) with the analogy to mechanics men-
got a parent random variable". This parent random variable x~R is usually
istics to postulate the PDF of the population for any random sample, and
assume that the sample can be regarded as having been picked out, or drawn
from the . population, each element of the sample independently from the rest.
present Introduction does not deal with statistical testing we shall keep
shape of its PDF. The individual members of a family vary according to the
value of these parameters. It is common to use if possible, the mean and the
standard deviation as the PDF's parameters. The less parameters the family of
of its parameters that woUld fit the actual random sample the best. In
other words, the shape of the postulated ~(x) is chosen first; then, its
Since we shall be dealing with the samples and the random var-
iables (populations) at the same time, we shall use, throughout these notes,
the latin letters for the sample characteristics, and the corresponding greek
Figure 3.13
=<
h, for (k - q < x < k + q)
o(x)
0, for (x < k- q) and (x > k + q).
57
~(x) <
=
h, for (i.x-kl.-< q) ·
! ~(x) d.x =1
-""
that must be satisfied for any ~ to be a PDF. Let us
00
f ~ (x) dx
k-4-q
k+q
= 0 + ! hdx + 0
k-q
k+q
k+q
= h ! dx = h [x]k = 2hq = l.
-q
k-q
0, for (x ~ k - q)
ll'(x) = !
X
-oo
~(x) d.x -1
2q !
k-q
X
dx. = ;q_ (x-k+q), for
( lx - kl ~ q) •
1, for (x :::, k + q) ,
o/(x)
/. 0 ---- - - - - - - - - - - - -
0,5
Figure 3.14
interval over which <l> :/: 0, and is constant everywhere else. Note
that:
cjl(x) =d '!' (x)
dX
The mean of the given PDF is computed from equation (3.20) as
follows:
00
k+q 1 2 k+q
11 =f x<P (x)dx = 12q f
k-q
xdx = -2q [:!....]
2 k-q
= -14q (k 2 + 2kq + q
2
- k
2 2
+ 2kq - q )
- 4kq -
- 4q - k.
This result satisfies our expectation for a symmetrical function
3 k+q
= ~q [~ Jk-q - 2k 2 + k2
follows:
<
~her, for ( lx-p.l) ~ her)
R(·~, a ; x) = ~(x)
0, for ( Ix - pI ) > her) •
1, for (x ~ ~ + ho) • -
Assume·· that we would like to compute the probability of
follows:
6o
{>(x).
I ,
~~~~~~~~~~~~~-.-x
(~-f3C1') I (~+JJo-)
(~-rr) cJ+o-)
Figure 3.15
].I+a
P ( 1,1-0! < X < 1J +a ) = f cp ( X ) dx
}.1-a
].I+ a
1 1
= f 2ha dx = 2ha [ 2 o]
}.1-a
3.15.
3 2
m3 =ll_·-~~
Figure 3.16
62
~(x) =
+ q), for (k < x < k + q).
~(x) = < h (q - Ix -
qo
k
for ( Ix - k I
I), for ( Ix - k
:: q) •
I < q)
From the above, we can see that the triangular PDF has the
-
--. ,
1
Thi.s gives . us : h = -,
q
and hence,
~(x) ·= < - 1
q
0, for
.
ix-~ l
q
-. (1·x-k 1' <
for
Clx-kl ~ q).
q) •
student.
<
lx-}.1 I ' for( lx··\11
6(/
< 16a)
0, for (x ~ ~- q).
'f(>c)=J.
t.o
Figure 3.17
f x (~ _ Ix-u I ) dx
- 11-fllq q + xq2-
k-q q 9.2
X
......._______-"'-]J dx + f 9. - x + 11 dx,
]1-q q2 11 q2
and we get:
X
f (q + x - 1J) dx
11-q
== 1 2 {x 2 -2~x+i+2q(X-!1) + q2 }
2q
2
= . (x-11) _ + (x-11)_ + ~
2q2 q 2
Similarly,
X ( 2
l f (q-x+).l) dx = - x;JJ) - + (x-~2_,
2
q 11 2q q
and
f g,_+x-11 dx=l
2 2
q
JJ-q
1
=1/
0, for (X 2_1-l- ./6a)
(x-1J)
120 2
2
+ (x-1J) +
/60 2
.
!, for (l..l - ./60 .:s_ x .:s_ u) •
(X-l..l) 2 + (X-1J) 1 f ( .j )
2 ; 60 + 2' or u .:s_ x .:s_ 1..l + 60 •
\ 120
can compute the probabilities: P(l..l-0 .:s_ x .:s_ 1..l + 0),P(~-20 ~ x ~ 1..l + 20)
and P (1J-30 .:s_ x .:s_ 1J + 30) as well as the third moment m3 about zero for
the triangular ·PDF. Again, we give here the results, and the verification
P(lJ-0 2_ X 2_ 1J + a) ; 0.66 ,
m
3
= l..l3 + 3a 2 1..l •
66 .
as:
*The superscripts and subscripts here are found very useful to distinguish
between the c~ponents xJ, j = 1, 2, ... , s of the multivariate X, and
the elements x~, i = 1, 2, •.. , nj of the univariate (random variable)
XJ. J.
67
s
We can speak of a probability of X£ [X0 , Xl] C R , and define it
as follows:
r-------------------------------,
xi
P(X0 ~X~ x1 ) =!X ~(X) dX E[O, 1] . ( 3-28)
0
Here the integral sign stands for the a-dimensional integration~ dX for
j = l , 2 , ••• , s
Note that in order to be able to call the function ~ a PDF, the following
f
R
s~ (X) dX =1
as follows:
X
~(X) = f ~(Y) dYe {Rs + [0, 1]} (3-30)
-oo
Example 3.19: Consider the univariate PDF shown in Figure 3.12. This
~(x) = N(~, o; x) •
Figure 3-18.
~
\ )< -- •· .v
X
\~.:..- ).l~ ~.
[~-
'2
~--:-
pi
2
0 2. J
69
namely
s xj
= IT ! l (3-31)
j=l xj
0
s s
IT IT P(xj < xj < xJ1' ) •
o-
j=l j=l
of the PDF's of· its constituents, then these constituents are known as
we have
The sequence
... , u )
s
= E* (X) (3-33)
where
(3-35)
where
(f~ =
J
~ - 2 ~ - 2 ~ 2 -2
E*(X-~) = E*(X-E*(X)) = E*(X) - ~ ( 3-37)
and
section 3. 3. 4) •
Let us now turn our attention to what the mean and the variance
we obtain
s
=J [xj ~ . (xj) ( II ~ R, (x.R,) dx.R,) dxj] ( 3-39)
Rs J .R-=1
R-r!j
( 3-40)
72
Similarly,
(3-41)
Thus for the statistically independent X, we can compute the mean and
let us define another statistical quantity needed for this matrix. This
xj and xk of a multivariate X as
cov (3.42)
let us write
( 3. 43)
we can write:
74
. k .
O'jk = E*(xJx -xJ j.lk-j.ljXk + ].lj].lk)
'k k
= E*(xJx ) - ].1 E*(xj) - ].I.E*(x) + ].lj].lk
k J
. k . k . k
= E*(xJx ) - ].lj].lk = E*(xJx) - E*(xJ) E*(x) =0
. k
Hence, for statistically independent components xJ and x , we get
(3-44)
r r
t
E* ( II X ) = II (3-45)
t=l t=l
0'2
1 0'12 0'13 crls
cr2
0'21 2 0'23 0 2s
~*
X
= (3-46)
.
0'2
crsl crs2 s
that the superscript T in the above formula stands for the transposition
student.
the diagonal elements are the variances of the components and the off-
always positive) and the inverse exists if and only if there is no absolute
as follows:
J::lI.
n -
Jr•i·
I. 2
1
t;3,
2
• • •
..... ,
• • I
E;l )
nl
~2 )
e R
e R
1
n2
(3-48)
n (_t;l, F;3'
n2
I.
lt,:S l«~. s
t;2, F;3,
s
• • • • • I
E;s
n
s
n
e: R s
(experimental) PDF and CDF of a multisample in very much the same was as
we have done for a random sample. Also, the distribution and cumulative
student.
... , (3-49)
r
l -2
s =
2 2 2 2
(sl, s2, s3, .•• , ss) = E- (n-M)
- 2 e: R
s ,
] ( 3-51)
(3-53)
Example 3.20: Let us determine the mean M, the variance s2 and the
standard deviations of a multisample n = (t; 1 , t; 2 , t; 3 ),
where
sl = ( 2, 3, 4, 7, 4) ,
s2 = (6, 4, o, 3, 2) and
s3 = ( 5; 2, 5, 5, 8) .
Here we have n 1 = n 2 = n 3 = 5. The mean M is given from
equation (3-49) as
78
as follows:
n 5
M = 1:._ 1:1 ~:- = 1:. 1: ~:-
l nl i=l ~ 5 i=l ~
= l5 20
(2 + 3 + 4 + 7 + 4) = 5 = 4 '
M2 = t (6 + 4 + 0 + 3 + 2) = l~ = 3,
M3 = t (5 + 2 + 5 + 5 + 8) = ~ = 5 ,
and we get
_,
M = (4, 3, 5) ,
~ 2 2 2
8 = (81, 82, 83) •
as follows:
nl 2 5
8
2 1 1
=--- 1: (~;-M 1 ) =-1 r
1
(~. - 4)
2
1 nl i=l ~ 5 i=l ~
1
= 5 [4 + 1 + 0 + 9 + 0] = 514 = 2.8 '
79
= ; [9 + 1 + 9 + 0 + 1] = 2 ~ = 4.0
= 51 [0 + 9 + 0 + 0 + 9] =
18
-s = 3.6 ,
and we get
2
Taking the square root of the individual members Sj' j = 1, 2, 3,
number of elements, say n, we can write the covariance Sjk between these
( 3-54)
Note that the covariance Sjk' as defined above, depends on the ordering of
. k
the elements in both components ~J and ~ , whereas the means Mj and ~ and
t he .
var~ances sj2 an d sk2 d o not. Therefore, to obtain a meaningful covariance
. k
sjk' each of the components ~J and ~ should be in the same order as it
was acquired. This can be visualized from the following example. Assuming
that the elements of ~j are observations of one vertical angle, and the
the value of the observed vertical angle, the matched pairs must be
respected.
80 a
5
8 12 = 8 21 = 51 i~l[(~i-
1
4)
2
(~i- 3 )
]
+ (3)(0) + (0)(-1)]
+ (3)(0) + (0)(3)]
l .
= - [ 0+3+0+0+0]
5
= -35 = 0.6 and
-
= 51 ( 0 - 3 + 0 + 0 - 3]• -6 : :
=~ -1.2 •
62
1 6 12 613 6 1. s
82
8 21 2 8 23 62 s
E = (3-54)
n
82
8 sl 8 s2 s
dimensional case either. While formulae for the mean and the variance
and (3-42) can be devised for the multisample, the ones used mostly in
results were:
E
n
= -1.4 4.0 -1.2
3.3.7 Correlation
may, of course, vary. We can see that the covariance Sjk E R may attain
defined as
( 3-57)
as follows:
- 1.4
= 1.67 . 2 = - o. 42
=
0.6 = 0.19 '
1.67·1.9
-1.2
2 • 1. 9
=- 0.31 .
Note that:
Since
sample n = (~ 1 , ~
2
, ~
3
), where:
1 4),
~ = (2' 1, 3, 5,
~2 = ( 4' 2. 6,_ 10, ._at~ ...
1;3 = (-4, -2, -6, -10, -8).
83
8 1 = ·12~ s2 = 83 = 2 12 '
8
12
=4 and 813 = -4.
Hence
8 12
P12 =8 •S = ...,l-2.......4-2._,1....2 = + l,
1 2
and
~13 _4
P13 = S1 ·S3 = - 1 '
such an arrangment using the two constituents .;1 and ~ 2 of the multi-
t;1 F;2 1
( ~ :t.-M1) •
~.
1
~
1
( F;i-M1) (E.l-M )2
"i 1
F.::~
2
(t,;i-M2) '(
(~~-M2)2 ~i-11 2)
2 -2 4 6 3 9 -6
3 -1 1 4 1 1 -1
4 0 0 0 -3 9 0
7 3 9 3 0 0 0
4 0 0 2 -1 1 0
-
E 20 14 15 20 -7
. --- -·--·-
1
il'!.1 = ~ (20) = 1.~' M = - ( 15)
2 5 =3 '
s12 = 5
1
(14) = 2.8, 2 1
82 = "5 (20) =4 '
and
p12 = 1.67
-1.4
. 2 =- 0.4?- .
86
,61, 70, 102, 107, 113, 114) 117,'119, 120, 126, 120, 129, 129, 13:2, !37,
i3V, 130~ 130, 142, 143, 146, 1~6, 1·17, 1471 148, 149, 149, 150, 150, 153.
153, 156, 157, 158, 150, 159, 159, 159, 162, 162, 164, 166, 166, 166, 167,
16!>, 169, 169, 169, 170, 170, 171, 17~; 172, 172, 173, 173, 175, 175, 176,
176, 1.76, 17.7, 177, 178, 179, 180, 180, 181, 181, 181, 182, 183, 184, 184,
l85, 186, 187, 188, 188, 190. 192, 102, 193, 194, 194, 194:, 195, 195, 195,
· ::JO, 10~I., 10°.
··1nn "'' 1n~ 900, •01
iT1..1 1 - -
~01 , -9 01 , -9 0"'"":t
, -
9 •o•
9 9- 0....-, •O·'·!:': -•o~u 1 -~os 1 9..,.. 09.
... ·-:> 1 - • ,
90!)
-
"09 21.~~ 216 <:11!)
I .- • J J~ 919 I 219I -•)•)1
I • - J
2?9 9"'3 2"'7
IW-1 ,.,_, l .- t -
"33.. J --'*'J
·~-.t. 2'-'6
0
93~, I
t .-.
Required: (i) Glassify this sample according to your own choice, and
polygons.
121 and 174 ems, by using your distribution histogram, your distribution
(3) Verify the results given in Example 3.18 for the mean, the variance
(l) The following table gives the weights as recorded to the nearest
Required: (i) Compute the mean, the standard deviation, the median
and the range of this random sample using both the original sample
(4) Verify the results given in Examples 3.17 and 3.18 for the
0, everywhere else.
( iv) Use both the PDF and CDF' to determine the following
probabilities: P(x~1.5),
P(x.?_2.5),
P( -1 .::_ X _-:_ 4) ,
(v) Compute the 3-rd and 4-th moments of the PDF about
zero.
¢(x)
=
< k·x , for (O < x < 2)
0 , everywhere else.
deviation of x.
0 , everywhere else.
\~
s2t 1213 st 6h ' or
¢(X) lx 2-rl < t 13)
, everywhere else,
where q, rare some real numbers and s, t are some positive real
numbers.
the two approaches are very different, however, in practice they coincide.
have some spread or dispersion (by spread we mean that not all the
the side of a rectangle using a graduated ruler, we will have two possi-
a) b)
Figure 4·.1
90
of it will not produce any spread. This is simply because the beginning
of the side will be at a graduation line of the ruler, and at the same
time the end of the side will be at another graduation line, and hence
we get always the same result. On the other hand, if the end of the
side is located between two division lines on the ruler, there will be
influence the spread, we can try to account for them in one way or the
time, etc.
some cases, the systematic errors remain constant in both magnitude and
example, the error in the rod reading due to the inclination of the line
91
of sight of the level, with respect to the bubble axis, can be eliminated
by taking the backsight and the foresight at equal distances from the
level.
in the observations. These blunders are usually gross errors due to the
blunders has to be carried out before starting to work with the observations.
The ways for intercepting blunders are numerous and are as different as
the experiments may be. We are not going to venture into this here.
i.e. are still not identical, and we begin to blame some unknwon or
errors.
same quantity t; where ~· is the unknown value, and can be written as:
n
~i
= nl
~
+ E: i' ~
L-
- 1I 2I •••I
n •
92
The sequence
( 4-3)
(or the sequence L, equation (4-l), for this matter) is declared a random
freely in practice.
Figure 4.2
* It may happen, and as a matter of fact often does happen, that we are
able to spot some depen~ence of e (for whatever this means) on one or
more parameters,.e.g. temperature, pressure, time, etc., that had not
been suspected and eliminated before. Then we s~ that the e's change
systematically or predictably with the parameter in question, or we say
that there is a correlation between the e 's and the parameter. Here, we
may say that the observations still contain systematic errors. In such a
case we may try to eliminate them again, after establishing the law
governing their behaviour.
93
This explanation leads to the derivation of·the well known model - the
account, along with the derivation of the law, due to de Moivre, are
Appendix I) :
where its argument E is the random error, i.e. a special type of random
variable with mean eqlal to zero, and C is the only parameter of the dis-
Figure 4.3.
From the above Figure we note the following characteristics of
( i) G is symmetrical around 0.
dition:
00
00
2 2 2
a
e:
= E* ( e: -~ )
E
= f e: G( C; e:) de:
00
2
= ig_ C'IT f E exp (-2e: 2 /C)de:: (4.6)
-oo
Recalling that
00
2 2 2 f1f
f t exp (-a t ) dt = (a> 0),
0 4a 3 '
* The same result can be obtained using slightly weaker (more general)
assumptions through the "central limit theorem".
95
=--
2a
3 '
where
a = lc2 .
Hence,
12 · clc c
= 21lc • 21l2 =4
and we get
C = 4cr~2 • (4-8)
2 .
Consequently, the variance cr , or rather the standard deviation cr ~ can
~ . ~
Note from equation (4-8) that ~ = IC/2, which equals to the abscissas
Example 4.1. Let us compute, approximately, the probability P(-cr < £ < cr)
- E- - ~
that:
Hence
96
2 2 2 4 6
exp ( -€ /(2o )) = 1 - _e- + ....£_
4 _....£__ +·
E
2og
2
Bog 6
48oe:
and
E:
3 5
= __1_ [ 20 ~:~ 1 2 • 2oe_ + _L . 2oe:
o(21T) ~ 2o 3 Bo ~ 5
e: e:
7
1 2oE' + ]
- 48o 6 • 7 •••
e:
= ~~~1.
1T.
(J {C
- 0.167 + 0.025- 0.003]
e:
Thus:
Y = e: + ~
y ' ( 4-10)
in equation (4-9), where y is the argument of the new PDF- the generalized
Gaussian. Such generalized Gaussian PDF is usually called normal PDF and
is denoted by N, where:
2
(y-~ )
'-T). 2
y
.
( 4-11)
used to have, in the power of the Gaussian law (also called the "normal
to this law and display a histogram conforming to the normal PDF, they
are normal. On the other hand, if they do not, they are regarded as
The normal PDF contains only two parameters - the mean ~~and
Note here that the family of G(~; e:) is a subset of the family
of N(~, ~; y). Also note that the following condition has to be satis-
fied by N:
The formula for the normal CDF corresponding toN is given as:
2
1 y . (x-lly).
'!'N(y) = ali";) L"'exp(- 2 ) dx' ( 4-12)
y 2o
.Y
where x is a dummy variable in the integration.
98
shown that:
and
< y < ll + 3a ) ; 0. 997 ..
P( J..l y -3ay- - y y
in example 3.18).
variable with mean J..l and standard deviation a • Note that the above
X X
tor x.
The transformation of the normal variable y (equation (4-10))
y-]J
to a standardized normal variable t = ~results
a
in a new PDF
whose mean J..lt is zero and whose standard deviation at is one. This
PDF is called the standard normal PDF, a particular member of the family
S:dtnce both the parameters lit~· '.0 and crt' • -l;~ are. determined once
for all, the standard normal PDF is particularly suitable for tabulation
standard normal PDF for different values oft. Note again that
00 l
£oo N(t)dt = /(2n)
The CDF corresponding to N (t) is given by
2
l ft exp (- L) dx,
I!'N(t) = 1(2n) -"" 2 (4-15)
or
2
l 1 ft exp (- ~ )dx,
I!'N(t) =-
2
+
1(2n) 0 ( 4-16)
Hence, care must be taken when using different tables for computations.
~ (-t )
N o
=l - ~ (t ) .
N o
100
~N(t) = 12 + 212
1
erf (t), ( 4-17)
where, erf (t) is known as the error function, and is obviously given by
PDF and CDF for computations concerning a given normal random variable
= 1. - 0.8413 = 0.1587 •
N{\:)
Hence, K2 = (0.1587)(1000) = 159 students •
(iii) P(h _> 74.6) = P(t > 74 • 6- 66
- 5
= P(t ~ 1.72)
= l. - 0.9573 = ~.
Figure 4. 4-iii
Hence, K~ = (0.0427)(1000); 43 students.
H1 - 66 = 5(0.4) = 2,
i.e.
H1 = 66 + 2 = 68 inches,
inverse solution.
103
But:
and we get
we have to write:
and we get
P(t <-t )
- 3
=1 - 0.20 = 0.80.
By interpolation in the above mentioned
Figure 4.5-iii
table we get: (-t 3 ) = 0.842, which
and, H3 = 66-5(0.842)
= 66-4.210 = 61.79 inches.
101~
N(t)
= l. - 0.95 = 0.025
2
and we get:
t
0
= 1.96, which corresponds to
t 0 = 5K = -
1.96 ,
and
105
can write:
we get:
= 0.6826 = 0.683,
which is the same result as obtained
in example 4.1.
106
section 4.2 while we developed the analytic formulae for the PDF's
mostly used in the error theory. Let us get back to it and state the
normal PDF N(~t' crt; t). Other PDF's are used rather seldom. The
estimate is often used in this context) the mean ~t of the parent PDF,
L = ( t . ) = ( t'+e . ) , i = l , 2 , ••• , n ,
~ ~
we get:
l n l n l l n
M1 =-f. £. =- .2:1 (£1+e.) =- (n£') +- .E 1 e.; = £1 + M (4-20)
n i=l ~ n ~= ~ n n ~= .... e
SLnce the random errors c.'s are postulated to have a parent Gaussian
PDF N(O, cre; e),which implies that ~e = 0, then we should expect that
(4-21)
107
keeping in mind that by the unknown value £'we mean the unknown mean
Similarly, we get
2 1 n n m 2
(.ti-~)2 1 2 1 s 2
SL =n i~l =- i~l[ti-(.t'+ ME))
n
= - k .( c:. - ME) =
m.i=l ~ £
(4.22)
The above result indicates that the variance s12 of the sample L is
c:. This is actually why si is sometimes called the mean sguare error of
as:
( 4-23)
2, 7, 4, 8, 6, 4) is postulated to be
as follows:
10
1 l
M =
10 i~l
~.
J.
- 10 (50) = 5,
1
~~ =~ = 5 and o 2 = o~ ; s1 = 2,
transformations as follows:
e =(e.),
J.
i = l, 2, ... , 10 ·,i.e.:
R..-)..ln R-.-5
~
t . :
~ (j £
;v : _].__
2
•
' ~ = 1 ' 2 ' ... ' 10
.
Hence, the transformed sample that has
value £'or )..1 £. All w·e can get is an estimate £ for it from the
following equation
£: M_
--r,
= £'+ Me: = £'+ E' *) ( 4-24)
-
and hope that e:, in accordinace with the basic postulate of the error-
* From now on, we shall use the symbol i for the mean~ of the sample
L. The "bar" above the symbol will indicate the sample mean to make
the notation simpler.
110
Residuals with inverted signs are usually called corrections. It should be noted
a varia:bl e. The observed value £. is fixed and so is the mean £ for the
~
particular sample. In other words, for a given sample, the residuals can
called residuals and not errors, because errors are defined as s. = ( 9.•• - J1 n)
l l lv
"variable residuals" etc. which are not strictly correct. If one wants
ently. The difference v. between the observed value £. and any arbitrarily
l ~
(4.26)
should be called discrepency, or misclosure, to distinguish it from the
values vary with the choice of £0 • Hence one can talk about "minimization
formulae for computing the sample mean £ and the sample variance s2 Such
L
simplified formulae facilitate the computations especially for large samples
111
(4.22), (4.25) and (4.26). Here we state only the results, and the
and
r 1. = i.l. - i = v.l. - v.- (4.30)
We take ! 0 = 972.0 m,
10 1
v- =1-
10 i=l I: v. =10 (10.50) =1.05 m,
l.
._
No. . .1.
l.
•··
"· . ' .. :~.vl..= 1. - R.o
r
i
= R..J. - i r.
2
. 10
4
(m) l.
(m) = v.J. -
- v
J.
(m2)
= o.oo
The normal PDF (or its relatives) are by no means the only bell-
shaped PDF's that can be postulated. Under different assumptions, one can
view of fitting them to any experimental PDF. In other words the additional
with such PDF's are more troublesome. In this context let us just mention
that some recent attempts have been made to design a family of PDF's that
are more peaked than the normal PDF in the middle. Such PDF's are called
samples. We shall have to wait for any definite word in this domain for
some time.
Hence, the normal is still the most ~opul~r PDF and likely to remain
namely: The root mean square er·ror (RMS) mentioned in section 4. 7, and
the range (Ra) mentioned in section 3.1.5. Besides the RMS and the range
P(lrl < P)
e "p)e =0.50· = .P(lrl 4. 31) !(
.. I
which means that there is ·50.%. probability that. the resid:ual is smaller and
the CDF of the corresponding absolute vaJues of the sample residuals, and
take the value of r which corresponds to the CDF = 0.5 as the value of p
e
•
ae = f lxl ~ (x) dx
- CIO
oo (x-~ )2
1
= --=--- I Ix I exp (- - ~ ) ' dx • (4.33)
a I( 21r -oo 2a
X X
Similarly for pe' by taking the symmetry of the normal curve into account,
we can write:
(x-~ )
2
1 c~x-pe)
=a l(27r) ! exp (- X ) dx = 0.25 (4.34)
X _CIO
2a 2
X
and
It can be shown for the normal PDF N(Jl a · x) that "a " "a " and "p "
x' x' x ' e e
are related to each other by the following approximate relation:
or (4.36)
a : a ~ p ,;, 1. 0 : 0. 80 : 0. 67 .
x e e
The relative or proportional error r , of the sample L, is defined
e
as the ratio between the sample RMS and the sample mean, i.e.
re = s1 I I. (4.37)
In practice, the relative error isusually usedto describe the uncertainty
of the result, i.e. the sample mean. In that case, the relative error is
defined as:
....J-.1-e_=_S_Q_/_:€___,1 (4.38)
where S is the standard deviation of the mean I and will be derived later
I
in Chapter 6 .. In this respect, one often hears expressions like "proper-
tional accuracy 3 ppm (parts per million)", which simply means that the
(N (I, SL;t)) for the random sample L. It is very common to represent the
l[ I + SL ] I
or
the fact that the probability P(\1 2- a2 .::_ t .::_ 1-1 2 + ai~ is approximately 0.68 for
116
of £ is expressed as:
(4.40)
where K is determined in such a way as to make
The values (I - KS 1 ) and (I + KS1 ) are called the lower and the
Example 4.1: Let us compute the average error,the relative error and the
1 10 1
a =- l: Ir. I = 10 (1. 90) = 0.19 m •
e 10 i=l ~
[I - K s < £ < I + Ks ] .
L - - L
from the standard normal tables (see example 4.3) the last
II- B we get:
K = 1.96.
(In practice K =2 is usually used for the 95% confidence
that is:
or
[973.05 + 0.47] m.
N{t) ~
X
= 35 and cr X = 4.
The most probable error p is computed so that
e
P(ll~x - p e < X < 11
~x
+ pe ) =
-tp 0 t
= P(-t p-< .t < t )
p
= 0.50, (Figure 4.7a)
Figure 4.7a
where t
p cr
X
(4.35)):
118
P(x < y
- x
+ p )
e
= P(t <
-
t )
p
= 0.75 (Figure 4. 7b) .
p
e
= 4 tp = 4 (0.675) = ---
2.7 .
0 tf
Figure 4. Zb Note that in the second case of example t~.3, the value 3.375
is nothing else but the most probable error of the given random
variable h.
4.11 Exercise 4
1. Prove that the Gaussian PDF given by equation ( lt.• t~), has two points of
2. For the Gaussian PDF given by equation (!~.• 8), determine approximately
by integrating the PDF, then check your results by using the standard
normal tables.
3. Prove by direct evaluation that the standard normal PDF has a standard
l~. Show that the standard deviation a, average error a and the most pro-
e
bable error p of the normal PDF satisfy the following approximate
e
relations:
a a
e
1.0 0.80 ~' o. 67.
119
5. Determine: the average error, the most probable error, the relative
error and the 90% confidence interval of the random sample given in
6. Assume that the sample H = (-5, -4., -3, -2, -1, 0, l, 2, 3, L~, 5) is
following probabilities:
8. For the random variable in the previous problem, determine the values
Z. such that
l
9.
D
/'
/
/
/ h
- , . - ~.
are:
The field results of these observations are given in the following table:
121
Field Observations
.l(m) a 13 e
P-45.63 65° 32' 03" 37° 13' 08" 42° 53' 15"
.55 32 04 13 11 52 30
.59 31 59 13 10 53 00
.65 32 01 13 . 13 51 00
.58 31 58 13 06 52 15
13 12 52 45
..
51 .15"
53 00
51 45
52 15
(i, The micrometer of the vertical circle of the used theodolite was not
(ii) The nominal length of the used tape is 20m at the calibration temper-
is y =5 • 10- 5 / 1° F •
122
Required
( i) Compute the estimated values for the quanti ties 9.-, a, (3 and 6 •
(ii) For each of the above observed quantities compute its standard de-
(iv) Assume that each of these observed quantities has a postulated normal
parent PDF, construct the 95% confidence interval for each quantity.
(v) Compute the estimated value of the tower's height h to the nearest
centimeter.
123
5. LEAST-SQUARES PRINCIPLE
sample L = (~.),].
i = 1, 2, .•• , n, what is the value ~ 0 that makes the
v.]. = 2.
].
- ~
0
, i = 1, 2, ... , n, ( 5-l)
2
S* =-1n n
I (2.-2°)
i=l J.
2
=-1n n
.I v.
2
J.=l ].
(5-2)
value of 8* 2 .
(5-3)
We know that:
1 n -2 n
=- E [2(i.-R. 0 )(-l)] = - E (i.-i 0 ) = 0 ,
n i=l ~ n i=l ~
that is:
n
E (i.-i0
i=l ~
) =0 •
n n·
0 0
E R.. = E R. = ni ,
i=l ~ i=l
which yields
o n
R. =-n1 E R.. - R.
i=l ~
(5-4)
The result (5-4) is nothing else but the "sample mean" I again. In
other words, the mean of the sample is the value that minimizes the
endent of the PDF of the sample. This means that the sample mean I is
always "the mi:"l.imum variance estimator of R." whatever t.he PDF may be.
125
We say that the normal PDF, N(~~' cr~; ~) ; N(~ 0 , S*; ~) is the most
that ~the normal distribution N(~ 0 , S*; ~),at the same places as Lis
n
= II N( ~o' S*; L)
1.
. 8~.
~
i=l
be maximum with respect to the existing free parameters. By examining equation
(5.6), we find that the only free parameter is ~0 (note that S* is a function
of ~0 as follows:
Note that 8~'s are some values depending on L and therefore are determined
uniquely by L.
126
is (for the postulated normal PDF) again the value rendering the smallest
n
[ A( ~ 0 ) J = max I II N(~ 0 , S*; ~.) 8~.]
1. 1.
~ 0 sR i=l
n n (~. -~0)2
l
= max-, II II exp (-
1.
) 8~.
1.
1
~ 0 sR i=l S*I(27T) i=l 28* 2
n (£.-~0)2
= max l )n II exp ( -
1.
8~.] (5.8)
0 [( S* ll(21r) 1.
Q, sR
i=l 2S* 2
n
Here II 8~. is determined by L, and hence does not lend itself to maximiza-
1.
i=l
tion. It thus can be regarded as a constant, i.e.
n (!l,.-.Q.o)2
max [ A( ~ 0 )] =max II exp ( - 1
2 ) ] . (5.9)
i=l 2S*
R. 0 sR
Let us denote the second term in-the RHS of equation (5.9) by Q, which can
be expressed as:
n (~.-~o)2
Q= .Ill exp ( -x. ) , where x. = -1---~- (5.10)
1.= 1. 1. 2S* 2
n n
£n Q = .Q.n II exp ( -x.)) =
1.
E tn ( exp ( -x. ) ) ,
1.
i=l i=l
or
n
Q = exp ( E ( -x. ) ) , (5.11)
i=l 1.
127
n (,Q,.-,Q,o)2 n
J. ) l (,Q,.-,Q,o)2]
IT exp (- = exp [-- i: (5-12)
J.
i=l 28* 2 28* 2 i=l
n
[( l )
S*I(2'TT)
Since the only quantity in equation ( 5-14) that depends on ,Q, 0 is S*, we
can write:
n
[;x.(,Q,O)J =max [(..1.)] =
,Q, 0 ER S*
(S*)n will be attained for the same argument as the minimum of S* (see
Figure 5-l.
Figure 5-l
128
( 5-16)
as*
- as*-2=
-=- 0 '
that is:
n
a
-I. v.2 :::: 0 . (5-17)
].
ato i=l
We ha.ve thus shown that under the postulate for the underlying
PDF, the mean 1 of the sample 1 is the ~aximum probability estimator for
(5.18)
o.:f squares of discrE)ancies and that this property is required, for a lar~e
family of postulated PDF's, to yield the maximum probability for the underlying
129
PDF. Hence the sample mean ~' which automatically satisfies the condition
most probable value of the mean ~~ of the underlying PDF under the con-
and sufficient condition for the sample mean to be both the least squares
equivalent.
The whole development we have gone through does not say any-
thing about the most probable value of the standard deviation a~ of the
same principle is used for all kinds of estimates (not only the mean of
mind, namely
(ii) The least-squares principle does not tell anything about the
postulated PDF.
out the equivalence between the sample mean 9., and the estimate for the
parent population mean ~9., determined from the condition that the sum of
discrepancies be minimum. We have also shown that 9., is the most probable
normal or any other synunetrical PDF. We shall show now that the same
write:
(Ll, L 2 , Ls)
"}
L • ., " I
( 5-19)
j Q,j ) sRs
Lj (X,j
1' 9.,2'
... , n.
J
L = (Q,l, Q,2, • • • I
Q, s) t:Rs (5-20)
0 0 0 0
·- (V ,
l
v2 , ... I vs ) ( 5-21)
i.e.
S*~ 1
=-
n.
( 5-23)
J
J
. 2
'l'he minimization of the variances, i.e. minimization of each E[ (V'J) ] ,
( 5-25)
sati.sfi.es the cond:i.ti.on ( 5-2l~). On the other hand, the result ( 5-25)
be written as:
~
s
~(£) = IT N(£j, S*. Q, j )
j=l 0 j '
2
s (£j-£j)
l " 0
= IT exp [- ]' ( 5-2'7)
j=l s~l(2rr) 2S~ 2
J J
where Q,j is the random variable having mean ~j and standard deviation s~·.
0 J
Following a similar procedure as in section 5.2, we end up again with
( 5-28)
maximizes the probability that the members of the parent population will
,_.
occur at the same places as the members of the multisample L.
( 5-29)
5.5 Exercise 5
~ =J x ~ (x) dx
as* ---
2. Prove that---= 0, is the necessary and sufficient condition for the
&Q.o
rectangular (uniform) PDF, R( .Q. 0 _,__ S*; .Q.), to be the most probable
that the analytic expression for the uniform PDF is given in example
that had been obtained through some measurement or through any other data
between these two, since even the "primary" samples may be regarded as
although applied to random variable rather than sample (see sections 4.5 and
4.6), namely the transformation of the Gaussian PDF to the normal and to
the standard normal PDF's, respectively.
in the above two cases. As a matter of fact, it may not be even possible
to derive the sample at all from the primary sample which is usually
134
x, where
and
(6.1)*
X = (xl ' x2 ' . • . 'xn )
are called the mathematical model for the statistical transformation; and
l F ( .Q. , x) =0 J (6.2)
be formulated as:
x =F (t), (6.3)
* Note that ~ and x are nothing else but the multivariates corresponding
to the multisamples L and X respectively.
135
Example 6.1: After having measured the two perpendicular edges a and b
a. = f 2 (a, b) = ab •
x=[::l = [:1 =
(ii) "completeness" of the primary multisa.mple L, i.e. whether all its con-
stituents have the same number of elements in order to deduce the variance-
covariance matrix ~1 ;
(iii) our willingness to match the individual s-tuples of elements from the
usually do not even try to carry out the transformation and put up with
~
E1 for the primary multisample L, from which we then compute the statistical
E(L)
-
= E*(£), EL = E*t'
-
E(X) = E*(x), and EX= E~ (6.4)
i· = p£ and SL = a R.
for the univariate case as discussed in section 4.7. This postulate allows
us to work with continuous variables in the mathematical model and write it
as:
F(L, X) =0 (6.5)
understanding tacitly that each value X has· its counterpart L.
- A
From now on·, we shall write E for E(L),a.nd .X for the statistical
(if s < n). If it is solvable then we may still have~ distinctly different
cases:
137
The first case we have met in example (6.1) where the determina-
t.ion of X from L does not present any problem from the statistical point of
view. The only problem is to obtain L:X from L and L:L. This problem, known
F (L, X) = 0
To make things easier, let us deduce it first for one particular explicit
transition
We can write:
(6.9)
we can write:
E(X) = E(B L + C) =B -
E (L) + C =B -L + C.
Hence
EX =E ( (BL
I Ex = B EL BT. I (6.10)
L:L = 2 3 1
0 1 4
If a multisample X= (x1 , x 2 ) is to be derived from L
X
1
=
X= B L
2,1 2,3 3,1
i.e.
{::1 =
f: -:]
1
f::1
This indicates that the coefficients matrix B is given by:
B = [:
0
1
-3
0
I •
140
~
X
= B
2,2 2,3 3,3 3,2
-1
7
-121
1
{-3~ i]
0
= [39
5
5]
23
i.e.
matrix can be deduced even for a more general case, namely the non-linear
where
dF
dL
IL= Lo
(L - L0 ) =
s
L:
i=l
-, C!F
dR,. R,.
~ ~
=
141
Taking the first two terms only, which is permissible when the values of
the elements in~ 1 are much smaller than the values of 1i' we can write:
(6.12)
where is again ann by s matrix but this time composed from all the partial
B
ax.
~
derivatives alj *).Applying the expectation operator we obtain
.Q, =.Q.o
j j
realizing that E(F(L 0 )) = F(L 0 ) and E(L 0 ) =L
0 (because L0 is a
(6.13)
Subtracting (6~13) from (6.12) we get:
Ex
. (6.15)
* Explicitly' if we have
xl xl (.Q.l' Q,2' Q, )
' s
x2 x2 (.Q.l' .R.2' . '
.Q,
s
)
X = =
X
n
X (tl, .Q,2'
n
. '
.R.
s
)
case)of the more general explicit relation, yielding therefore the same
covariance law. It should be noted that the physical units of the individual
Example 6.3: Let us ;!:;·ak.e .again the example . 6.1 and form the variance
EL = [s~ sabl
sab s~ J
and the model is non-linear, although explicit, i.e.
X = F ( L) , or ( d, a ) = F (a, b) •
~ere:
.
B =
r aa
-aa
a~
Ml
a
3b
a
ab •
a.
B =
2,2
and by applyi.ng · the covariance law (equation ( 6.15 )) we
·. t!~
get:
I:
X
= sd•l
82
= B I:
L
BT
8ad a
Example 6.4: Let us assume that the primary multisample L = (a, b) which
quantities will be
i =l~J..
· a'·
= [II (a)~a + b~1>) 2 ]].;
where a and bare the estimates (means) of the two:meas'U.r.ed
Hence
144
X=
= [
142. 57 em
· 8016. 50 cm2
l
E1
l
After computing the variance-covariance matrix we get
·ro.oo4 0 em 2
~L = 0 0.0056
as statistically independent.
.6. 3) we get:
B =
-a
a"
b
b
A
-a
=
I 0.898
62~58
o.439l
128.1
~ =B ~ BT
X L
= [0.898 ro.~o4
0.~056]
0.4391 [ 0.898 62.581
62.58 128.1 0.439 128.1
= [0.0043
0. 5397
0.5397 ]
107 . 5627 , with units
[ 2
~:3 =~]
em •
Furthermore
EX = BI:LBT
s2X . (6.16)
If, moreover, 1 is assumed uncorrelated, we have
r1 = diag (S~ ,
1
si2 ' . . . ' (6.17)
s
8.2
X
= I:
i=l
B~
l
si i' (6.18)
matrix.
8 2 = (ay)2 8 2 + (ar)2 82
y aa a a~ s
= (;.,1) 2 (4) 2 + (-1) 2 . (3) 2 = 16 + 9 = 25 )
Example 6.6: Figure 6.3 shows a levelling line between two bench marks
t 1.,\on a 1 · t o 2
k
.II(, i 1
•
1. • e • 8h . = k ~.,
l.
where k is a constant.
l.
Let us deduce the exPression for the M8E of the overall level
s
~H = HC - HA = r hi •
i=l
The mathematical model in this case is
+ ••• + h
s
A Hence:
: 2· (~)2
Figure 6.3 2
8~H = (a~H)2
dhl sh + dh
82
h2 + • ..
1 2
s s
= r k ~.
l.
=k z ~.l. '
i=l i=l
levelling line A - C.
147
Let us consider the example 6.3 and assume that the errors in
82 = (~)2 82 + (~)2 82 1
=~ (a2 8 2 + b2 8 2)
d aa a <lb b a b
Note that the same results can be obtained from Example 6.3 immediately by
putting 8 ab = 0.
between the two derived quantities d and ~ , we have to apply the covariance
8 da = a~ ( 8 a 2 + sb 2) '
that is Sda ~ 0, and ~X (X= (d, a)) is not a diagonal matrix, even though
X = (d, a) for any further treatment in which case we cannot assume that
d and a are uncorrelated any more and we must take the entire ~X into
account.
Example 6. 7: Let us solve Example 6.2 again, but this time we will
Hence:
I
X = Bl: LBT )
L:x =[:
0
l
-:] 3
0
0
3
0
0
l
0
2
0 0 4 -3 0
=[3: l:l = r xl
82
8
x2xl
xh]
8
82
x2
law above.
x = 10 m, with S
X
=3 em,
-a = 90° with sa = 2 "
'
i3 = 45° with ss = 4"
Figure 6.4
'
s aS = -1 arc sec
2
and sXCI. = sx(3 = 0 .
It is required to compute the statistical estimates for
X= F (L), where
sin a.
y-· = _..;..._
2<
sin S
~
= _:;;;.__
sin y
.'
however the angle y is not observed, i.e. it is not an
a and (3 by putting
150
and we get:
y =X sin a
--~~~-
sin (a + S)
Z =X sin S
--~~~--
sin (a + S) •
z"' = 10 m.
(a, S,
(a, S,
x)1 = [x
x)
sin a/ sin (a + S)J
x sin S/ sin (a + S)
~ 2:/. 'Oy
\ 'Oet ' as ax
B =
az az 'Oz
aa ' as a:x:
z -y
sin(a + s) tan (a + S) X
=
-z z
tan (a +S) sin ( et' + S) X
82 s
a. sa.s a.x
= s2 s
IL ssa. s Sx.
s s2
xa. 8xs X
4 -1 0
= -1 16 0
0 0 9
B=
-Z(lOO) y( 100) z
X
p"tan(a. + S) p"sin(a. + S)
B =
[0.007 0.007 1.414]
0.005 0.010 1.000
and conseQuently
152
i.e.
S
y
= 118 = 4.2 em J
S
z
= l9 = 3 em.
The results of the above example show that the high precision
the precision of the measured length x. Hence, one can use the error
the accuracy of the derived quantities and decide on the needed accuracy
sometimes called systematic errors, for which the law governing their
X =f (L) , (6-19)
in which x is a single quantity, f i.s a single-valued function and
Taylor's series expansion, around the approximate values 1° = (~~' R.~, ... ,~~),
X
. f(L 0 )
=
of (L-1°)
+ aLl
L = 10
s
= X
0
+ L:
of (~.-~~) (6-20)
i=l F.i
J.
R, • = ~~ J. J.
J. J.
154
we get:
0
X
= a~.l *) ' (6-21)
I= T l(a 2 + 8 2 ) ~ (6-22)
with o being the non-random error and 8 being the M8E. Combining
the two errors in x as given above and using equations (6-18) and (6-21)
we get:
s of 2 s
T
X
= /[~; 1
~
a:r-l ali) +
i=l
z:
s af 2 2 2
= I[ i=l
z: ( (-;-;;-) ( o~. +8. ) } +
o~i
0~.]
l l J
or
s 2
T = ·'j t l: ( ..u:_) 2
T .. + q] ' ( 6-23)
x i=l a~. l:
l
*For the validity of the Taylor's series expainsion, we can see that
the requirement of o~i being small in comparison to ~i is obviously
essential.
155
same parameters.
Example 6.9: Let us solve Example 6.2 again considering the primary
covariance matrix:
We have:
3 ax 1 2
82 = I: (aT"") 82 = 39
xl i=l ~
~.
~
'
3 ax 2 2
82 = L: (aT)
x2
82
~.
= 15 .
i=l ~ ~
The influences o and o- due to the given non-random errors
xl x2
in L are computed from eguation (6-21) as follows:
3
0 = L:
xl i=l
= -1.5 + 0 - 1.5 = - 3
3 3x 2
0 = L:
().Q., o.Q..
x2 i=l l
l
= -3 + 2 + 0 = - 1 •
T
xl
= IK-3) 2 + 39] = 1[48] =~ '
s
where k was a constant and iAC =i=l
E i. is the entire length of
~
s a~H
o~H = i=l
E
3h.
oh.
~
~
where
= k'h. ~
Then we get
s s
~H.
o~H = i~l ;'.~: .•. k.'' i~l hi = k'
many decimal places like 1/3, 5/11, etc., by rational numbers with a
ways. We either truncate the original number after the required number
158
(6-24)
number of decimal :places and Int stands for the integer value.
7T =T= ~
= 3.141.
The second :process, i.e. the rounding-of~ can be described
by the formulae:
= 3.142.
It can be seen that the errors involved in the above two
* To normali.ze the number, say 3456.21, we write it in the form 3.45621 • 103.
159
8 =a - ;, E [0, 10-n)
aT
8 =a- ~ E [-0.5 10-n, 0.5 10-n)
~
and we may postulate that-8 has a parent random variable distributed
aT
according to the rectangular (uniform) PDF (see section 3.2.5):
0 0
PDF of rounding errors PDF of truncation errors
Figure 6.5
160
From example 3.17, section 3.2.5 we know that 0 = q/13, where q equals
-n
half of the width of the R. In our case, obviously q = 0.5 10 so
that (J = 0.289 10
-n
.
propagates according to the "total error law" and the errors in rounding
number x:
.x=f(L), ( 6-28)
where
L=(.Q..), i = l, 2, ... ,s
l
we can ~orrite the formulae for the errors in .x due to truncation and
s s
= l[(l: lL l:
i=l at.l i=l
l
(6-30)
12
This indicates clearly that the error in .x due to the rounding process
places;
161
decimal :places.
Solution:
1000
8 =I{[ E ~X . (0.5 . l0- 5 )] 2 +
~ i=l oai
1000 ~x 2 1 10
+ .E (-~-) ( 12 10- )}
~=l oai
• 0.005001 = 0.005 .
1000 2
(lL) ( - l . 10-10)}
8 = I{ E
XR i=l aa.~ 12
l . 10-10)}
= I{ (looo) (-
12
= l{lo- 8 (0.833)}
. 0.000091
'
which is much smaller than the corresponding 8
XT
162
when the maximum tolerable errors of the results, which are usually
tolerance limits.
for both the random and the inevitable non-random (systematic) errors.
we hence expect the random errors to have the parent Gaussian PDF,
the actual results should not have the total error, composed of the non-
IT ~ 1{0 2 + 2
(3a ) ), (6-31)
with a relative error (see 4.10) not worse than 10-4 , using a
precision not better than 3cr < 1 mm, i.e. tolerance limits
segment?
Solution:
error in D, is given by
TD = lOOCm. 10
-4 = 0.10 m = 10 em.
50
D = i=l
L: d~, where d. = -12 (F.+B.)
... 1 1 1
we get:
50 3D
L: cd.
i=l
ad.1 1
where
Hence,
50
L: 1· c.1 <- 50 mm =5 em.
i=l
164
em2
or
2 75 cm2 , :8-,. 33 2
aD !.. -9 = _.. - em
by a we get:
and
2 2
o ~ 2od = 2(0.lb) = 0.33 em.
~. =F.-B.
2 2 2
Then:
2
3~. 3~. 2
2 2 2 2
a~. = 0~ = (3F~) oF. +(3B~) OB
2 2 2 2 2
= (1)202 + (-1)202
= 2o 2 •
2 2 - 2
o~ ~ 2o = 2(0.33) = 0.66~ em
or
of the differences ~'s with~ o~, i.e. within+ 0.8 em, and
95% of~ within~ 2o~, i.e. within~ 1.6 em. These specifications
F(L, X) =0 • (6.33)
to generally f.it the abbve model for whatever X we choose, i.e. yielding
Denoting
L - E =v (6.34)
we may reformulate our mathematical model (6.33) as:
"
F(L, X) = F(L + V, X) =0 (6.35)
where V is called the vector of discrepancies.
Note that V plays here very much the same role as the v's
have played in section 4.8. From the mathematical point of view, there
ment problem.
model
X =~ (6.36)
different from each other and cannot therefore all satisfy the model.
j = 1, 2, . . . , m (6.37)
where the v's are the discrepancies. We have to point out that, although
we seek now the same result as we have sought in section 4.7, t.he formu-
on. While we have been taking all the n observations into account in section
4.7, we shall now work only with them distinctly different values 1.,
J
168
m: -
I: t j
i = j=l. P(ij·· }= ~ i."J
.1=1
PJ ' (6.38)
rather than the first (equation (3.3)) as used in section 4.T. Here,
according to section 3.1.3, Pj = cj/n with cj, being the count of the
m
X= I: . 'l pj J (6.39}
j=l . .1
or
(6.40)
in vector notation,
(see section 3.1.3}. Note that, with the weights being nothing else but
which we are more "certain", i.e. which are repeated more often in the
* L = (i , "i , • • •, "i )
can be regarded in this context as a sample of
"groupe~" o~servation~, i.e. each constituent t , j = 1, 2, • ·• .,m,
has a count ( frequenby). cj associated with itjin the original sample
L •
169
.
we were see k ~ng such Nno as t o rnak e
1 2 n n o 2
v. l: = ;1 l: (~ - ~ ). (6.41)
. 1 ~
n ~= i=l i
m 2
min [ l: P.v. (6.42)
~o 8 R j=l J J
( 6. 43)
(6.44)
general since we can regard the former formulation, i.e. euqation 6.41 as
the observations i!. Hence we shall use (6.43) exclusively from now on.
~
The same holds true even for the two formulae for i and we shall use
equation (6.40).
m A
s 2 = 2
E P. v j (6.45)
L J
j=l
or in matrix notation,
[s L
2 = VT PV.
l (6.46)
does not depend on any specific underlying PDF. If L has got normal
A
parent PDF (or any symmetric distribution), xis the most probable estimate
sample ? In other words, we may ask what value of variance can be as soc-
i.e.
s* X
= ( 6. 4 7)
Here l:L is not yet defined. All we know is that L- (i 1 , i 2 ,•. im)
vations uncorrelated and let us also assume that there can be some
On the other hand the value of x (i.e. the sample mean) can be
i.e. the ungrouped observations ti' i = 1, 2, ... , n, which all have equal
n
1
X = n1 l: £. =
~ n
(6. 50)
:.i=l
2
Hence, we can compute the variance of the mean, i.e. S", again by applying
X
n 2 n
l: = (!) l: (6.51)
n
X i=l i=l
in which all the 'varianc~ s~. are again assumed to have the same value
~
2: ( ~. ;;.. x) . (6.52)
~
i=l
s~ = (6.53)
X
which indicates that the variance of the sample mean equals to the variance
of the sample computed from equation (6.52) divided by the total number
m
I: (P 2
j
j=l
n •
173
8 2 n 8 2
L 1
--=
n n L: (l)
n
i=l
Using the oame manipulation as in section 6.4.2 when dealing with the
V's and ~'s, and also earlier, in section 3.1.4 when prooving equation
m 82
L
E [ PJ. (--;-) ], (6.55)
j=l
which can
K, j = 1, 2, . . . , m, (6.56)
8tj = p
K
j
' j = 1, 2, .•. , m, (6.57)
which shows that in order to get the correct result from (6.49) we have to
variance, i.e.
(6.58)
174
= 1·s 02 = K (6.59)
2
where s 0 , constant for a specific sample, is known as the variance of unit
whose weight equals to one. In the case of sample mean I, S0 equals to si.
From equations (6.46) and (6.53) we can write:
S~ = ~lPv ( 6. 60)
x n
hinges on the acceptance of the "variances" s~. and s 2R:.. They have been
1 1
introduced solely for the purpose of deriving formulae (6.53) ,(6.58) that
are consistent with the rest of the adjustment calculus •. The more rigorous
has also a standard deviation Si associated with it. This standard devi-
number of observations.
We can now ask ourselves the following question: Does the mean
-
L of a multisample L also have a variance-covariance matrix associated with
s2- s- - s- -
~1 ~1 ~2 ~l~s
E- = s- - s2- s- - (6.61)
L ~2.Q,l ~2 .Q,l~s
s- - s2-Q,
2s 21 s
175
where
l 82
n. !/.,.
1 1
and
= Ls
n. !/.,,!/.,,
1 1 J
Here we have to require again that ni = nj, i.e. that both components of
the multisample have the same number of elements (see section 3.3.5).
we have
= ns = n
and
(6.62)
By analogy, the variance-covariance matrix obtained via the
the mean of multisample is associated with the mean of the derived multi-
r"" = BEEBT I
is the variance-covariance matrix of the statistical estimate X, i.e. of
(6.53)
X = F(L) •
Development of these is left to the student, who should also compare results
vector X.
__o 2
--..;....;·0;..:50;..::.5.;:;.6...:c;::;m=-- = 0.0011 em2
E- =[ 0. 0008 0 ] 2 1
em = 5 ~L •
L 0 0.0011
or
2
zA =fo. ooo81 em 0.1079 cm3 ]
4
X 0.01079 21.51254 em
given by
min n
A
(6.64)
XER
The· condition (6. 64) for the majority of mathematical models, is enough to specify
2 2 2
P = diag (K/Sl , K/S l , . . . , K/S l ) ,
l 2 m
that is
2
1/S I ).
m
Using the notation developed for the multisample, this can be rewritten as:
Q
which indicates that the matrix P is obtained by multiplying the constant
(6.65)
sample L to be uncorrelated.
ference between a sample and multisample - they can be hence treated in much
the same way. Thus, there is not basic difference between the apparently
178
trivial adjustment of the sample mean and the general problem of adjust-
A'
ment. The only difference is that in the first case X is a vector of one
K (K having been a scalar equal to s?X in the adjustment of the mean of sample),
in the least squares method, where X has several constituents. Let us just
say at this time that we usually compute the weight matrix P, as it is called
p =K l: -1 (6.66)
E
where K is an arbitrarily chosen constant, the meaning of which will be
shown later. This can be done because, as will also be shown later, the
. multisample X;
AX =L , (6.67)
179
BL =C (6. 68)
non-linear by nature, can be linearized. This is the reason why the two
AX + C =L (n > u) (6.69)
AX - (L + V) =0
or
180
observation equations.
" that would minimize the quadratic
We wish to get such X = X
form VTPV in which P is the assumed weight matrix for the observations
-1
From equation (6.66) we have P = K E_ , where K is a constant scalar and
L
Er; is the variance-covariance matrix of L. Since E- is symmetric, the
L
(6.73)
and we obtain, writing d/dX for the whole vector of partial derivatives
d/dXi,
abbreviated form:
~
N X= U (6.76)
equations, or simply the normal equation matrix and U = ATPL is the vector
+ Note that the normal equatiomcan be obtained directly from the mathemati-
cal model by pre-multiplying it by ATP •
182
given by
(6. 77)
vector X, let us use a different weight matrix, say P', such that
=X
This result indicates that the factor K in equation (6.66) for computing
the weight matrix P from LL' can be chosen arbitrarily without any influ-
A
section 6.4.4.
for this purpose and this is the convention we are going to use (see also
A
6.4.2). The values v. are computed directly from equation (6.70) in the
~
same units as these of the vector L. - Then the adjusted observations will
be given by £ =L+ V.
will be
x = ~l + v 1 , with weight p 1 ,
A
x = tn + v , with weight P .
n n
Or, in matrix form
AX
A
=L - A
+ V
where
1 ~
1 iJ.
2
A= -L
.
1 t
n
184
Exa;mple 6.16: Let us have a levelling line connecting two junction points>
lengths d., i = l, 2, 3.
~
A X =L
3,2 2,1 3,1
where X = (Hl' H2) and
Hl = hl + HG = Ll )
-H1 + H2 = h2 = L2
-H = h - H = L
2 3 J 3
1 0
-1
0
1
-1
[::l =
Hl = HG + ( hl + v 1 ) '
-Hl + H2 = ( h2 + v 2) '
H2 = -HJ + (ii 3 + v 3 ) .
as:
V = A X E
3,1 3,2 2,1 3,1
186
where:
vl (iil + HG)
v
3,1
= v2
v3
X =
2,1 [::1 1 =
(h3
ii2
- HJ)
'A
3,2
= r-~ ~1
0 -1 •
We assumed that the observed values hl' h2 and h 3 are
-
uncorrelated. We will also assume that HG and HJ are
errorless, Hence:
2
L:- = diag (S-2 ' S- S-2
1 h. h2 l
' h3
But S-2 is proportional to d., i = 1, 2, 3;
h. l
l
thus
p = K L:-
1
-1 .
= d1ag
1 1
(d, d' L) .
1 2 d3
Applying the method of least-squares the normal equations
are
A
N X = u
2,2 2,1 2,1 '
where
1
N = ATPA = -1 0 0 1 0
dl
[: 1 _: J 0
1
0 -1 1
d2
1
0 0 0 -1
d3
This giv:es
187
1 1
(L +d),
dl
--
d2
2
N
=
2,2
--
1
d2
'
1 1
( -d2 . + -
d3
)
and 1
-l 0 0 (hl + HG)
T-
U = A PL =
r: 1 _:l dl
0
0
1
d2
0
0
.L
h2
(ii 3 - HJ)
d3
Hence
u =
A
The solution X is given by
(\
-1
X N u
2,1 = 2,2 2,1
where
-1 dl d2 d3
N (L + L)
= (dl+ d2+ d3) d2 d3
1 (.L + L)
d2 ' dl d2
V = AX - L and find
i = 1, 2, 3.
188
L=L+V.
Remembering that HG and HJ are assumed errorless we get:
h.l. = h.l.
+ v.'
l.
i = 1, 2, 3.
1 a c 6.16 4
c
2 a d 12.57 2
4 a b 1.09 4
5 b d 11.58 2
6 b c 5.07 4
2
Assume that the variances Sh., i = 1, 2, .•. , 6, are
" l.
proportional to the corresponding lengths t.. The elevation
l.
p6ints b, c, and d.
189
Solution:
A X =· L 1
6,3 3,1 6,1
where
X = (~, He, Hd) •
3,1
The 6 independent observation equations will be(one
hl + v·l = Hc H
a = Hc o.o = H
c'
h2 + v 2 :::
Hd H
a
::: Hd - o.o = Hd,
h~ + v 3 :::
Hd H
c
h4 + v4 :::
~ H
a
:::
~- o.o :::
~ '
h5 + v 5 = Hd ~'
E6 + v 6'· = Hc - ~ .
The above set of equations can be rewritten i<l the following
v
1
=
/.
H
c
6.16 ,
A
v2 :::
Hd 12.57
A
·'·
v3 = -Hc + Hd 6.41
A
'
v4 = Hb 1.09
'· .,
v
:J
= -~ .... "Hd 11.58
/'.
v6 = -~ + Hc 5.07
In matrix form we can write
190
v ... A X E
6,1 6,3 3,1 6,1
where
vl 6.16
v2 12.57
v3 H 6.41
b
v = v4 X =
H
E = 1.09
6,1 3,1 c 6,1
v ..) 11.58
Hd 5.07
v6
and the design matrix, A,is
0 1 0
0 0 1
0 -1 1
A = 1 0 0
6,3
-1 0 1
-1 1 0
h.,
l.
we will treat them as uncorrelated. Hence, the variance-
EE = diag (~, 2, 2, 4, 2, 4)
6,6
N X & u
3,3 3,1 3,1
yielding the solution
-1 u
X = N
3,1 3,3 3,1
where
191
N = AT P. A
3,3 3,6 6,6 6,3
Thus:
0 0 1 -1 0.25 0 0 0 0 0
N=
[:
0
1
-1
1
0
0
0
1
-:] .0
0 0.5
0
0
0.5
0
0
0
0
0
0 l(
0 0 0 0.25 0 0
0 0 0 0 0.5 0
0 0 0 0 0 0.25
0 1 0
0 0 1
0 -1 1
X
1 0 0
-1 0 1
-1 1 0
and
[ 0 0 0
N ;: 0.25 0 -0.5
0.25
0
-0.5
0
-0.25
0.25
J 0
0
1
0
0
1
0 0.5 0.5 0 0.5 0 0 -1 1
1 0 0
-1 0 1
-1 1 0
Finally:
N = ll.OO-0.25
-0.25
1.00
-0.5 ]
-0.5
3,3
-0.5 -0.50 1. 50
Hence:
0.81.6
0.81
N
-1 = [ 0.8
1.6 0.8
3,3 o.8 o.8 1.2
Computing U =ATPL- , we get.
-0.25]
U=l0.~5
0 0 0.25 -0.5 6.16
0 -0.5 0 0 0.25 12.57
3,1 0 0.5 0.5 0 0.5 0 6.41
1.09
11.58
5.07
and
-6.78501
u = [ -0.3975 .
3,1 15.2800
Performing the mult~plication N-l U, we get X as:
1.6 0.8
o.8J [-6. 785ol 1.051
x [
= o.8 1.6 0.8 -0.3975 = [ 6.16
3,1 o. 8 0.8 1.2 15,2800 i2.59
Therefore, we have obtained' the following· estimates
~ = 1.05 m,
H
c
= 6.16 m,
Hd = 12.59 m.
V =A X - L .
193
Namely:
o.oo m
0.02 m
v = 0.02 m
6,1
-0.04 m
-0.04 m
0.04 m
and we get:
0.04
11.54
5.11
in metres.
points b, c and d from Ha using the adjusted hi. The resulting values
must not differ from the adjusted values Hb' He and Hd.
(6.77), i.e.
194
X= BE (6.81)
where
(6.82)
we get:
LX= BEE BT. (6.83)
=K
that is
("X =K -N
On the other hand, by. putting p ·=
-1
=K (ATPA)-l,
KL-
,;.1
1
in (6.86) we get
(6.86)
L
E" =! K (AT E--1A)-l = (ATEE -lA)-1 , (6.87)
X 1C L
which shows that EX does not depend on the choice of the factor K. In fact,
this statement is valid only if we know the correct values of the elements
195
VTPV = ( n - u) ~ • (6.88)
The multiplier in the right-hand side is nothing else but the difference
df=n-u. (6.89)
df must be greater than zero in order to be able to perform a least-squares
K = (6.90)
Usually)in the literature, K is known as the a priori variance factor and
* Here, the vector V is the vector of residuals from the least squares
adjustment .
196
" -1
l:"
X
= K" N = K(ATPA)-l
"T"
V PV (ATPA)-l
= (6.91)
df
(6.91) we get:
"T "
~~ = y (V PV) ~ ( Tp )-1 = ~"
6x df Y A A 6x
The above result indicates that fx given by equation (6.91) is independent
of the choice of the a priori variance factor K. We recall that the same
"
holds true for the estimated solution vector X (equation 6.79).
Then we have to take into account their respective weights. We know that
see that the matrix of normal equations, N, can be immediately used as the
weight matrix of the vector X, since the inverse N-l is proportional to the
-1
variance-covariance matrix Z:~. Accordingly, the matrix N is known also
as the weight coefficient matrix and the square roots of its diagonal
sample mean (see 6.4.3). It is left to the student to show that the
On the other hand, using eq. (6.91), we obtain the estimated variance of
the mean x as
(6.92)
n n
l: (~. _ I)2 = 1 l: (6.93)
1. n - 1
i=l i=l
(compare with eq. 3.6) which is used in statistics wherever the mean I of the
show that using the estimated variances for the grouped observations (see
and
2 n A
A
1 2
S- = E pi (6.95)
~ n ~
(n-1) z pi i=l
i=l
of the variance factor K, introduced for the first time in 6.4.5. Let
write:
2 2
pl sl = P2S2 = •.• = Pnsn 2 = K
(6.97)
Comparison with equation (6.59) gives some insight into the role the
since it ties together formulae (6.66) and (6.65), where K can be also
2 2 2 A A A
~
cr
2 vrpv
=-- (6.98)
0 df
~ 2
~A
= cr Q. (6.99)
X 0
~"
X
" in example 6.16.
of the adjusted parameters X The
p . 1 1 1
3,3 ;:: .-diag [d, d' d]
1 2 3
and df =n - u =3 - 2 = 1.
Hence,
[1, 1, 1],
and
A 2 - 2•
cr
0
(HJ - HG -~hi) /
-1 l.
As we have seen, N = Q is given by
dl d2 d3 d2 + d3 1
-1
Q =N = d2d3 d2
2,2 ~:ldi
dl + d2
L •
d2 dld2
"T
v = [o.oo, o.o2, o.o2, -o.o4, -o.o4, o.o4]
1,6
in metres,
df =n - u =6 - 3 = 3.
Hence
and
= 0.002
3
.!.
- 0.00067 ( uni tless) •
Q = N-1 = [1.6
0.8
0.8
1.6 o.BJ
0.8 2
in m .
3,3 o.8 0.8 1.2
Finally,
~"
X = ~ 0 2Q = 10-4
10.67
5.33
5.33
10.67
5·33]
5.33 in m ,
2
3,3 5.33 5.33 8.0
~:~~
or"" [10.67 5.33
'Ei = 5. 33 10.67 ] in cm2 •
5.33 5.33 8.0
201
other words: taking P =~ 2 ~L-l and seeking min VTPV, provides such a
0 XERU
solution X that satisfies at the same time the condition
applied to random multivariate (section 5.4) and we are not going to prove
it here.
n 1
q,(LJ3;L) = II exp [ - (6.101)
i=l s.I(27T)
~
n
1 1
q,(L 0 ;,S;L) = - - - = - n - - exp [--
2
L
(27T)n/2 II S i
i=l
i=l
1 1 T
= n
exp [- 2; V PV] ,
II s.l
i=l
202
T
which is maximum if both V PV and trace (EX) are minimum. This is valid
for any fixed K.
purpose of getting the solution vector X along with its EX' we can assume
2
any relative weights, i.e. P = cr 0 EL -l , with cr 2 chosen arbitrarily. On
0
T
the other hand, t~e~mai"irix
of no}"lna;l':lquations, i.e. N =A PA, and the
... 2 AT A
model with the observations and to test the correctness of the assumed
looking into the assumed LL and use the obtained cr~ from the adjustment
and covariances of the observations are beyond the expected range known
"overflow" into the value of the quadratic form VTPV and con~equently,
"2 •
into a
0
obtained &2 ~
0
0.00067. Thus the ratio a 2
e
;cro2 equals to 1500 which is con-
matrix E- was postulated too "pessimistically" and that the actual variances
L
B L = c '
(r < n), (6.102)
B (L + V) ~ C =0
or, as we usually write it
B- ·v + w = 0 *)) (6.103)
where:
(w= BE-c. (6.104)
crepencies and W is. the vector of constant valu~s ~. We re9all that "n" is
the number of observations and "r" is the number of j,ndependent
F(TJ) = F(L 0 ) + ~~ I
1=1°
(L-1°) + • • • '
"
We wish again to get such estimate V of V that would minimize
T
the quadratic form V PV, where P = ao2 E-L -1 is the assumed weight matrix
. . T
of the observations L. The formualtion of this condition, e • mJ.n V PV,L
Ve:Rn
is not as straightforward, as it is in the parametric case (section 6.4.6).
This is due to the fact that V in equation (6.103) can not be easily
denote
lV = -P-1 BT K .} (6.106)
B(-P-l BT K) + W = 0 ,
or
·~-lBT) K= W l (6.107)
MK = W, (6.108)
where
(6.109)
(6.110)
Once we get the correlates K, we can compute the estimated residual vector
L=L+V. (6.111)
the formula for the adjusted observations 1 can be written in terms of the
1 =1 + v
=1 - P-l BT K
=1 - P-lBT(BP-lBT)-l (B L - C) . (6.112)
[ L = (I - T) L+ HC ) \ (6.113)
(6.114)
Example 6.20: Let us solve example 6.16 again, using this time the"con-
L:h.l = llH.
After reformulation we get:
B v + w = 0
1,3 3,1 1,1
where
[ vl
B = [1, 1, 1], v = v2
v3
and
w = ii + 112 + h3 - llH
= L. h.l - llH .
i
example 6 .16) :
P (l 1 L)
3,3 = diag dl' d2!' d3
and
by equation (6.108) as
.M K = w
1,1 1,1 1,1
where
L:d .•
i l
208
K = M-1 W
1
= 'i:'A'"
.d.
U:ii. -
i 1
AH).
1 1
(6.106) as:
A
V = -P-1 T
BK
0 0 1 r-
dl
e.h.
J.
- AH
=- 0 d2 0 1 Ea.
1 1
0 0 d3 1
dl
~h. - H
1 J.
-- d2 L: d.
i J.
d3
and we get:
AH - th.J.
v.
,J.
= L:
l.
d.'
1
idi
This is the same result as obtained in example 6.16 when
i = 1' 2' 3.
209
This yields:
llH - Eh.~ ~
h. = h. + d ••
~
~ ~
~d.
~ ~
H ::a
HG + hl
1
= HG - dl
+ hl + - - (llH - ~iii J
~d.
~·
~ ~
and
H2 = HJ - h3
d3
= HJ - h3 - (llH - Eh. ) •
~ ~
BU.
i ~
Example 6.21: Let us solve example 6.17 again, but this time using the
for convenience.
No. of observations, n = 6,
a c
No. of unknown parameters, u = 3.
Figure 6.8 Then df =6 - 3 = 3, and we shall see that we can again
loop I = a - c - b - a ,
loop II = a - c - d - a >
loop III =a - d - b - a •
v 2 - v 4 - v 5 + (ii 2 - ii 4 - ii 5 ) =0 •
matrix notation as
B V + W = 0 ,
3,6 6,1 3,1
where:
211
1 0 0 -1 0 -1
B = 1 -1 1 0 0 0
3,6
0 1 0 -1 -1 0
rl _ 1
1,6
and
ii4 _ ii6)
w = (hl - h2 + h3)
3,1 (h - h4 - h )
2 5
0.0
w= 0.0 in metres •
3,1
-0.1
and
-1
P = diag (4, 2, 2, 4, 2, 4).
6,6
The normal equations for the correlates K are
M K = W
3,3 3,1 3,1
where
212
12 4 4
4 8 -2
4 -2 8
By inverting M we get:
= o.oo
0.02
0.02 m
-0.04
-o.o4
0.04
and are again identical with the results of example 6.17.
L =E + V , i.e.
in metres.
213
For instance;
= L-lr - T
IL- T - TL-I + TL-T •
T (6.115)
L L L
T
It is not difficult to see that both (HET ) and (TLEI) are square symmetric
T
= L--
L
TL- (2I- T
L
). (6.116,)
knowing that P = cr 0 2 LE
-1
, i.e.
we get:
(6.ll7)
Noting that
we get finally
r-~-LL_=_cro-2-P--1-(-I--_B_T_(B-P---1-BT_)___l_B_P___l_)-.-,]
(6.118)
~ 2
cr (6.119)
0
matrix L£· Under the same assumptions as stated in section 6.4.8, the
are satisfied then T and P-l will be diagonal matrices. On the other
A
hand, we can experience uncorrelated L even for some other general T and
P.
Example 6.22: Let us determine the variance-covariance matrix "' for the
LA
L
conditional adjustment formulated in example 6.20.
We have
b.H - ~h.
~ ~
v.
~
= d
i
i = 1, 2, 3,
~d.
~ ~
p 1 1 1
and = diag (d, d).
d2 ' 3
3,3 l
Thus we get
AT A (tlH - :EE. )2
"2 VPV 'l
0
o
= ----
r
= ------~~
4:d.
l l
6.121, i.e.
Hence,
0 0 1
dl [ LJ [1, l, l]
I: d.
T = 0 d2 0 l l l
0 0 d3 l
and we obtain
dl dl
d2 d2
d3 d3
Further we get
a.l a.l
a.3 a.3
where a..
l
= i = 1, 2, 3.
Finally we get:
217
E - 2 (I - T) P-l
h
= cr 0
3,31
2 al al al
(AH - l.Eh·)
• l.
= Ed.
a2 a2 a2
i l.
a a3 a
3 3
.1\
We have
in metres,
r = df = n - u =6 - 3 = 3.
Hence,
hT "
V PV = 0.002 (unitless),
"T A
V PV
" !
cr
o
;i:-
.~ r
= 0 · 3002 = 0.00067 (unitless).
6.121 as
E"
L
= cr"o 2 (I - T) P-1
6,6
where
-1
P = diag [4, 2, 2, 4, 2, 4]
6,6
218
T = p -1 (BTM-1 B).
6,6
M
-1
= (BP-lBT)-1 = I have:
-0.1
-0.1
0.15
-0.1
0.2
0.1
-0.1
0.1
0.2
l
and
[~ -n
0 0 -1 0
B = -l 1 0 0
3,6
1 0 -1 -1
Hence J
(BTM-1 B) =
0.15 -0.1 0.1 -0.05 0 -0.05
-0.1 0 •. 2 -0.1 -0.1 -0.1 0
0.1 -0.1 0.2 0 -0.1 0.1
=
-0.05 -0.1 0 0.15 0.1 0.05
0 -0.1 -0.1 0.1 0.2 -0.1
-0.05 0 0.1 0.05 -0.1 0.15
and: T = p -1 (BTM-1 B) =
6,6
'-
-0.2 0 o.4 ().2 -o.4 0.6
Hence
(I - T) p -1 =
219
Finally we get
~ 2
L:~
L
= (J
0
(I - T) P-l as
6,6
in metres squared.
6.5 Exercise 6
found to be:
with S- = 3 em,
a
1.2 · 10- 4 . The RMS ofT is 0.08 centigrads (1 grad= 100 centigrads).
221
x = 100.0 m y = 200.0 m
x0 = 150.0 m y0 = 150.0 m
to the formula
xl, Yl A A
x2 and y2.
of an unknown station P.
x2 = 546.4 m
E -
(x2, y2
- ) =l -o.j em
2
L~
y2 = 300.0 m -0.5 3
places.
places.
/
/
9. To determine the height h of a tower, the /
a = 100 m.
a precision of 2" (i.e. s-a ,; s--S ~ 2"), what are the required precisions
(i.e. Sa and s8) such that their contributions to the standard error Sh
of the derived height h - which is specified not to be worse than
several times, from which it was found that the standard deviation
SI, not worse than o': 5, how many times should we measure each angle
12. The following table shows the means I. of the daily measurements of
~
the same distance t during a five day period, along with their
I. (m)
~
101.01 100.00 99.97 99-96 100.02
SI _(em) 2 1 4 5 3
~
225
Required: Compute the weighted mean of the distance ~' say i, along
with its associated RMS, i.e. Si.
g
0
= 979832.12 mgal at the initial
point P is known.
0
observation.
Station
From To I:J.g (mgal) b.T (hr)
p pl - 9.82 2.5
0
p2 Po -27.78 1.5
pl p2 +38.42 2.0
Assume that the observed differences I:J.g's are uncorrelated, and their
HA = 300.000 m,
A
HB = 302.245 m.
Section Station h. t.
J.. J..
No. From To
(m) (km)
1 pl B 1.245 1.0
2 A pl 0.990 0.5
3 pl p2 0.500 1.0
4 p2 B 0.751 1.0
5 p3 B 1.486 0.5
6 p3 p2 0.740 1.5
Note that the arrows in the giYen figure indicate the directions of
~ fi ~
(i) The estimated elevations H1 , H2 and H3 of points P1 , P2 and P 3 .
adjustment.
16. Two physical quantities Z and Y are assumed to satisfy the following
linear model
Z = aY + S ,
where a and S are constants to be determined. The observations Y.l
and Z. obtained from an experiment are given in the following table.
l
y l 3 4 6 8 9 ll 14
z 1 2 4 4 5 7 8 9
Assume that the Y's are errorless, and the Z's were observed with equal
precision.
A
17. Solve problem No. 16 again, but this time consider the Z's error-
less, and the Y's with equal variances. Compare the results for
2 2 28 22 17.70
3 5 110 29 25.02
4 3 125 53 33.67
5 2 25 44 09.30
6 2 29 19 17.50
7 5 55 03 29.32
8 3 68 33 32.33
Assume that all the measurements were done with the same instrument
given by:
D.ACD, E: = 1'!515
A D
The results of the direction observations are summarized in the
following table .
... ,
..
Occupied Target Direction Observed Direction
Station Station No.
c 30 00 00 00.00
B D 36 28 22 l/.26
A 38 110 29 2T.l3
D 26 00 00 00.00
c A 27 29 19 17.52
B 28 35 03 26.80
A 11 00 00 00.00
D B 12 35 07 29.00
c 13 68 33 32.60
230
coordinates are:
X y
A: 0. 0 m, 0.0 m,
E: 200.0 m, 0.0 m.
A
B 1 00° 00' oo'!o
c 2 60 00 10.0
D 3 00 00 00.0
B c 4 6o 00 05.0
A 5 119 59 50.0
A 6 00 00 oo.o
c B 7 59 59 55.0
D 8 120 00 00.0
E 9 180 00 05.0
E 10 00 00 oo.o
D c 11 59 59 45.0
B 12 119 59 55.0
A 13 00 00 00.0
E
D 14 60 00 15.0
freedom.
(iii) The approximate values for the ,x, y coordinates of points B,C,D.
vector ~E.
232
2".
233
APPENDIX I
Gaussian PDF
m + 1 values
s.
~
= (2. ~
- m)~ i = 0, 1, . . . m, (I-1)
corresponding to particular distinguishable combinations of the m elementary
errors.
because there are only m + 1 different values available. The counts, ci'
of the individual values e. (see section 3.1.1) can be computed using the
l
m i
c.=(~)= m (m-1) (m-2) ..• (m-:i+l) = II j/ II j. (I-2)
l l i ( i-1) ( i-2) . . . 1 j=m-i+l j=l
c
P(.e.)
l
= .i = (~)/2m
l '
(I-3)
n
(see section 3.1.2).
(iii) The above formula describes the actual PDF of our sample e
the analytic expression for the "corresponding" (we shall see later what
finite differences.
oP(e.)
l
= P(e.)-
l
P(e.l -1 )
and we get from e~uation (I-3)
6P(e.)/P(e.)
~ ~
= 1- i/(m-i+l). ( I-4 )(
(I-1)
2i - m = E. I !J.
~
or
=1 - 2e/8e
= _______ ..;;..__
2E - OE
1 + m/2 - e/oe (1 + m/ 2) 0E - E
e which is defined the same way as the· discrete e in (i) with the exception
that m is let to grow beyond all limits, i.e. m ~ oo, By letting m grow
would contradict our experience teaching us that the errors are always finite
m grows to infinity, the absolute value o.f elementary error t:. grows to zero,
equation as
0£ + 0 0£ + 0
which is nothing else but an ordinary differential equation for the con-
d P( £) 2£ - de:
- - ~----~-------
P(e:) (1 + m/2) de: - s
To simplify the solution of this differential equation let us multiply
both the numerator and denominator of the right hand side by de: and assume
dP •
-p=-
2 e:de:
C/2 = - -c4 e:de: (I-6)
!~ = - !~ e:de: + const.
4 e:2
. - c:2 + const.
The question now arises whether we are free to regard both K and
C as two independent parameters of the above PDF. We know that the basic
00 00
2
f -00 P(e:) de: =J -00
K exp (-2e: /C) de: =
00
2
=K J exp (-2e /C) de =1
-oo
00
and K =1 I I
-oo
2
exp (-2e /C) de .
Hence the answer is that K must not be regarded as an independent parameter.
We obtain
00 2 = .; Crr
f 2 !0 exp (-2~:: /C) de (I-9)
-00
2
r,RDINATES OF THE
STANDARD NORMAL CURVE
1 t2
Y = -= exp (- 2)
V27r
t 0
·-
1
__......
2 3 .. 4 5 6 7 8 9
0.0 .3~)89 .3989 .3989 .3988 .3986 .3984 .3982 .3980 .3977 .3973
0.1 .3970 .3965 .3961 .3956 .3951 .3945 .3939 .3932 .3925 .3918
0 ') .3910 .3902 .3894 .3885 .3876 .3867 .3857 .3847 .3836 .3825
0.:3 .3814 .3802 .3790 .3772 .3765 .3752 .3739 .3725 .3712 .3697
0.4 .3683 .3668 .3653 .3637 .3621 .3605 .3589 .3572 .3555 .3538
0.5 .3521 .3503 .3485 .3467 .3448 .3429 .3410 .3391 .3372 .3352
0.6 .3332 .3312 .3292 .3271 .3251 .3230 .3209 .3187 .3166 .3144
0.7 .3123 .3101 .3079 .3056 .30:34 .3011 .2980 .2966 .2943 .2920
I e.s
!
0.9
.2897
.2661.
.2874
.2637
.2850
.2613
.2827
.2589
.2803
.2565
.2780
.2541
.2756
.2516
.2732
.2492
.2709
.2468
.2685
.2444
1.0 .2·i20 .2:396 .2371 .234'7 .2323 .2299 .2275 .2251 .2227 .2203
1.1 .21.79 .2155 .2131 .2107 .2083 .2059 .2036 .2012 .1989 .1965
1.2 .1942 .1919 .1895 .1872 .1849 .1826 .1804 .1781 .1758 .1736
1 .u'J .1714 .1691 .1669 .1647 .1626 .1604 .1582 .1561 .1539 .1518
1.4 .1497 .1476 .1456 .1435 .1415 .1394 .1374 .1354 .1334 .1315
1.5 .1295 .1276 .1257 .1238 .1219 .1200 .1182 .1163 .1145 .1127
1.6 .1109 .1092 .1074 .1057 .1040 .1023 .1006 .0989
l 1.7 .0940 .0925 .0909 .0893 .0878 .0863 .0848 .08:33
.0973
.0818
.0957
.0804
l.R .0790 .0775 .0761 .0748 .0734 .0721 .0707 .0694 .0681 .0669
1.9 .0656 .0()44 .0632 .0620 .0608 .0596 .0584 .0573 .0562 .0551
2.0 .0540 .0529 .0519 .0508 .0498 .0488 .0478 .0468 .0459 .0449
2.1 .0440 .0431 .0422 .0413 .0404 .0396 .0387 .0379 .0371 .0363
2.2 .0355 .0347 .0339 .0332 .0325 .0317 .0310 .0303 .0297 .0290
2.3 .0283 .0277 .0270 .0264 .0258 .0252 .0246 .0241 .0235 .0229
2.4 .0224 .0219 .0213 .0208 .0203 .0198 .0194 .0189 .0184 .0180
~
?..5 .01.75 .0171. .0167 .0163 .0158 .01.54 .0151 .01.47 .0143 .0139
2.. t>" .0136 .0132 .0129 .0126 .0122 .0119 .0116 .0113 .0110 .0107
2.7 .0104 .0101 .0099 .0096 .0093 .0091 .0088, .0086 .0084 .0081
2.8 .0079 .0077 .00'75 .0073 .0071 .0069 .0067 .0065 .0063 .0061
2.9 .0060 .0058 .0056 .0055 .0053 .0051 .0050 .0048
. .0047 .0046
3.0 .0044 .0043 .0042 .0040 •0039 .0038 .0037 .0036 .0085 .0034
3.1 .0033 . 0032 .0031 . .0030 .0029 .0028 .0027 .0026 .0025 .0025
3.2 .002<1 .0023 .0022 .0022 .0021 .0020 .0020 .0019 .0018 .0018
u.,_,
'} ry
.OD17 .0017 .0016 .0016 .0015 .0015 .0014 .0014 .0013 .0013
3.t! .0012 .0012 .0012 .0011 .0011 .0010 .0010 .0010 .0009 .0009
~.5 .0009 .00.08 .0008 .0008 .0008 .0007 .0007 .0007 .0007 .0006
3.(3 .0006 .0006 .0006 .0005 .0005 .0005 .0005 .0005 .0005 .0004
3..7 .OfJO'i .0004 .0004 .0004 .0004 .0004 .0003 .0003 .0003 .0003
:l.8 .0003 .0003 .0003 .0008 .0003 .0002 .0002 .0002 .0002 .0002
:J.9 .0002 .0002 .0002 .0002 .0002 .0002 .0002 .0002 .0001 .0001
239
APPENDIX II - B
·:·~~··:' ;~(~~!_'~~2f~~·(
. . Y\W'aJ~L£;. ,.,,
~~2i~.
;..-· •· .. ~,.· :... ,,:;~·;·.w.::._:.:(:"··.-·..-..\~~~.:;.~ ...... ,.. ·~ . . . . . .:...... -
t 0 1 2 3 4 5 6 7 8 9
0.0 .5000 .5040 .5080 .5120 .5160 .5199 .5239 .5279 .5319 .5359
0.1 .5398 .5438 .5478 .5517 .5557 .5596 .5636 .5675 .5714 .5754
0.2 .5793 .5832 .5871 .5910 .5948 .5987 .6026 .6064 .6103 .6141
0.3 .6179 .6217 .6251? .6293 .6331 .6368 .6406 .6443 .6480 .6517
0.4 .6554 .6591 .6628 .6664 .6700 .6736 .6772 .6808 .6844 .6879
0.5 .6915 .6950 .6985 .7019 .7054 .7088 .7123 .7157 .7190 .7224
0.6 .7258 .7291 .7324 .7357 .7389 .7422 .7454 .7486 .7518 .7549
0.7 .7580 .7612 .7642 .7673 .7704 .7734 .7764 .7794 .7823 .7852
Q.8 .7881 .7910 .7939 .7967 .7996 .8023 .8051 .8078 .8106 .8133
0.9 .8159 .8186 .8212 .8238 .8264 .8289 .8315 .8340 .8365 .8389
1.0 .8413 .8438 .8461 .8485 .8508 .8531 .8554 .8577 .8599 .8621
1.1 .8643 .8665 .8686 .8708 .8729 .8749 .8770 .8790 .8810 .8830
1.2 .8849 .8869 .8888 .8907 .8925 .8944 .8962 .8980 .8997 .9015
1.3 .9032 .9049 .9066 .9082 .9099 .9115 .9131 .9147 .9162 .9177
1.4 .9192 .9207 .9222 .9236 .9251 .9265 .9279 .9292 .9306 .9319
1.5 .9332 .9345 .9357 .9370 .9382 .9394 .9406 .9418 .9429 .9441
1.6 .9452 .9463 .9474 .9484 .9495 .9505 .9515 .9525 .9535 .9545
1.'7 .9554 .9564 .9573 .9582 .9591 .9599 ,9608 .9616 .U625 .9633
1.8 .9641 .9649 .9656 .9664 .9671 .9678 .9686 .9693 .9699 .9706
1.9 .9713 .9719 .9726 .9732 .9738 .9744 .9750 .9756 .9761 .9767
2.0 .. 9772 .9778 .9783 .9788 .9793 .9798 .9803 .9808 .9812 .9817
2.1 .9821 .9826 .9830 .9834 .9838 .9842 .9846 .9850 .9854 .9857
2.2 .9861 .9864 .9868 .9871 .9875 .9878 .9881 .9884 .9887 .9890
2.3 .9893 .9896 .9898 .9901 .9904 .9906 .9909 .9911 .9913 .9916
2.4 .9918 .9920 .9922 .9925 .9927 .9929 .9931 .9932 .9934 .9936
2.5 .9938 .9940 .9941 .9943 .9945 .9946 .9948 .9949 .9951 .9952
2.6 .9953 .9955 .9956 .9957 .9959 .9960 .Q961 .9962 .9963 .9964
2.7 .9965 .9966 ,9967 .9968 .9969 .9970 .9971 .9972 .'9973 .9974
2.8 .9974 .9975 .9976 .9977 .9977 .9978 .9979 .9979 .9980 .9981
2.9 .9981 .9982 .9982 .9983 .9984 .9984
. .9985 .9985 .9986 .9986
3.0 . .9987 .9987 .9987 .9988 .9988 .9989 .9989 .9989 .9990 .9990
3.1 .9990 .9991 .9991 .9991 .9992 .9992 .9992 .9992 .9993 .9993
3.2 .9993 .9993 .9994 .9994 .9994 .9994 .9994 .9995 .9995 .9995
3.3 .9995 .9995 .9995 .9996 .9996 .9996 .9996 .9996 .9996 .9997
3.4 .9997 .9997 .9997 .9997 .9997 .9997 .9997 .9997 .9997 .9998
3.5 .9998 ·.9998 .9998 .9998 .9998 .9998 ~9998 .9998 .9998 .9998
3.6 .9998 .9998 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999
3.7 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999
3.8 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999 .9999
3.9 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
240
APPENDIX II - C
A
AREAS
under the
STANDARD
NORMAL CURVE
from 0 to t 0 . t
t 0 1 2 3 4 5 6 7 8 9
0.0 .0000 .0040 .0080 .0120 .0160 .0199 .0239 .0279 .0319 .0359
0.1 .0398 .0438 .0478 .0517 .0557 .0596 .0636 .0675 .0714 .0754
0.2 .0793 .0832 .0871 .0910 .0948 .0987 .1026 .1064 .1103 .1141
0.3 .1179 .1217 .1255 .1293 .1331 .1368 .1406 .1443 .1480 .1517
0.4 .1554 .1591 .1628 .1664 .1700 .1736 .1772 .1808 .1844 .1879
.0.5 .1915 .1950 .1985 .2019 .2054 .2088 .2123 .2157 .2190 .2224
0.6 .2258 .2291 .2324 .2357 .2389 .2422 .2454 .2486 .2518 .2549
0.7 .2580 .2612 .2642 .2673 .2704 .2734 .2764 .2794 .2823 .2852
0.8 .2881 .2910 .2939 .2967 .2996 .3023 .3051 .3078 .3106 .3133
0.9 .3159 .3186 .3212 .3238 .3264 .3289 .3315 .3340 .3365 .3389
1.0 .3413 .3438 .3461 .3485 .3508 .3531 .3554 .3577 .3599 ..3621
1.1 .3643 .3665 .3686 .3708 .3729 .3749 .3770 .3790 .3810 .3830
1.2 .3849 .3869 .3888 .3907 .3925 .3944 .3962 .3980 .3997 .4015
1.3 .4032 .4049 .4066 .4082 .4099 .4115 .4131 .4147 .4162 .4177
1.4 .4192 .4207 .4222 .4236 .4251 .4265 .4279 .4292 .4306 .4319
1.5 .1332 .4345 .4357 .4370 .4382 .4394 .4406 .4418 .4420 .4441
1.fl .4452 .4463 .4474 .4484 .4495 .4505 .4515 .4525 .4535 .4545
1.7 .4554 .4564 .4573 .4582 .4591 .4599 A608 .4616 .4625 .4633
.· 1.8 .4641 .4649 .4656 .4664 .4671 .4678 .4686 .4693 .4699 .4706
1.9 .4713 .4719 .4726 .4732 ,4738 .4744 .4750 .4756 .4761 .4767
2.0 .4772 .4778 .4783 .4788 .4793 .4798 .4803 .4808 . .4812 .4817
2.1 .4821 .4826 .4830 .4834 .4838 .4842 .4846 .4850 .4854 .4857
2.2 .4861 .4864 .4868 .4871 .4875 .4878 .4881 .4884 .4887 .4890
~.3 .4893 .4896 .4898 .4901 .4904 .490() .4909 ..4911 .4!>13 . .4916
2.•i .4918 .4920 .4922 .4925 .4927 .4929 .4931 .4932 .4934 .4936
2.5 A938 .4940 .4941 .4943 .4945 .4946. .4948 .4949 .4951 .4952
2.6 .4953 .4955 .4956 .4957 .4959 .4960 .4961 .4962 .4963 .4964
2.7 .4965 .4966 .4967 .4968 .4969 .4970 .4971 .4972 .4973 .4974
2.8 .4974 .4975 .4976 .4977 .4977 .4978 .4979 .4979 .4980 .4981
2.9 .4981 .4982 .4982 .4983 .4984 .4984 .4985 .4985 .4986 .4986
.
3.0 .4987 .4987 .4987 .4988 .4988 .4989 .4989 .4989 .4990 .4990
3.1 .4990 .4991 .4991 .4991 .4992 .4992 .4992 .4992 .4993 .4993
3.2 .4993 .4993 .4994 .4994 •.;,994 .4994 .4994 .4995 .4995 .4995
3.3 .4995 .4995 .4995 .4996 .4996 .4996 .4996 .4996 .4996 .4997
3.4 .4997 .4997 .4997 .4997 .4997 .4997 .4997 .4997 .4l:l97 ..1!)98
3.5 .4998 .4998 .4998 .4998 .4998 .4998 .4998 .4998 .4998 .4998
3.6 .4998 .4998 .499~) .4999 .4999 .4999 .4999 .4999 .4999 •.4999
3.7 .4999 .4999 .4999 .4999 .4999 .4999 .4999 .4999 .4999 .4999
3.8 .49()9 .4999 .4999 .4999 .4999 A999 .4999 .4999 .4999 .4999
·~.D .5000 .5000 .5000 .5000 .5000 .5000 .5000 .5000 .5000 .5000
241
BIBLIOGRAPHY
Wells, D.E. and Krakiwsky, E.J., 1971: The Method of Least Squares,
Department of Surveying Engineering, U.N.B., Lecture Notes No. 18.