Theory of the combination of observations which is subject to the least error (1821–1826) by Carl Friedrich Gauss, translated from French by Wikisource
→
Based on the 1855 French translation by Joseph Bertrand.
Carl Friedrich Gauss4458592Theory of the combination of observations which is subject to the least error1821-1826Wikisource
In the previous Memoir, we assumed that the quantities to be determined, with the help of imperfect observations, depended on certain unknown elements, in terms of which they could be expressed: the problem then consisted in deducing from the observations, as accurately as possible, the values of these elements.
In most cases, this is indeed how the question arises. But sometimes the problem is presented differently, and at first glance it is even doubtful whether or not it can be reduced to a problem of the required form. Indeed, frequently the quantities under observation are not given in the form of a function of certain elements, and they seem to only be reducible to such a form by difficult or even ambiguous operations; whereas the nature of the problem does provide certain conditions that the observed values must rigorously satisfy.
Upon closer inspection, however, it can be seen that this case does not differ essentially from the previous one, and that it can be reduced to it. Indeed, if denotes the number of observed quantities, and denotes the number of constraint equations, then nothing prevents us from choosing, among the former, quantities, and considering them as our only unknowns, the others being regarded as functions of these, defined by the constraint equations. By this artifice, we can return to the case of the previous Memoir.
Nevertheless, although this approach often leads to the result in a quite convenient manner, one cannot deny that it is less natural, and it is therefore desirable to treat the problem in another form, which admits a very elegant solution. Moreover, since this new solution leads to faster calculations than the previous one whenever is less than or equivalently, whenever the multitude of elements that we have denoted by in the previous commentary, is greater than it is preferable to use the new solution, even if it would be easy, given the nature of the problem, to unambiguously eliminate the constraint equations.
Let etc. denote the quantities whose values are provided by observation. Suppose an unknown depends on these and is expressed by a known function let etc. be the values of the differential quotients
corresponding to the true values of etc.. If one substituted the true values for etc. in the function one would also obtain the true value of but if the observations are affected by errors etc., there resulting total error for will be represented by
provided (when is not linear) that one can neglect the squares and products of errors etc., as we will always assume.
Although the magnitude of errors etc., is uncertain, the uncertainty attached to the value found for can generally be measured by the mean error to be feared in the adopted determination. According to the principles developed in the previous commentary, this average error is
where etc. are the average errors of the various observations. If all observations are affected by the same degree of uncertainty, this expression becomes
and it is clear that in this calculation, it is permissible to replace etc., with the values taken by the differential coefficients
when one replaces etc. with their observed values.
When the quantities etc. are independent, the unknown can only be determined in one way, and the uncertainty attached to the result cannot be avoided or reduced. In this case, the observations provide a value of the unknown which is not arbitrary.
It is quite different when the quantities etc. are subject to certain relations, which we will suppose are expressed by conditional equations
where etc. are given functions of the unknowns etc. In this case, there are infinitely many ways to represent our unknown as a combination of the quantities etc., since it is clear that the function can be replaced by any other function such that the difference vanishes identically upon setting
If the observations were rigorously exact, this substitution would not change the result at all; but, due to inevitable errors, each form adopted for will correspond to a different result, and the error committed, instead of being
will be
where etc. denote the differential quotients
Although it is impossible to assign the value of the various errors, we can, however, compare the mean errors to be feared in the various combinations. The most advantageous combination will be the one that gives the minimum value to the mean error. This error being
The infinite number of functions by which can be replaced, will differ from each other in our investigations only by the values they provide for etc.: therefore, we must first seek the relationships that exist between the systems of values that these coefficients can take. Let
denote the values taken by the coefficients
when we substitute the true values of etc. It is clear that if increments etc. are given to etc. which do not change etc. and thus leave each of them with a value of zero, these increments, which satisfy the equations
will not change the value of and consequently, we will have
We easily conclude that etc., must have the form
where etc. are determined multipliers. Conversely, it is clear that for all values of etc., we can form a function for which the values etc., will precisely be those provided by these equations, and this function can, according to the above, be substituted for The simplest form we can give it is
but the most general is
where denotes a function of etc. which vanishes identically when etc., are zero, and whose value, in the current case, will be maximum or minimum. However, this makes no difference for our purposes.
It is now easy to assign values to etc. such that the sum
achieves its minimum value. It is clear that to achieve this goal, the knowledge of the absolute mean errors etc. is not necessary. Rather, it suffices to know only their ratios. Indeed, instead of these quantities, let us introduce the weights of the observations, etc., i.e. numbers inversely proportional to the squares etc. The quantities etc. must then be determined in such a way that the polynomial
achieves its minimum value. Suppose that etc., are the determined values of etc. that correspond to this minimum.
We now introduce the following notation:
The minimality condition clearly requires that we have
(1)
After these equations have provided the values of etc., we set
(2)
and the most suitable function to determine our unknown, corresponding to the minimum mean error, will be one whose differential coefficients, for the considered values of the variables, are equal to etc. The weight of the determination thus obtained will be
(3)
meaning that will precisely be the value taken by the polynomial considered above, for the values of the variables etc. which satisfy the equations (1).
In the previous article, we showed how to determine the function that provides the most suitable determination of the unknown Let us now examine what value results from this. Let's denote it by : it will be obtained by substituting, in the observed values of the quantities etc. Let be the value that takes when the same substitutions are made, and let be the true value of the unknown, obtained by substituting the true values of etc., either in or in We have
and, consequently,
Replacing etc. with their values provided by (2), and setting
(4)
we will have
(5)
It is not possible to calculate etc. by means of the formulas (4), as the errors etc. appearing in them have unknown values, but it is easy to see that the quantities etc., are none other than the values of etc., corresponding to the observed values of etc. In this way, the systems of equations (1), (3), (5) exhibit the complete solution to our problem. In fact, it is clear that the remark made at the end of art. 2, regarding the quantities etc., can be applied to the calculation of etc. etc. etc., or in other words, the true values of etc. can be replaced with the observed values.
In place of formula (3), which represents the weight of the most probable determination, several other expressions can be given, which are worth developing.
First, observe that if the equations (2) are multiplied by etc. resp. and then added, the result is
The left hand side is and therefore, denoting the right hand side by we have
and similarly
Next, multiplying the equations (2) by etc. and adding, we find that
and thus we obtain a second expression for the weight,
If, finally, if we multiply the same equations (2) by etc. and add, we obtain a third expression for the weight,
where, with the same notation as before,
One can easily derive from this a fourth expression for the weight,
The general solution we have just outlined is particularly adapted to the case where there is only one unknown to determine. However, when seeking the most plausible values for several unknowns, all of which depend on the same observations, or when it is unknown which unknowns should be derived from the observations, it is necessary to proceed differently, as we will now explain.
Let us consider the quantities etc. as indeterminate, and set
(6)
Suppose that we deduce, by elimination, that
(7)
Let us first of all observe that the symmetrically placed coefficients are necessarily equal, i.e.
This follows from the theory of elimination, but we will also give a direct demonstration below.
Comparing equations (7) and (9) shows that the auxiliary quantities etc., are the values taken by the indeterminates etc., when we set etc., and thus we have
(12)
Multiplying equations (10) by etc. respectively, and adding, we obtain
and similarly,
(13)
Now, since is the value taken by after the observed values are substituted for etc., it is easy to see that if we apply the corrections etc., to each of these quantities, the value of becomes , and similarly etc., vanish under this assumption. In a similar way, equation (11) also shows that is the value taken by after the same substitutions have been made.
We will refer to the application of the corrections etc. as compensation of observations. This evidently leads us to the following very important consequence. Namely, the observations, compensated in the way we have describde, exactly satisfy all of the conditional equations, and cause any function of the observed quantities to take the value resulting from the most suitable combination of the unaltered observations. Since the conditional equations are too few to deduce the exact values of the errors, we will have found, by the above, the most plausible errors. We will henceforth refer to the quantities etc. by this name.
Since the number of observations is greater than the number of conditional equations, an infinite number of systems can be found, other than the system of the most plausible corrections, which exactly satisfy the condition equations.
It is important to examine the relationships that link these various systems. Let etc., be such a system of corrections, different from the most plausible system. Then we have
Multiplying these equations by etc., and adding, we obtain, with the help of the equations (10),
But combining equations (13) in the same manner gives
(14)
From the combination of these results, one easily deduces
Therefore, the sum
is necessarily greater than the sum
which can be stated as follows:
Theorem.The sum of squares of the corrections by which the observations can be reconciled with the condition equations, being multiplied by the weights of the corresponding observations respectively, gives a minimum, if the most plausible corrections are adopted.
This is the very principle of least squares, from which equations (12) and (10) can be easily and immediately derived. The minimum value of the sum, which we will now denote by is equal, by equation (14), to
The determination of the most plausible errors, being independent of etc., clearly provides the most convenient preparation, regardless of how the observations are to be subsequently used. Furthermore, it is easy to see that, to achieve this goal, it is not necessary to perform elimination indefinitely, i.e. to calculate etc., it suffices to deduce from the equations (12), by a definite elimination, the auxiliary quantities etc., which we will call, in what follows, the correlates of the condition equations
and to substitute these quantities into equation (10).
This method leaves nothing to be desired when only the most plausible values of the quantities provided by the observation are required, but it is different when one also seeks the weights of each of the found values. Whichever of the four previous formulas one wants to use, it is essential to know etc., or equivalently, etc.; for this reason, it will be useful to more closely study the elimination that provides these quantities, and to obtain a more convenient method for determining the weights.
We must now subject the function to a transformation similar to that which has been indicated in Theoria Motus Corporum Coelestium art. 182, and was further developed in Disquisitione de elementis ellipticis Palladis. To this end, let us set
Having solved our primary problem, we will now address some secondary questions, which will shed more light on the subject.
We will first investigate whether it is possible for the elimination that provides etc., as functions of etc., to become impossible in certain cases. This would obviously happen if the functions etc., were not independent of each other. Suppose, for a moment, that this is the case, and that one of them can be expressed as a function of the others, such that we have an identity
where etc. are determined numbers. Then we will have
and if we set
it follows automatically that
and consequently,
where etc. are, by their nature, all positive, this equation requires
Let us now consider the complete differentials etc. corresponding to the values of etc., immediately provided by the observations. By our previous results, these differentials
will be related to each other in such a way that when they are multiplied by etc. respectively and added, the sum is identically zero, so that among the equations
at least one can be regarded as superfluous, because it will be satisfied as soon as the others are.
Upon closer examination of the question, it is seen that this conclusion is only applicable to values of the variables that differ infinitesimally from those provided by observation. In particular, there are two cases to distinguish: first, the case where one of the equations
is generally and absolutely contained within the others, and can therefore be eliminated; and second, the case where, for the particular values of etc. provided by the observations, one of the functions etc., for example, achieves its maximum or minimum value, or, more generally, a value where its differential vanishes, while the other equations remain satisfied. However, since we only consider variations of our variables whose squares are negligible, this second case (which in practice will rarely occur) can be assimilated into the first, and one of the constraint equations can be removed as redundant. If the remaining equations are independent in the sense we have indicated, it is certain, according to the above, that the elimination is possible. However, we reserve the right to return to this matter, which deserves to be examined as a theoretical subtlety rather than as a question of practical utility.
In the previous commentary, arts. 37 sqq., we have shown how to approximate the weight of a determination a posteriori. If approximate values of quantities are provided by equally precise observations, and they are compared with the values resulting from the most plausible assumptions one can make about the elements on which they depend, it has been seen that one must add the squares of the obtained differences, divide the sum by and then the resulting quotient can be regarded as an approximate value of the square of the average error inherent in this kind of observation. If the observations are of unequal precision, the only modification that must be made is that one must multiply the squares of the differences by the respective weights of the corresponding observations, and the mean error obtained in this way relates to observations whose weight is taken to be unity.
In the current case, the sum of the squares we are talking about obviously coincides with the sum , and the difference with the number of the condition equations. Consequently, for the average error of observations whose weight is , we will have the expression so that the determination will be increasingly reliable for larger values of
However, it is worthwhile to establish this result independently of the reasoning in the first investigation. To this end, it will be helpful to introduce some new notation. Suppose that, corresponding to the values
The series of observations that provide the quantities etc., affected by random errors etc., can be considered as a test that does not reveal the magnitude of each error, but instead, by means of the rules explained above, allows us to determine the quantity as a function of the errors. In such a test, some errors may be larger and others smaller; but the greater the number of errors used, the greater the probability that differs little from its average value. The difficulty thus boils down to finding the average value of
By the principles outlined in the first Memoir, which it is unnecessary to reproduce here, we find for this average value
Let denote the average error corresponding to observations with weight 1, so that we have
Then the previous expression can be written as follows:
but we have found
and thus the right-hand side is unity, as can be easily recognized by comparing equations (6) and (7). Similarly, we find that
and so on.
Hence the average value of is and insofar as it is permissible to regard the random value of as equal to the average value, we conclude that
One can assess the confidence warranted by this determination by calculating either the mean error to be feared, or its square. The latter will be the square root of the average value of the expression
which can be developed by reasoning similar to that which was explained in the first commentary (arts. 39 sqq.). We will omit these for the sake of brevity, and simply give the result.
The mean error to be feared in determining the square is expressed by
where is the average value of the fourth powers of the errors with weight 1, and is the sum
This sum cannot generally be simplified; however it can be shown, by a method similar to that used in art. 40 of the previous commentary, that its value is between and With the hypothesis on which we originally established the method of least squares, the term containing this sum disappears, because and the precision that must be attributed to the determination is therefore the same as if one had operated on observations affected by exactly known errors, following the precepts of arts. 15, 16 of the previous commentary.
For the compensation of observations, we have seen above that there are two operations to perform: first, it is necessary to determine the correlates of the conditional equations, that is, etc., which satisfy equations (12); secondly, we must substitute these quantities into equation (10). The compensation thus obtained can be called perfect or complete, as opposed to an imperfect or incomplete compensation. We will use the latter term to describe those resulting from the same equations (10), in which values of are substituted which do not satisfy the equations (12), i.e. which satisfy only some or none. We will not consider such a system of corrections here, and we will not even describe them as a compensation. When equations (10) are satisfied, systems (12) and (13) become equivalent, and the difference of which we speak can then be stated as follows: Fully compensated observations satisfy the conditional equations
Partially compensated observations satisfy only some of these equations, and perhaps none; a compensation for which all equations are satisfied is necessarily complete.
It follows from the very definition of a compensation that a combination of two compensations is again a compensation, and it does not matter whether the rules given to obtain a perfect compensation are applied to the raw observations or to imperfectly compensated observations.
Let etc., be a system of incomplete compensations resulting from the formulas
(I)
Since the observations thus changed do not satisfy all of the conditional equations, let etc., be the values taken by etc., when the values obtained for etc., are substituted into them. We must then find the values etc., which satisfy the equations
(II)
Once this is done, the complete compensation of the observations thus modified will be accomplished by the new changes etc.; etc., being deduced from the formulas
(III)
Let us now examine how these corrections agree with the complete compensation of the raw observations. First of all, it is clear that we have
In these equations, if we replace etc. with their values from (I), and if we replace etc., with their values from (II), then we obtain
It follows that the correlates of the conditional equations (12) are
and then equations (10), (I), and (III) show that
Hence the perfect compensation produces the same value for each unknown, whether calculated directly or obtained through an incomplete compensation.
When there are a large number of conditional equations, determining the correlates etc. may require calculations so lengthy that the calculator may be deterred: in such cases, it may be advantageous to obtain a complete compensation using a series of approximations based on the theorem of the preceding article. For this purpose, the conditional equations will be divided into two or more groups, and first, a compensation will be sought that satisfies the equations of the first group. Then, the values modified by this initial calculation will be treated, and they will be corrected again, considering only the equations of the second group. This second calculation will generally yield results that no longer satisfy the equations of the first group, and if only two groups were formed, one must then return to the first group and satisfy its equations with new corrections. The observations will then undergo a fourth compensation, in which only the conditions of the second group are considered, and by alternating in this way between the two groups of equations, corrections will be formed that will necessarily become smaller and smaller. If the choice of groups has been made skillfully, one will arrive at stable values after a few iterations.
When forming more than two groups, one must proceed in the same manner, with the various groups being used successively until the last, after which one returns to the first group to repeat the process in the same order. We have only indicated this method, the success of which will depend greatly on the skill of the calculator.
We have yet to provide the proof of the lemma which was assumed in art. 8. For the sake of perspicuity, let us use notation which is better adapted to this matter.
Let etc., be indeterminates, and suppose that the equations
have, through elimination, led to the following:
By substituting, in the first two equations of the second system, the values of etc., provided by the first, we obtain two identical equations:
These equations being identical, we can substitute any quantities we wish for etc. Let's choose in the first
and in the second
Subtracting the two identities element-wise yields
This can be written more succinctly as:
where denote two arbitrarily chosen indices; from this, we conclude that the equalities
and more generally,
and so
Moreover, since the order of the indeterminates is arbitrary, it is evident that, under the assumed hypothesis, we have
in general.
Since the method outlined in this commentary is intended primarily for useful application in high geodesy, the reader may appreciate it if we include some examples drawn from this branch of science.
The conditional equations that exist among the angles of a system of triangles can generally be divided into three categories.
I. The sum of the horizontal angles formed around the same vertex, encompassing the entire horizon, must be equal to four right angles.
II. The sum of the angles of each triangle can always be considered as known; for even when the triangle is situated on a curved surface, the excess of the sum of its angles over two right angles can be calculated with such approximation that it is permissible to consider the result as absolutely exact.
III. Finally, a third kind of relation is obtained by examining the ratios of the sides in triangles that form a closed network. If, indeed, the triangles are placed in such a way that the second triangle has a side in common with the first, and a side in common with the third; if the fourth triangle has two sides and respectively in common with the third and the fifth, and so on, until the last triangle, which has a common side with the previous one, and with the first of all a common side then the quotients
can be calculated by means of the angles opposite to them in the triangle whose two compared sides are part of it, and since the product of these fractions is obviously unity, there will be a relation between the sines of the various measured angles (reduced by one-third of the spherical or spheroidal excess when operating on a curved surface).
In somewhat complicated networks, it often happens that there are too many equations of the second and third category, and some of them are contained in the rest. On the other hand, it may happen in rare cases that some new equations must be added to those of the second category. For example, this will occur when the network contains polygons not divided into triangles by measurements; one must then introduce equations related to figures with more than three sides. On another occasion, we will return to this and provide more details on these various circumstances, but examining them now would divert us from our purpose. However, we cannot refrain from making an essential remark here for those who would like to make the rigorous application of our theory: we always assume that the quantities denoted by etc., have been observed directly, or deduced from observations, in such a way that their determinations are independent of each other, or at least can be regarded as such. In common practice, one observes the angles that can be considered as the elements etc. themselves. But one must not forget that if the system contains, in addition, triangles whose angles have not been directly observed and have been deduced from those that are known, by additions or subtractions, then these angles should not be counted among the quantities determined by the observation, and they must be included in the calculation as functions of the elements used to form them. It will be different if one adopts the method of observations of Struve (Astronomische Nachrichten II, p. 431), which consists of determining all directions around the same vertex, by relating them all to a single arbitrary direction. The angles measured in this way will then be taken for etc., and the angles of the triangles will all appear as differences. The equations of the first category should, in this case, be suppressed as superfluous, for they will be identically satisfied. The method that I myself followed in the triangulations carried out during the last few years differs from the two previous methods; however, it can be reconciled, as to the result, with Struve's method, in the sense that, at each station, one must regard etc., as the angles formed by the directions emanating from it, with a single arbitrarily chosen line.
We will give two examples: the first relates to the first mode of operation, and the second is related to observations made according to the second method.
The first example comes from the work of de Krayenhof, Précis historique des opérations trigonométriques faites en Hollande. We will seek to compensate the part of the observations related to the system of triangles that are contained between Harlingen, Sneek, Oldeholtpade, Ballum, Leeuwarden, Dockum, Drachten, Oosterwolde, and Gröningen. Between these points, nine triangles numbered 121, 122, 123, 124, 125, 127, 128, 131, 132 were formed in the cited work. The observed angles are as follows:
Triangle 121.
0. Harlingen
50° 58′ 15″ 238
1. Leeuwarden
82° 47′ 15″ 351
2. Ballum
46° 14′ 27″ 202
Triangle 122.
3. Harlingen
51° 05′ 39″ 717
4. Sneek
70° 48′ 33″ 445
5. Leeuwarden
58° 05′ 48″ 707
Triangle 123.
6. Sneek
49° 30′ 40″ 051
7. Drachten
42° 52′ 59″ 382
8. Leeuwarden
87° 36′ 21″ 057
Triangle 124.
9. Sneek
45° 36′ 07″ 492
10. Oldeholtpade
67° 52′ 00″ 048
11. Drachten
66° 31′ 56″ 513
Triangle 125.
12. Drachten
53° 55′ 24″ 745
13. Oldeholtpade
47° 48′ 52″ 580
14. Oosterwolde
78° 15′ 42″ 347
Triangle 127.
15. Leeuwarden
59° 24′ 00″ 645
16. Dockum
76° 34′ 09″ 021
17. Ballum
44° 01′ 51″ 040
Triangle 128.
18. Leeuwarden
72° 06′ 32″ 043
19. Drachten
46° 53′ 27″ 163
20. Dockum
61° 00′ 04″ 494
Triangle 131.
21. Dockum
57° 01′ 55″ 292
22. Drachten
83° 33′ 14″ 515
23. Gröningen
39° 24′ 52″ 397
Triangle 132.
24. Oosterwolde
81° 54′ 17″ 447
25. Gröningen
31° 52′ 49″ 094
26. Drachten
66° 12′ 57″ 246
The consideration of these triangles shows that the twenty-seven angles directly provided by observation have thirteen necessary relationships among them, namely: two of the first kind, nine of the second kind, and two of the third kind. However, it is not necessary to write down all these equations here in their final form, because for the calculation we only need the quantities which were denoted in the general theory by , , , etc., , , , etc. etc. Therefore, we can immediately write down the equations (13), which bring out these quantities. Instead of , , etc., we will simply write here etc.
In this way, the two equations of the first kind correspond to the following:
Next, we find, for the spheroidal excesses of the nine triangles: Then, we have the conditional equation of the second kind:
and so on for the others, and we have the following nine equations:
The conditional equations of the third kind are more easily expressed using logarithms. The first one is:
It seems unnecessary to develop the other one in finite form. To these two equations correspond the following, in which the coefficients refer to the seventh decimal place of Brigg's logarithms:
No reason leads us to attribute unequal weights to the various observations, so we will assume
Denoting the correlates of the condition equations, in the same order as these equations were written, by
we determine them by the following equations:
We then deduce by elimination:
Finally, the most plausible errors are given by the formulas
and we obtain the following numerical values, to which we add, for comparison, the corrections adopted by de Krayenhof:
de Kr.
de Kr.
|}
The sum of the squares of our corrections is 97.8845; the average error, as indicated by the 27 observed angles, is therefore
The sum of the squares of the corrections by de Krayenhof is 341.4201.
24.
The triangles whose vertices, in the triangulation of Hanover, are placed at Falkenberg, Breithorn, Hauselberg, Wulfsode, and Wilsede, will provide us with a second example.
These observations allow us to form seven triangles.
Triangle I.
Falkenberg
008° 00′ 47″ 395
Breithorn
028° 17′ 42″ 299
Hauselberg
143° 41′ 29″ 140
Triangle II.
Falkenberg
086° 27′ 13″ 323
Breithorn
055° 44′ 54″ 345
Wilsede
037° 47′ 53″ 635
Triangle III.
Falkenberg
041° 04′ 16″ 563
Hauselberg
102° 33′ 49″ 504
Wulfsode
036° 21′ 56″ 963
Triangle IV.
Falkenberg
078° 26′ 25″ 928
Hauselberg
068° 08′ 02″ 752
Wilsede
035° 25′ 34″ 281
Triangle V.
Falkenberg
037° 22′ 09″ 365
Wulfsode
073° 16′ 39″ 603
Wilsede
069° 21′ 11″ 508
Triangle VI.
Breithorn
027° 27′ 12″ 046
Hauselberg
148° 10′ 28″ 108
Wilsede
004° 22′ 19″ 354
Triangle VII.
Hauselberg
034° 25′ 46″ 752
Wulfsode
109° 38′ 36″ 566
Wilsede
035° 55′ 37″ 227
We have here seven conditional equations of the second kind (there is obviously no need to form any of the first); to form them, we must first find the spheroidal excesses of the seven triangles, and for this it is essential to know the length of one side. The one that connects Wilsede to Wulfsode is 22877.94 meters. We conclude that the spheroidal excesses of the various triangles are:
I...II...III...IV...V...VI...VII...
If we denote the angles that determine the directions indicated above by etc., then the angles of the first triangle, marked with the same indices, will be
and the first condition equation is therefore
The remaining six triangles will yield six similar equations, but a little attention shows that these equations are not independent. Indeed, the second is identical to the sum of the first, fourth, and sixth; the sum of the third and fifth is identical to that of the fourth and seventh: for this reason we will ignore the second and fifth. Instead of writing the remaining equations in finite form, we will write here the corresponding equations (13), with symbols etc. taking the place of etc.:
We can obtain eight conditional equations of the third kind from the triangles of the system, to which end one may combine three of the four triangles I, II, IV, VI, or three of the triangles III, IV, V, VII; however, a little attention shows that it suffices to consider two belonging to each of these two systems of triangles, and those will encompass all the others. Thus, for the sixth conditional equation, we have
and for the seventh,
to which correspond the equations (13),
If we attribute the same certainty to the various directions, setting
,
and if we denote the correlates of the seven conditional equations by then their determination will depend on the following equations:
From this, by elimination, we deduce
and the most probable errors are given by the formulas
from which we derive the following numerical values:
The sum of the squares of these errors is equal to hence, the average error resulting from the 18 observed directions is
To give an example of the last part of our theory, let us determine the precision with which the compensated observations determine the Falkenberg-Breithorn side, using the Wilsede-Wulfsode side. The function by which it is expressed, in this case, is:
Its value, deduced from the corrected observations, is
Differentiation of this equation provides the following, with etc. expressed in seconds,
From this we find that:
Taking the meter as the unit of length, the methods indicated above yield,
or
We conclude that the mean error to be feared in the value of the Falkenberg-Breithorn side is (where denotes the average error to be feared in the observed directions, this error being expressed in seconds), and consequently, if we adopt the value of announced above, this mean error to be feared is
Moreover, inspecting the system of triangles immediately shows that one cans completely disregard the Hauselberg station without breaking the network that connects the other four. However, it would not be good practice to eliminate the operations related to this point,[3] as they certainly contribute to increasing the precision of the whole. To show more clearly what increase in precision results from this, we will finish by again performing the calculation, after excluding all results related to the Hauselberg point. Of the eighteen directions mentioned above, eight then cease to be used, and the most plausible errors on those that remain are:
The value of the Falkenberg-Breithorn side then becomes a result little different from that obtained earlier. But calculating the weight yields
or
and the mean error to be feared, in meters, is
We see that by adding the operations related to Hauselberg, the weight of the determination of the Falkenberg-Breithorn side is increased in the ratio of to that is, in the ratio of unity to
↑In the previous calculations, it was sufficient to use three letters from each series to clarify the formulas' law; it seemed necessary here to include a fourth one to make the algorithm more evident. We have followed the series of letters in the natural way, with However, lacking the required letters, we have followed the series with and the series with .
↑The starting points to which the individual directions are referred are here regarded as arbitrary, although in reality they coincide with the meridian lines of the stations. The observations will be published in full at a later date; meanwhile, the figure can be found in "Astronomische Nachrichten" Vol. I. p. 441
↑The greater part of these observations had already been made before Breithorn's point was discovered and incorporated into the system.