Agreement Criteria For check of conformity of empirical distribution to theoretical (hypotheses) it is possible to impose a theoretical curve on the histogram (Fig. 6).
Fig. 6. Histogram and theoretical density function
Thus random divergences connected with the limited volume of supervision, or divergences testifying about wrong selection of leveling function (hypothesis) will inevitably be found out. For the answer this question so-called «agreement criteria» are used. For this purpose random variable U describing a divergence of empirical and theoretical distributions in the assumption of the validity of theoretical distribution is entered. The measure of divergence U gets out so that function of its distribution did not depend on a kind of leveled (empirical) distribution and quickly enough converged on number of supervision n to limiting function . Then the actual degree of a divergence u is defined and the probability is estimated. Small value speaks that the received divergence u by virtue of cleanly casual reasons is improbable, and theoretical distribution will badly be coordinated with empirical. However, high probabilities cannot be considered as the exhaustive proof of the validity of the theoretical law of distribution and testify only to absence of the bases it to reject. Sometimes act differently: in advance count a measure of a divergence which can be exceeded with the specified small probability, and at the considered theoretical distribution is rejected. There is a set of agreement criteria among which the most common are Pearson’s criterion and Kolmogorov-Smirnoff criterion . In Pearson’s agreement criterion a measure of a divergence of theoretical and empirical distributions is the weighed sum of squares of deviations (27) where k – number of intervals of splitting of values of a random variable, – quantity of the observations which have got in the i -th interval, – theoretical probability of occurrence of value from the i -th interval, n – general number of observations. In practical problems it is recommended to have in each interval of splitting not less than 5-10 observations [3]. Let's designate through t number of the independent communications imposed on probability . Their general number is equal to quantity of characteristics of the theoretical distribution selected on empirical data, plus 1 (a condition of normalization ). Thus, the scheme of application of criterion to an estimation of a coordination of theoretical and empirical distributions is reduced to the following: 1) Measure of a divergence from the formula (27) is defined. 2) Number of degrees of freedom r = k – t is defined. 3) On r and by means of special table [3] the probability of that value having distribution with r by degrees of freedom, will surpass the given value is defined. If this probability is rather small, the hypothesis (a theoretical curve) is rejected as improbable. If this probability is rather great, it is possible to recognize a hypothesis not contradicting the received experimental data. Be how much small owe probability р to reject or reconsider a hypothesis, is not solved on the basis of mathematical reasons and calculations. In practice if it appears, that р <0.1, it is recommended to check up or repeat experiment. If appreciable divergences will appear again, it is necessary to search for another law of distribution, more suitable for the description of empirical data. If the probability p > 0.1 (is rather great), it cannot be considered as the proof of validity of a hypothesis yet, and speaks only that the hypothesis does not contradict experimental data. In Kolmogorov-Smirnoff criterion a measure of a divergence theoretical F ( x ) and empirical distributions is the maximal module of a difference (28) A.N.Kolmogorov has proved, that at irrespective of kind F ( x ) probability of an inequality (29) aspires to the limit (30)
For check of a hypothesis by Kolmogorov-Smirnoff criterion it is necessary to construct functions of distribution for theoretical
F
(
x
) and empirical
distributions to define a maximum
d
of module of a difference between them and to find
. After this it is necessary to find probability
from special table [2]:
|
Contents
>> Applied Mathematics
>> Mathematical Statistics
>> Treatment of Experiment Results
>> Agreement criteria