Lecture
Here the idea is that around a recognizable object volume cell is built
. At the same time, an unknown object belongs to that image, the number of training representatives of which in the constructed cell turned out to be the majority. If we use statistical terminology, then the number of image objects
trapped in this cell characterizes the estimate of the volume averaged
probability density
.
To assess the averaged need to solve the question of the relationship between volume
cells and the number of objects of one or another class (image) that fell into this cell. It is reasonable to assume that the smaller
the more subtly be characterized
. But at the same time, the fewer objects will fall into the cell of interest, and therefore, the less reliable the estimate
. With an excessive increase
reliability of an assessment increases
, but the subtleties of its description are lost due to averaging over too large a volume, which can lead to negative consequences (an increase in the probability of recognition errors). With a small amount of training sample
It is advisable to take extremely large, but to ensure that within the cell density
little changed. Then their averaging over a large volume is not very dangerous. Thus, it may well happen that the cell volume relevant for one value
may not be suitable for other cases.
The following procedure is proposed (for now, we will not take into account the belonging of an object to a particular image).
In order to evaluate based on a training set containing
objects, center the cell around
and increase its volume as long as it does not contain
objects where
there is some function from
. These
objects will be closest neighbors
. Probability
vector hits
to the area
determined by the expression
.
This is a smoothed (averaged) density distribution. . If you take a sample of
objects (by a simple random selection from the general population),
of them will be inside the area
. Probability of hitting
of
objects in
described by a binomial law having a pronounced maximum around the mean
. Wherein
is a good estimate for
.
If we now assume that so small that
inside it changes slightly, then
,
Where - area volume
,
- point inside
.
Then . But
, Consequently,
.
So, the assessment density
is the value
. (*)
Without proof we give the statement that the conditions
and
(**)
are necessary and sufficient for convergence to
in probability at all points where the density
continuous.
This condition is satisfied, for example, .
Now we will take into account the belonging of objects to one or another image and try to estimate the posterior probabilities of the images.
Suppose we place a volume cell around
and grab the sample with the number of objects
,
of which belong to the image
. Then according to the formula
estimation of joint probability
there will be a magnitude
,
but
.
Thus, the posterior probability estimated as the fraction of the sample in the cell related to
. To minimize the error level, you need an object with coordinates
attributed to the class (image), the number of objects of the training sample of which is maximum in the cell. With
such a rule is Bayesian, that is, it provides a theoretical minimum of the probability of recognition errors (of course, the conditions
).
Comments
To leave a comment
Pattern recognition
Terms: Pattern recognition