Computer Vision How To Approach Vision Problems
Computer Vision How To Approach Vision Problems
Image preprocessing. In Binary vision do we need to use an opening and/or a closing, etc?
In gray-scale or colour imaging do we need to use a Smoothing technique? If there is any
non-linearity do we need to geometrically correct the image?
Binary vision. Is the domain such that a sufficient distinction could be created between the
foreground and background to allow segmentation by thresholding?
Colour vision. Do we need to do some sort of analysis of the colours present in the scene?
Could we solve the problem using some sort of summary representation such as a histogram
or K-means.
Region based vision. If the data cannot be segmented by binary vision but needs to be
broken into distinct objects or regions would a region based technique such as split and
merge allow reliable segmentation?
Edge based vision. If the boundary of the objects would be sufficient to allow recognition
(or whatever is needed in the question) then perhaps the edges should be detected and
analyzed in some fashion (e.g. Hough, features, etc.). Also what sort of edge detector
should we use (Roberts, first derivative, second derivative, ) and does this need to be
followed by Contour Following, and the extraction of straight line segments, etc?
In the vision course associated with these notes application problems come up in two
instances. In tutorials open vision problems are presented to the class to be tackled in groups,
the idea being that each group would develop a series of steps to solve the problem. It is
essential when doing so that the solution starts at the correct starting point (e.g. the specified
images) and finishes at the specified target (e.g. recognized characters).
139
Computer Vision
The second instance that these problems are faced is during the examination but in this case
typically questions are asked in two parts. The first part asks about some specific area of
vision, and the second part usually builds upon this to solve an application problem. It is
essential that student provide details of the techniques that they use, and to give some idea of
what is required a sample question and sample answer are provided here. Note that the level
of detail required varies depending on the complexity of the answer. If there are 5 steps to
solve a problem then less detail is needed to describe each these steps than if there were only
2 steps to solve the problem.
Sample Answer
a. First derivative edge detectors measure the rate of change of the
image intensity colour (for grey-scale images), whereas the second
derivative edge detectors look at the rate of change of the rate of
change of the image intensity. Considering a single row of an image
with two significant intensity changes:
140
Computer Vision
Computer Vision
Similarities:
-
Differences
-
142
Computer Vision
b. To address this problem we assume that we have an image of just one bottle.
Then we apply
- Sobel edge detection (as in part a).
- Non maxima suppression:
If the gradient image is thresholded (to identify significant edges) the
resulting edges are generally quite wide. Hence at this stage nonmaxima suppression is used:
Quantize edge orientations
for all points (i,j)
Look at the 2 points orthogonal to edge
if gradient(i,j) < gradient(either of these 2 points)
output(i,j) = 0
else output(i,j) = gradient(i,j)
-
Computer Vision
two on the middle line and compare edge points on each side relative to
those points (sort of like folding in the middle to see if the two sides
match).
-
To extract the chain of edge points we can start from any one of
the label edge points already located and then use a technique
similar to heuristic search for images borders to build up the
rest of the chain of edge points. Starting at the chosen label
point search forwards repeatedly for other edge points (there
should be just one) until reaching the start point again. This
would not work for all contours but would work for the label in
this case.
144