One of the most interesting projects
I've worked on in the past couple of years was a project about href="https://en.wikipedia.org/wiki/Image_processing" rel="noreferrer">image
processing. The goal was to develop a system to be able to recognize Coca-Cola
'cans' (note that I'm stressing the word 'cans', you'll see
why in a minute). You can see a sample below, with the can recognized in the
green rectangle with scale and
rotation.
src="https://i.stack.imgur.com/irQtR.png" alt="Template
matching">
Some constraints on the
project:
- The
background could be very noisy. - The
can could have any scale or
rotation or even orientation (within reasonable
limits). - The image could have some degree of fuzziness
(contours might not be entirely straight). - There could be
Coca-Cola bottles in the image, and the algorithm should only detect the
can! - The brightness of the image
could vary a lot (so you can't rely "too much" on color
detection). - The can could be partly
hidden on the sides or the middle and possibly partly hidden behind a
bottle. - There could be no can at all
in the image, in which case you had to find nothing and write a message saying
so.
So you
could end up with tricky things like this (which in this case had my algorithm totally
fail):
src="https://i.stack.imgur.com/Byw82.png" alt="Total
fail">
I did this project a while ago, and
had a lot of fun doing it, and I had a decent implementation. Here are some details
about my
implementation:
Language:
Done in C++ using OpenCV
library.
Pre-processing:
For the image pre-processing, i.e. transforming the image into a more raw form to give
to the algorithm, I used 2
methods:
- Changing
color domain from RGB to rel="noreferrer">HSV and filtering based on "red" hue, saturation above a
certain threshold to avoid orange-like colors, and filtering of low value to avoid dark
tones. The end result was a binary black and white image, where all white pixels would
represent the pixels that match this threshold. Obviously there is still a lot of crap
in the image, but this reduces the number of dimensions you have to work
with.
- Noise filtering using median filtering
(taking the median pixel value of all neighbors and replace the pixel by this value) to
reduce noise. - Using href="http://en.wikipedia.org/wiki/Canny_edge_detector" rel="noreferrer">Canny Edge
Detection Filter to get the contours of all items after 2 precedent
steps.
Algorithm:
The algorithm itself I chose for this task was taken from href="https://rads.stackoverflow.com/amzn/click/com/0123725380"
rel="noreferrer">this awesome book on feature extraction and called href="http://en.wikipedia.org/wiki/Generalised_Hough_transform"
rel="noreferrer">Generalized Hough Transform (pretty different from the
regular Hough Transform). It basically says a few
things:
- You can
describe an object in space without knowing its analytical equation (which is the case
here). - It is resistant to image deformations such as
scaling and rotation, as it will basically test your image for every combination of
scale factor and rotation factor. - It uses a base model (a
template) that the algorithm will "learn". - Each pixel
remaining in the contour image will vote for another pixel which will supposedly be the
center (in terms of gravity) of your object, based on what it learned from the
model.
In the end, you
end up with a heat map of the votes, for example here all the pixels of the contour of
the can will vote for its gravitational center, so you'll have a lot of votes in the
same pixel corresponding to the center, and will see a peak in the heat map as
below:
src="https://i.stack.imgur.com/wxrT1.png"
alt="GHT">
Once you have that, a
simple threshold-based heuristic can give you the location of the center pixel, from
which you can derive the scale and rotation and then plot your little rectangle around
it (final scale and rotation factor will obviously be relative to your original
template). In theory at
least...
Results:
Now, while this approach worked in the basic cases, it was severely lacking in some
areas:
- It is
extremely slow! I'm not stressing this enough. Almost a
full day was needed to process the 30 test images, obviously because I had a very high
scaling factor for rotation and translation, since some of the cans were very
small. - It was completely lost when bottles were in the
image, and for some reason almost always found the bottle instead of the can (perhaps
because bottles were bigger, thus had more pixels, thus more
votes) - Fuzzy images were also no good, since the votes
ended up in pixel at random locations around the center, thus ending with a very noisy
heat map. - In-variance in translation and rotation was
achieved, but not in orientation, meaning that a can that was not directly facing the
camera objective wasn't
recognized.
Can
you help me improve my specific algorithm, using
exclusively OpenCV features, to resolve the
four specific issues
mentioned?
I hope some people will also learn
something out of it as well, after all I think not only people who ask questions should
learn. :)
No comments:
Post a Comment