Meaningful Matches in Stereovision

Computer Science – Computer Vision and Pattern Recognition

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

IEEE Transactions on Pattern Analysis and Machine Intelligence 99, Preprints (2011) 1-12

Scientific paper

10.1109/TPAMI.2011.207

This paper introduces a statistical method to decide whether two blocks in a pair of of images match reliably. The method ensures that the selected block matches are unlikely to have occurred "just by chance." The new approach is based on the definition of a simple but faithful statistical "background model" for image blocks learned from the image itself. A theorem guarantees that under this model not more than a fixed number of wrong matches occurs (on average) for the whole image. This fixed number (the number of false alarms) is the only method parameter. Furthermore, the number of false alarms associated with each match measures its reliability. This "a contrario" block-matching method, however, cannot rule out false matches due to the presence of periodic objects in the images. But it is successfully complemented by a parameterless "self-similarity threshold." Experimental evidence shows that the proposed method also detects occlusions and incoherent motions due to vehicles and pedestrians in non simultaneous stereo.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Meaningful Matches in Stereovision does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Meaningful Matches in Stereovision, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Meaningful Matches in Stereovision will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-380411

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.