DCSH-Matching patches in RGBD images

Yaron Eshet, Simon Korman, Eyal Ofek, Shai Avidan

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We extend patch based methods to work on patches in 3D space. We start with Coherency Sensitive Hashing (CSH), which is an algorithm for matching patches between two RGB images, and extend it to work with RGBD images. This is done by warping all 3D patches to a common virtual plane in which CSH is performed. To avoid noise due to warping of patches of various normals and depths, we estimate a group of dominant planes and compute CSH on each plane separately, before merging the matching patches. The result is DCSH-an algorithm that matches world (3D) patches in order to guide the search for image plane matches. An independent contribution is an extension of CSH, which we term Social-CSH. It allows a major speedup of the k nearest neighbor (kNN) version of CSH-its runtime growing linearly, rather than quadratic ally, in k. Social-CSH is used as a subcomponent of DCSH when many NNs are required, as in the case of image denoising. We show the benefits of using depth information to image reconstruction and image denoising, demonstrated on several RGBD images.

Original languageEnglish
Title of host publicationProceedings - 2013 IEEE International Conference on Computer Vision, ICCV 2013
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages89-96
Number of pages8
ISBN (Print)9781479928392
DOIs
StatePublished - 2013
Event2013 14th IEEE International Conference on Computer Vision, ICCV 2013 - Sydney, NSW, Australia
Duration: 1 Dec 20138 Dec 2013

Publication series

NameProceedings of the IEEE International Conference on Computer Vision

Conference

Conference2013 14th IEEE International Conference on Computer Vision, ICCV 2013
Country/TerritoryAustralia
CitySydney, NSW
Period1/12/138/12/13

Fingerprint

Dive into the research topics of 'DCSH-Matching patches in RGBD images'. Together they form a unique fingerprint.

Cite this