Improving disparity estimation based on residual cost volume and reconstruction error volume

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Autoren

  • J. Kang
  • L. Chen
  • F. Deng
  • C. Heipke

Externe Organisationen

  • Wuhan University
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Seiten (von - bis)135-142
Seitenumfang8
FachzeitschriftInternational Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives
Jahrgang43
AusgabenummerB2
PublikationsstatusVeröffentlicht - 12 Aug. 2020
Veranstaltung2020 24th ISPRS Congress - Technical Commission II - Nice, Virtual, Frankreich
Dauer: 31 Aug. 20202 Sept. 2020

Abstract

Recently, great progress has been made in formulating dense disparity estimation as a pixel-wise learning task to be solved by deep convolutional neural networks. However, most resulting pixel-wise disparity maps only show little detail for small structures. In this paper, we propose a two-stage architecture: we first learn initial disparities using an initial network, and then employ a disparity refinement network, guided by the initial results, which directly learns disparity corrections. Based on the initial disparities, we construct a residual cost volume between shared left and right feature maps in a potential disparity residual interval, which can capture more detailed context information. Then, the right feature map is warped with the initial disparity and a reconstruction error volume is constructed between the warped right feature map and the original left feature map, which provides a measure of correctness of the initial disparities. The main contribution of this paper is to combine the residual cost volume and the reconstruction error volume to guide training of the refinement network. We use a shallow encoder-decoder module in the refinement network and do learning from coarse to fine, which simplifies the learning problem. We evaluate our method on several challenging stereo datasets. Experimental results demonstrate that our refinement network can significantly improve the overall accuracy by reducing the estimation error by 30% compared with our initial network. Moreover, our network also achieves competitive performance compared with other CNN-based methods.

ASJC Scopus Sachgebiete

Zitieren

Improving disparity estimation based on residual cost volume and reconstruction error volume. / Kang, J.; Chen, L.; Deng, F. et al.
in: International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives, Jahrgang 43, Nr. B2, 12.08.2020, S. 135-142.

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Kang, J, Chen, L, Deng, F & Heipke, C 2020, 'Improving disparity estimation based on residual cost volume and reconstruction error volume', International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives, Jg. 43, Nr. B2, S. 135-142. https://doi.org/10.5194/isprs-archives-XLIII-B2-2020-135-2020
Kang, J., Chen, L., Deng, F., & Heipke, C. (2020). Improving disparity estimation based on residual cost volume and reconstruction error volume. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives, 43(B2), 135-142. https://doi.org/10.5194/isprs-archives-XLIII-B2-2020-135-2020
Kang J, Chen L, Deng F, Heipke C. Improving disparity estimation based on residual cost volume and reconstruction error volume. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives. 2020 Aug 12;43(B2):135-142. doi: 10.5194/isprs-archives-XLIII-B2-2020-135-2020
Kang, J. ; Chen, L. ; Deng, F. et al. / Improving disparity estimation based on residual cost volume and reconstruction error volume. in: International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives. 2020 ; Jahrgang 43, Nr. B2. S. 135-142.
Download
@article{6022add6538f403d88ddffea50541a6e,
title = "Improving disparity estimation based on residual cost volume and reconstruction error volume",
abstract = "Recently, great progress has been made in formulating dense disparity estimation as a pixel-wise learning task to be solved by deep convolutional neural networks. However, most resulting pixel-wise disparity maps only show little detail for small structures. In this paper, we propose a two-stage architecture: we first learn initial disparities using an initial network, and then employ a disparity refinement network, guided by the initial results, which directly learns disparity corrections. Based on the initial disparities, we construct a residual cost volume between shared left and right feature maps in a potential disparity residual interval, which can capture more detailed context information. Then, the right feature map is warped with the initial disparity and a reconstruction error volume is constructed between the warped right feature map and the original left feature map, which provides a measure of correctness of the initial disparities. The main contribution of this paper is to combine the residual cost volume and the reconstruction error volume to guide training of the refinement network. We use a shallow encoder-decoder module in the refinement network and do learning from coarse to fine, which simplifies the learning problem. We evaluate our method on several challenging stereo datasets. Experimental results demonstrate that our refinement network can significantly improve the overall accuracy by reducing the estimation error by 30% compared with our initial network. Moreover, our network also achieves competitive performance compared with other CNN-based methods.",
keywords = "Disparity Refinement, Reconstruction Error, Residual Cost Volume, Stereo Matching",
author = "J. Kang and L. Chen and F. Deng and C. Heipke",
note = "Funding Information: The author Junhua Kang would like to thank the China Scholarship Council (CSC) for financially supporting her study at the Institute of Photogrammetry and GeoInformation, Leibniz Universit{\"a}t Hannover, Germany, as a visiting PhD student. Furthermore, we gratefully acknowledge the support of NVIDIA Corporation for the donation of GPUs used for this research; 2020 24th ISPRS Congress - Technical Commission II ; Conference date: 31-08-2020 Through 02-09-2020",
year = "2020",
month = aug,
day = "12",
doi = "10.5194/isprs-archives-XLIII-B2-2020-135-2020",
language = "English",
volume = "43",
pages = "135--142",
number = "B2",

}

Download

TY - JOUR

T1 - Improving disparity estimation based on residual cost volume and reconstruction error volume

AU - Kang, J.

AU - Chen, L.

AU - Deng, F.

AU - Heipke, C.

N1 - Funding Information: The author Junhua Kang would like to thank the China Scholarship Council (CSC) for financially supporting her study at the Institute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, Germany, as a visiting PhD student. Furthermore, we gratefully acknowledge the support of NVIDIA Corporation for the donation of GPUs used for this research

PY - 2020/8/12

Y1 - 2020/8/12

N2 - Recently, great progress has been made in formulating dense disparity estimation as a pixel-wise learning task to be solved by deep convolutional neural networks. However, most resulting pixel-wise disparity maps only show little detail for small structures. In this paper, we propose a two-stage architecture: we first learn initial disparities using an initial network, and then employ a disparity refinement network, guided by the initial results, which directly learns disparity corrections. Based on the initial disparities, we construct a residual cost volume between shared left and right feature maps in a potential disparity residual interval, which can capture more detailed context information. Then, the right feature map is warped with the initial disparity and a reconstruction error volume is constructed between the warped right feature map and the original left feature map, which provides a measure of correctness of the initial disparities. The main contribution of this paper is to combine the residual cost volume and the reconstruction error volume to guide training of the refinement network. We use a shallow encoder-decoder module in the refinement network and do learning from coarse to fine, which simplifies the learning problem. We evaluate our method on several challenging stereo datasets. Experimental results demonstrate that our refinement network can significantly improve the overall accuracy by reducing the estimation error by 30% compared with our initial network. Moreover, our network also achieves competitive performance compared with other CNN-based methods.

AB - Recently, great progress has been made in formulating dense disparity estimation as a pixel-wise learning task to be solved by deep convolutional neural networks. However, most resulting pixel-wise disparity maps only show little detail for small structures. In this paper, we propose a two-stage architecture: we first learn initial disparities using an initial network, and then employ a disparity refinement network, guided by the initial results, which directly learns disparity corrections. Based on the initial disparities, we construct a residual cost volume between shared left and right feature maps in a potential disparity residual interval, which can capture more detailed context information. Then, the right feature map is warped with the initial disparity and a reconstruction error volume is constructed between the warped right feature map and the original left feature map, which provides a measure of correctness of the initial disparities. The main contribution of this paper is to combine the residual cost volume and the reconstruction error volume to guide training of the refinement network. We use a shallow encoder-decoder module in the refinement network and do learning from coarse to fine, which simplifies the learning problem. We evaluate our method on several challenging stereo datasets. Experimental results demonstrate that our refinement network can significantly improve the overall accuracy by reducing the estimation error by 30% compared with our initial network. Moreover, our network also achieves competitive performance compared with other CNN-based methods.

KW - Disparity Refinement

KW - Reconstruction Error

KW - Residual Cost Volume

KW - Stereo Matching

UR - http://www.scopus.com/inward/record.url?scp=85091075776&partnerID=8YFLogxK

U2 - 10.5194/isprs-archives-XLIII-B2-2020-135-2020

DO - 10.5194/isprs-archives-XLIII-B2-2020-135-2020

M3 - Conference article

AN - SCOPUS:85091075776

VL - 43

SP - 135

EP - 142

JO - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives

JF - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives

SN - 1682-1750

IS - B2

T2 - 2020 24th ISPRS Congress - Technical Commission II

Y2 - 31 August 2020 through 2 September 2020

ER -