Multi-run: An approach for filling in missing information of 3D roadside reconstruction

Show simple item record

dc.contributor.author Geng, Haokun en
dc.contributor.author Chien, HJ en
dc.contributor.author Klette, R en
dc.contributor.editor Huang, F en
dc.contributor.editor Sugimoto, A en
dc.coverage.spatial Auckland, New Zealand en
dc.date.accessioned 2016-11-03T03:55:17Z en
dc.date.issued 2016 en
dc.identifier.citation In F. Huang, & A. Sugimoto (Eds.), Image and Video Technology – PSIVT 2015 Workshops: RV 2015, GPID 2013, VG 2015, EO4AS 2015, MCBMIIA 2015, and VSWS 2015, Auckland, New Zealand, November 23-27, 2015. Revised Selected Papers, 9555, pp. 192-205 en
dc.identifier.isbn 9783319302843 en
dc.identifier.issn 0302-9743 en
dc.identifier.uri http://hdl.handle.net/2292/30950 en
dc.description.abstract This paper presents an approach for incrementally adding missing information into a point cloud generated for 3D roadside reconstruction. We use a series of video sequences recorded while driving repeatedly through the road to be reconstructed. The video sequences can also be recorded while driving in opposite directions. We call this a multi-run scenario. The only extra input data other than stereo images is the reading from a GPS sensor, which is used as guidance for merging point clouds from different sequences into one. The quality of the 3D roadside reconstruction is in direct relationship to the accuracy of the applied egomotion estimation method. A main part of our motion analysis method is defined by visual odometry following a traditional workflow in this area: first, establish correspondences of tracked features between two subsequent frames; second, use a stereo-matching algorithm to calculate the depth information of the tracked features; then compute the motion data between every two frames using a perspective-n-point solver. Additionally, we propose a technique that uses a Kalman-filter fusion to track the selected feature points, and to filter outliers. Furthermore, we use the GPS data to bound the overall propagation of the positioning errors. Experiments are given with trajectory estimation and 3D scene reconstruction. We evaluate our approach by estimating the recovery of (so far) missing information when analysing data recorded in a subsequent run. en
dc.description.uri http://librarysearch.auckland.ac.nz/UOA2_A:Combined_Local:uoa_alma51208414500002091 en
dc.publisher Springer International Publishing en
dc.relation.ispartof 7th Pacific Rim Symposium on Image and Video Technology - PSIVT 2015 Workshops en
dc.relation.ispartofseries Image and Video Technology – PSIVT 2015 Workshops: RV 2015, GPID 2013, VG 2015, EO4AS 2015, MCBMIIA 2015, and VSWS 2015, Auckland, New Zealand, November 23-27, 2015. Revised Selected Papers en
dc.rights Items in ResearchSpace are protected by copyright, with all rights reserved, unless otherwise indicated. Previously published items are made available in accordance with the copyright policy of the publisher. Details obtained from http://link.springer.com/bookseries/558 en
dc.rights.uri https://researchspace.auckland.ac.nz/docs/uoa-docs/rights.htm en
dc.title Multi-run: An approach for filling in missing information of 3D roadside reconstruction en
dc.type Conference Item en
dc.identifier.doi 10.1007/978-3-319-30285-0_16 en
pubs.begin-page 192 en
pubs.volume 9555 en
dc.rights.holder Copyright: Springer en
pubs.author-url http://link.springer.com/chapter/10.1007/978-3-319-30285-0_16 en
pubs.end-page 205 en
pubs.finish-date 2015-11-27 en
pubs.publication-status Published en
pubs.start-date 2015-11-23 en
dc.rights.accessrights http://purl.org/eprint/accessRights/RestrictedAccess en
pubs.subtype Conference Paper en
pubs.elements-id 531855 en
dc.identifier.eissn 1611-3349 en
pubs.record-created-at-source-date 2016-11-03 en


Files in this item

Find Full text

This item appears in the following Collection(s)

Show simple item record

Share

Search ResearchSpace


Browse

Statistics