dc.contributor.author |
Geng, Haokun |
en |
dc.contributor.author |
Chien, HJ |
en |
dc.contributor.author |
Klette, R |
en |
dc.contributor.editor |
Huang, F |
en |
dc.contributor.editor |
Sugimoto, A |
en |
dc.coverage.spatial |
Auckland, New Zealand |
en |
dc.date.accessioned |
2016-11-03T03:55:17Z |
en |
dc.date.issued |
2016 |
en |
dc.identifier.citation |
In F. Huang, & A. Sugimoto (Eds.), Image and Video Technology – PSIVT 2015 Workshops: RV 2015, GPID 2013, VG 2015, EO4AS 2015, MCBMIIA 2015, and VSWS 2015, Auckland, New Zealand, November 23-27, 2015. Revised Selected Papers, 9555, pp. 192-205 |
en |
dc.identifier.isbn |
9783319302843 |
en |
dc.identifier.issn |
0302-9743 |
en |
dc.identifier.uri |
http://hdl.handle.net/2292/30950 |
en |
dc.description.abstract |
This paper presents an approach for incrementally adding missing information into a point cloud generated for 3D roadside reconstruction. We use a series of video sequences recorded while driving repeatedly through the road to be reconstructed. The video sequences can also be recorded while driving in opposite directions. We call this a multi-run scenario. The only extra input data other than stereo images is the reading from a GPS sensor, which is used as guidance for merging point clouds from different sequences into one. The quality of the 3D roadside reconstruction is in direct relationship to the accuracy of the applied egomotion estimation method. A main part of our motion analysis method is defined by visual odometry following a traditional workflow in this area: first, establish correspondences of tracked features between two subsequent frames; second, use a stereo-matching algorithm to calculate the depth information of the tracked features; then compute the motion data between every two frames using a perspective-n-point solver. Additionally, we propose a technique that uses a Kalman-filter fusion to track the selected feature points, and to filter outliers. Furthermore, we use the GPS data to bound the overall propagation of the positioning errors. Experiments are given with trajectory estimation and 3D scene reconstruction. We evaluate our approach by estimating the recovery of (so far) missing information when analysing data recorded in a subsequent run. |
en |
dc.description.uri |
http://librarysearch.auckland.ac.nz/UOA2_A:Combined_Local:uoa_alma51208414500002091 |
en |
dc.publisher |
Springer International Publishing |
en |
dc.relation.ispartof |
7th Pacific Rim Symposium on Image and Video Technology - PSIVT 2015 Workshops |
en |
dc.relation.ispartofseries |
Image and Video Technology – PSIVT 2015 Workshops: RV 2015, GPID 2013, VG 2015, EO4AS 2015, MCBMIIA 2015, and VSWS 2015, Auckland, New Zealand, November 23-27, 2015. Revised Selected Papers |
en |
dc.rights |
Items in ResearchSpace are protected by copyright, with all rights reserved, unless otherwise indicated. Previously published items are made available in accordance with the copyright policy of the publisher. Details obtained from http://link.springer.com/bookseries/558 |
en |
dc.rights.uri |
https://researchspace.auckland.ac.nz/docs/uoa-docs/rights.htm |
en |
dc.title |
Multi-run: An approach for filling in missing information of 3D roadside reconstruction |
en |
dc.type |
Conference Item |
en |
dc.identifier.doi |
10.1007/978-3-319-30285-0_16 |
en |
pubs.begin-page |
192 |
en |
pubs.volume |
9555 |
en |
dc.rights.holder |
Copyright: Springer |
en |
pubs.author-url |
http://link.springer.com/chapter/10.1007/978-3-319-30285-0_16 |
en |
pubs.end-page |
205 |
en |
pubs.finish-date |
2015-11-27 |
en |
pubs.publication-status |
Published |
en |
pubs.start-date |
2015-11-23 |
en |
dc.rights.accessrights |
http://purl.org/eprint/accessRights/RestrictedAccess |
en |
pubs.subtype |
Conference Paper |
en |
pubs.elements-id |
531855 |
en |
dc.identifier.eissn |
1611-3349 |
en |
pubs.record-created-at-source-date |
2016-11-03 |
en |