Filters








3,985 Hits in 10.9 sec

Detecting Changes in 3D Structure of a Scene from Multi-view Images Captured by a Vehicle-Mounted Camera

Ken Sakurada, Takayuki Okatani, Koichiro Deguchi
2013 2013 IEEE Conference on Computer Vision and Pattern Recognition  
For the images, we consider those captured by a camera mounted on a vehicle running in a city street.  ...  This paper proposes a method for detecting temporal changes of the three-dimensional structure of an outdoor scene from its multi-view images captured at two separate times.  ...  Acknowledgement This work was supported by Grant-in-Aid for Scientific Research on Innovative Areas "Shitsukan" (No. 23135501) and JSPS KAKENHI Grant Number 2230057.  ... 
doi:10.1109/cvpr.2013.25 dblp:conf/cvpr/SakuradaOD13 fatcat:mso2wbkpgfdwpnhw2yoaqgeu2a

Dense Optical Flow based Change Detection Network Robust to Difference of Camera Viewpoints [article]

Ken Sakurada, Weimin Wang, Nobuo Kawaguchi, Ryosuke Nakamura
2017 arXiv   pre-print
This paper presents a novel method for detecting scene changes from a pair of images with a difference of camera viewpoints using a dense optical flow based change detection network.  ...  On the other hand, in case of cameras mounted on a moving object, such as ground and aerial vehicles, we must consider the spatial correspondence between the images captured at different times.  ...  Additionally, to detect structural scene changes of an entire city from movies captured by vehicle-mounted cameras based on multi-view geometry, a large amount of image data and computational resources  ... 
arXiv:1712.02941v1 fatcat:wiibbn2wgbgenaeqz3bj6wedj4

Project AutoVision: Localization and 3D Scene Perception for an Autonomous Vehicle with a Multi-Camera System

Lionel Heng, Benjamin Choi, Zhaopeng Cui, Marcel Geppert, Sixing Hu, Benson Kuan, Peidong Liu, Rang Nguyen, Ye Chuan Yeo, Andreas Geiger, Gim Hee Lee, Marc Pollefeys (+1 others)
2019 2019 International Conference on Robotics and Automation (ICRA)  
Project AutoVision aims to develop localization and 3D scene perception capabilities for a self-driving vehicle.  ...  Emphasis is placed on leveraging multi-view geometry and deep learning to enable the vehicle to localize and perceive in 3D space.  ...  Given a set of images captured simultaneously from the multi-camera system, we detect the fiducial target in each image, and estimate its pose with respect to the multi-camera system by minimizing the  ... 
doi:10.1109/icra.2019.8793949 dblp:conf/icra/HengCCGHKLNYGLP19 fatcat:d7wcoms3drawhkd7klpu2sivlq

Project AutoVision: Localization and 3D Scene Perception for an Autonomous Vehicle with a Multi-Camera System [article]

Lionel Heng, Benjamin Choi, Zhaopeng Cui, Marcel Geppert, Sixing Hu, Benson Kuan, Peidong Liu, Rang Nguyen, Ye Chuan Yeo, Andreas Geiger, Gim Hee Lee, Marc Pollefeys, Torsten Sattler
2019 arXiv   pre-print
Project AutoVision aims to develop localization and 3D scene perception capabilities for a self-driving vehicle.  ...  Emphasis is placed on leveraging multi-view geometry and deep learning to enable the vehicle to localize and perceive in 3D space.  ...  AutoVision vehicle by extracting local SIFT features [23] from the images captured by the cameras mounted on the AutoVision vehicle and matching the descriptors of these 2D features against the descriptors  ... 
arXiv:1809.05477v2 fatcat:psxp34p4hbfzdjvxtnrvj26ysy

Towards Urban 3D Reconstruction from Video

A. Akbarzadeh, J.-M. Frahm, P. Mordohai, B. Clipp, C. Engels, D. Gallup, P. Merrell, M. Phelps, S. Sinha, B. Talton, L. Wang, Q. Yang (+6 others)
2006 Third International Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT'06)  
The paper introduces a data collection system and a processing pipeline for automatic geo-registered 3D reconstruction of urban scenes from video.  ...  We present the main considerations in designing the system and the steps of the processing pipeline. We show results on real video sequences captured by our system.  ...  Acknowledgement This work is partially supported by DARPA under the UrbanScape project, which is lead by the Geo-Spatial Technologies Information Division of SAIC.  ... 
doi:10.1109/3dpvt.2006.141 dblp:conf/3dpvt/AkbarzadehFMCEGMPSTWYSYWTNP06 fatcat:mcwdxl47rvfbzerp3zv43afexq

Varifocal Multiview Images: Capturing and Visual Tasks [article]

Kejun Wu, Qiong Liu, Guoan Li, Gangyi Jiang, You Yang
2021 arXiv   pre-print
VFMV images are captured by focusing a scene on distinct depths by varying focal planes, and each view only focused on one single plane.Therefore, VFMV images contain more information in focal dimension  ...  than multiview images, and can provide a rich representation for 3D scene by considering both FoV and DoF.  ...  The focus cues of VFMV images enhance the ability to detect light field feature. 3D Reconstruction In this experiment, 3D reconstruction is conducted by two views structure from motion.  ... 
arXiv:2111.10099v1 fatcat:rpouwj36rncv5a7m6mwk7cprra

Imaging sensor technology for intelligent vehicle active safety and driver assistant systems

Bakhtiar B. Litkouhi, Yuri Owechko, Vikas Kukshya, Rajesh Rajavel
2012 International Journal of Vehicle Autonomous Systems  
On the other hand, inexpensive vision sensors can capture the scene image in high spatial resolution and a wide fi eld of view, which makes them ideal for object recognition and lane following under most  ...  The list of such sensors includes imaging sensors operating in different wavelength bands of the visible (i.e., video cameras) and IR spectrum, as well as ranging sensors such as ultrasonic, radar and  ...  By analysing the distortions in a known coded light pattern projected onto a 3D scene, a reconstruction of the 3D structure can be performed.  ... 
doi:10.1504/ijvas.2012.051247 fatcat:dqwxovvtrvhxlmg6fve2cnueky

Challenges in Wide-area Structure-from-motion

Marc Pollefeys, Jan-Michael Frahm, Friedrich Fraundorfer, Christopher Zach, Changchang Wu, Brian Clipp, David Gallup
2010 IPSJ Transactions on Computer Vision and Applications  
The topic of this paper is wide area structure from motion. We first describe recent progress in obtaining large-scale 3D visual models from images.  ...  Our approach consists of a multi-stage processing pipeline, which can process a recorded video stream in real-time on standard PC hardware by leveraging the computational power of the graphics processor  ...  In Section 4.2 we discuss several approaches to obtain the absolute scale of motion from cameras mounted on a vehicles.  ... 
doi:10.2197/ipsjtcva.2.105 fatcat:qwp24vver5botb32uaqgqemvre

Stereo-vision based 3D modeling for unmanned ground vehicles

Stephen Se, Piotr Jasiobedzki, Grant R. Gerhart, Douglas W. Gage, Charles M. Shoemaker
2007 Unmanned Systems Technology IX  
Instant Scene Modeler (iSM) is a vision system for generating calibrated photo-realistic 3D models of unknown environments quickly using stereo image sequences.  ...  The photo-realism of the models enhances the situational awareness of the mission and the models can also be used for change detection. iSM has been tested on our autonomous vehicle to create photo-realistic  ...  Stereo imaging is a passive technique and can recover the structure of the environment by matching features detected in multiple images of the same scene.  ... 
doi:10.1117/12.718399 fatcat:6ym7rcwnybanbmfcbswp5ovs5i

Texture Segmentation of Road Environment Scene Using SfM Module and HLAC Features

Yousun Kang, Koichiro Yamaguchi, Takashi Naito, Yoshiki Ninomiya
2009 IPSJ Transactions on Computer Vision and Applications  
The proposed method can effectively recognize texture-based objects of a road scene by considering their three-dimensional structure in a perspective 2D image.  ...  This paper presents a new image segmentation method for the recognition of texture-based objects in a road environment scene.  ...  Input images were captured using a multi-band camera mounted on a moving vehicle.  ... 
doi:10.2197/ipsjtcva.1.220 fatcat:5lifs6lhffcxlgztjb74pti2te

Integrated 3D display and imaging using dual purpose passive screen and head-mounted projectors and camera

Shoaib R. Soomro, Hakan Urey
2018 Optics Express  
The second layer is an imaging surface made up of an array of curved mirrors, which form the perspective views of the scene captured by a head-mounted camera.  ...  Furthermore, the capture quality is determined by measuring the imaging MTF of the captured views and the capture light efficiency is calculated by considering the loss in transmitted light at each interface  ...  Acknowledgment We would like to thank Kaan Akşit and Osman Eldeş for helpful discussions on computational image reconstruction.  ... 
doi:10.1364/oe.26.001161 pmid:29401993 fatcat:djy3jt2lv5hl5heoiwq2mj4a2e

東日本大震災の被災と復興の画像アーカイブの構築とコンピュータビジョンによる被災市街地の時空間モデリング
Image Archive of 3.11 Earthquake and Tsunami Disasters and Spatio-Temporal Modeling of Town Areas Supported by Computer Vision Techniques

Koichiro DEUGUCHI
2017 Okan  
By those image sets, employing the computer vision techniques, we have been studying the spatial modelling of the temporal changes of city structure by the disaster and the afterward recovery process.  ...  Total amount of image data up to now, for six year activity, becomes more than 150 million 360 scenes of 100 Tera Bytes.  ...  Deguchi, Detecting Changes in 3D Structure of a Scene from Multi-view Images Captured by a Vehicle-mounted Camera, to appear in IEEE Conference on Computer Vision and Pattern Recognition 2013 (CVPR2013  ... 
doi:10.11487/trafst.11.2_116 fatcat:tkohq3jpqfhijethszfifk6skm

Practical camera auto-calibration based on object appearance and motion for traffic scene visual surveillance

Zhaoxiang Zhang, Min Li, Kaigi Huang, Tieniu Tan
2008 2008 IEEE Conference on Computer Vision and Pattern Recognition  
Secondly, calibrated camera makes it possible to recover metrics from images which are robust to scene or view angle changes.  ...  In addition, with calibrated cameras, we can make use of prior information of 3D models to estimate 3D pose of objects and make object detection or tracking more robust to noise and occlusions.  ...  Acknowledgement This work is funded by research grants from the National Basic Research Program of China (2004CB318110), the National Science Foundation (60605014, 60332010, 60335010 and 2004DFA06900),  ... 
doi:10.1109/cvpr.2008.4587780 dblp:conf/cvpr/ZhangLHT08 fatcat:rh62bzbggffpveq7uty7hubeku

Composition and Application of Current Advanced Driving Assistance System: A Review [article]

Xinran Li, Kuo-Yi Lin, Min Meng, Xiuxian Li, Li Li, Yiguang Hong, Jie Chen
2021 arXiv   pre-print
This paper makes a general introduction about ADAS by analyzing its hardware support and computation algorithms.  ...  The latest progress in this field has called for a review to sum up the conventional knowledge of ADAS, the state-of-the-art researches, and novel applications in real-world.  ...  [62] has rebuilt the surrounding 3D construction with the high-density information from stereo camera for scene detection by putting point clouds feature into a pre-trained mode to make scene classification  ... 
arXiv:2105.12348v3 fatcat:ymblcizg65e7hhmch5c5mly22q

Content-based 3D mosaics for dynamic urban scenes

Zhigang Zhu, Hao Tang, George Wolberg, Jeffery R. Layne, Kevin L. Priddy, Emre Ertin
2006 Intelligent Computing: Theory and Applications IV  
We propose a content-based 3D mosaic (CB3M) representation for long video sequences of 3D and dynamic scenes captured by a camera on a mobile platform.  ...  In the first step, a set of parallel-perspective (pushbroom) mosaics with varying viewing directions is generated to capture both the 3D and dynamic aspects of the scene under the camera coverage.  ...  We propose a content-based 3D mosaic representation (CB3M) for long video sequences of 3D and dynamic scenes captured by a camera mounted on a mobile platform.  ... 
doi:10.1117/12.664200 fatcat:stwwrlh4l5an5bqeritiwoxoza
« Previous Showing results 1 — 15 out of 3,985 results