440 Hits in 9.1 sec

Testing a Biologically-Based System for Extracting Depth from Brief Monocular 2-D Video Sequences

John A. Perrone, Michael J. Cree, M. Hedayati, Dale Corlett
2018 2018 International Conference on Image and Vision Computing New Zealand (IVCNZ)  
This allows us to recover depth information from monocular video sequences and here we report on a series of tests that assess the accuracy of this novel approach to 3-D depth recovery.  ...  Some passive (non-emitting) systems use stereo cameras but only a relatively small number of techniques attempt to solve the 3-D layout problem using the information from a single video camera.  ...  This paper represents a first step in demonstrating that a neural-based architecture can extract depth from monocular 2-D video sequences. VI.  ... 
doi:10.1109/ivcnz.2018.8634781 dblp:conf/ivcnz/PerroneCHC18 fatcat:5lpmllaxyzdnvmahc3zx4pbyzm

Estimating heading direction from monocular video sequences using biologically-based sensors

Michael J. Cree, John A. Perrone, Gehan Anthonys, Aden C. Garnett, Henry Gouk
2016 2016 International Conference on Image and Vision Computing New Zealand (IVCNZ)  
Our biologically-based method produced accurate heading estimates over a wide range of test angles and camera speeds.  ...  The traditional computer vision approach to this problem tracks visual features across frames in order to obtain 2-D image motion estimates from which the camera motion can be derived.  ...  One of the main goals of this line of research is to obtain 3-D depth estimates from the monocular 2-D video input, something that humans can do with very brief stimulus exposures.  ... 
doi:10.1109/ivcnz.2016.7804435 dblp:conf/ivcnz/CreePAGG16 fatcat:g664aogylna3rdeudguppl6kre

Research on Basketball Shooting Action Based on Image Feature Extraction and Machine Learning

Rong Ji
2020 IEEE Access  
First of all, the action posture data of basketball players is collected by image feature extraction method, and multi-dimensional motion posture features are extracted from time domain and frequency domain  ...  In order to effectively identify the basketball player's sports posture and improve the athlete's training effect, this paper proposes a basketball shooting gesture recognition method based on image feature  ...  Urtasun use a balanced Gaussian process dynamic model to guide the tracking of three-dimensional human motion in a monocular video sequence.  ... 
doi:10.1109/access.2020.3012456 fatcat:55yhmenbxrcl5phut443rpm3he

An Overview of Perception and Decision-Making in Autonomous Systems in the Era of Learning [article]

Yang Tang, Chaoqiang Zhao, Jianrui Wang, Chongzhen Zhang, Qiyu Sun, Weixing Zheng, Wenli Du, Feng Qian, Juergen Kurths
2020 arXiv   pre-print
learning-based monocular depth estimation, ego-motion prediction, image enhancement, object detection, semantic segmentation, and their combinations with traditional SLAM frameworks.  ...  In this review, we focus on the applications of learning-based approaches in perception and decision-making in autonomous systems, which is different from previous reviews that discussed traditional methods  ...  [116] 2019 where d i stands for the predicted depth and d gt i refers to the groundtruth of depth. thr denotes the threshold, usually 1.25, 1.25 2 , and 1.25 3 . • RMSE = 1 T ∑ i∈T d i − d gt i 2  ... 
arXiv:2001.02319v3 fatcat:z3zhp2cyonfqtlttl2y57572uy

Vehicle Surround Capture: Survey of Techniques and a Novel Omni-Video-Based Approach for Dynamic Panoramic Surround Maps

T. Gandhi, M.M. Trivedi
2006 IEEE transactions on intelligent transportation systems (Print)  
A novel approach for synthesizing the DPS using stereo and motion analysis of video images from a pair of. omni cameras on the vehicle is developed.  ...  Omni cameras, which give a panoramic view of the surroundings, can be useful for visualizing and analyzing the nearby surroundings of the vehicle.  ...  McCall for the help on the hardware of the car test bed and the final proofreading of this paper, as well as S. Cheng and S. Krotosky for the help on the stereo software.  ... 
doi:10.1109/tits.2006.880635 fatcat:irusuaxnzbanvau43t2nzfczne

Recent Advances in Monocular 2D and 3D Human Pose Estimation: A Deep Learning Perspective [article]

Wu Liu, Qian Bao, Yu Sun, Tao Mei
2021 arXiv   pre-print
Estimation of the human pose from a monocular camera has been an emerging research topic in the computer vision community with many applications.  ...  Recently, benefited from the deep learning technologies, a significant amount of research efforts have greatly advanced the monocular human pose estimation both in 2D and 3D areas.  ...  Furthermore, given a video sequence, 2D pose estimation can exploit temporal information to boost keypoint prediction in a video system.  ... 
arXiv:2104.11536v1 fatcat:tdag2jq2vjdrjekwukm5nu7l6a

StereoNeuroBayesSLAM: A Neurobiologically Inspired Stereo Visual SLAM System Based on Direct Sparse Method [article]

Taiping Zeng, Xiaoli Li, Bailu Si
2020 arXiv   pre-print
We propose a neurobiologically inspired visual simultaneous localization and mapping (SLAM) system based on direction sparse method to real-time build cognitive maps of large-scale environments from a  ...  Our proposed SLAM system is robust to real-time build a coherent semi-metric topological map from a stereo camera.  ...  Fig. 10 . 10 Our cognitive maps in sequences 02, 05, 06 and 09 from the KITTI odometry benchmark dataset. (A) Sequence 02. (B) Sequence 05. (C) Sequence 06. (D) Sequence 09.  ... 
arXiv:2003.03091v1 fatcat:z4amrhjqqvexjj334nwnv4wgpq

3D Recognition Based on Sensor Modalities for Robotic Systems: A Survey

Sumaira Manzoor, Sung-Hyeon Joo, Eun-Jin Kim, Sang-Hyeon Bae, Gun-Gyo In, Jeong-Won Pyo, Tae-Yong Kuc
2021 Sensors  
3D visual recognition is a prerequisite for most autonomous robotic systems operating in the real world.  ...  as a guide to those who are interested in vision-based robotics applications.  ...  Acknowledgments: Thanks for the help of reviewers and editors. Conflicts of Interest: The authors declare no conflict of interest.  ... 
doi:10.3390/s21217120 pmid:34770429 pmcid:PMC8587961 fatcat:qqoqlkaycbc4xk24j7j2nzqabe

Research and Implementation of Robot Vision Scanning Tracking Algorithm Based on Deep Learning

Haifeng Guo, Wenyi Li, Na Zhou, He Sun, Zhao Han, Danilo Pelusi
2022 Scanning  
It mainly uses the combination of traditional TLD algorithm and GOTURN algorithm to benefit from a large number of offline training data and updates the learner online, so that the whole system has better  ...  In order to solve the difficult problem of deep learning-based robot vision tracking algorithm research and implementation, a deep learning-based target tracking algorithm and a classical tracking algorithm  ...  Acknowledgments This work was supported by Basic Scientific Research Project of Liaoning Provincial Department of Education, "research on key technologies of health assessment of high safety equipment based  ... 
doi:10.1155/2022/3330427 pmid:35950087 pmcid:PMC9345732 fatcat:5hcje4iqtzagdjdlzqzu3twrkq

Visual simultaneous localization and mapping: a survey

Jorge Fuentes-Pacheco, José Ruiz-Ascencio, Juan Manuel Rendón-Mancha
2012 Artificial Intelligence Review  
SLAM is an essential task for the autonomy of a robot.  ...  The objective of this article is to provide new researchers in the field of visual SLAM a brief and comprehensible review of the state-of-the-art.  ...  Acknowledgments This paper has been made possible thanks to the generous support from the following institutions which we are pleased to acknowledge: CONACYT (Consejo Nacional de Ciencia y Tecnología)  ... 
doi:10.1007/s10462-012-9365-8 fatcat:ghi44bijkzf4xehlxvln5dl67y

Review of constraints on vision-based gesture recognition for human–computer interaction

Biplab Ketan Chakraborty, Debajit Sarma, M.K. Bhuyan, Karl F MacDorman
2018 IET Computer Vision  
This paper surveys major constraints on vision-based gesture recognition occurring in detection and pre-processing, representation and feature extraction, and recognition.  ...  Gesture recognition has applications ranging from sign language to medical assistance to virtual reality.  ...  The features can be obtained either from 2D or 3D image sequences (e.g., RGB-D).  ... 
doi:10.1049/iet-cvi.2017.0052 fatcat:fs2l3o27vveqjlanldjmziuudi

A Learning-Based Visual Saliency Prediction Model for Stereoscopic 3D Video (LBVS-3D) [article]

Amin Banitalebi-Dehkordi, Mahsa T. Pourazad, Panos Nasiopoulos
2018 arXiv   pre-print
Considering that the human visual system has evolved in a natural 3D environment, it is only natural to want to design visual attention models for 3D content.  ...  A new fovea-based model of spatial distance between the image regions is adopted for considering local and global feature calculations.  ...  As a result, we add 1% of the display height to d 1 and d 3 and 1% of the display width to d 2 and d 4 to reach the four points.  ... 
arXiv:1803.04842v1 fatcat:r24vlokfe5avtkxilwguep7ftm

Content-based image sequence representation [chapter]

Radu Jasinschi, Charnchai Pluempitiwiriyawej, José Moura, Pedro Aguiar
2004 Computer Engineering Series  
This is achieved by 2-D shape segmentation for which there is a wealth of methods and approaches. The chapter includes a brief description of active contour methods for image segmentation.  ...  A basic pre-processing step in content-based image sequence representation is to extract and track the relevant background and foreground objects.  ...  The work of the first author was partially supported by the (Portuguese) Foundation for Science and Technology grant POSI/SRI/41561/2001.  ... 
doi:10.1201/9780203486788.ch2 fatcat:5dar6sipxbfbljqen4ir4ztjlm

Recognition of Humans and Their Activities Using Video

Rama Chellappa, Amit K. Roy-Chowdhury, S. Kevin Zhou
2005 Synthesis Lectures on Image Video and Multimedia Processing  
The input is a monocular video sequence. We choose an appropriate two-frame depth reconstruction strategy [157] .  ...  However reliably recovering joint angles from a monocular video is a hard problem.  ... 
doi:10.2200/s00002ed1v01y200508ivm001 fatcat:anrgvxwlr5aylpk4qwjr7bc7ba

Surround-View Cameras based Holistic Visual Perception for Automated Driving [article]

Varun Ravi Kumar
2022 arXiv   pre-print
The dynamics changed from a primitive organism waiting for the food to come into contact for eating food being sought after by visual sensors.  ...  Humans have evolved a biological perception algorithm capable of driving cars, operating machinery, piloting aircraft, and navigating ships over millions of years.  ...  We present a novel self-supervised scale-aware framework for learning Euclidean distance and ego-motion by exploiting geometrical constraints in a sequence of images extracted from raw monocular fisheye  ... 
arXiv:2206.05542v1 fatcat:cdpn6afpvvf7hnsvry7cqbjq3u
« Previous Showing results 1 — 15 out of 440 results