Filters








8,411 Hits in 6.8 sec

Pixels, voxels, and views: A study of shape representations for single view 3D object shape prediction [article]

Daeyun Shin, Charless C. Fowlkes, Derek Hoiem
2018 arXiv   pre-print
The goal of this paper is to compare surface-based and volumetric 3D object shape representations, as well as viewer-centered and object-centered reference frames for single-view 3D shape prediction.  ...  We propose a new algorithm for predicting depth maps from multiple viewpoints, with a single depth or RGB image as input.  ...  More study is needed in this direction. Conclusion Recent methods to produce 3D shape from a single image have used a variety of representation for shape (voxels, octrees, multiple depth maps).  ... 
arXiv:1804.06032v2 fatcat:h3qu4m3afnbgphopq4vzgxdkka

Pixels, Voxels, and Views: A Study of Shape Representations for Single View 3D Object Shape Prediction

Daeyun Shin, Charless C. Fowlkes, Derek Hoiem
2018 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition  
The goal of this paper is to compare surface-based and volumetric 3D object shape representations, as well as viewer-centered and object-centered reference frames for single-view 3D shape prediction.  ...  We propose a new algorithm for predicting depth maps from multiple viewpoints, with a single depth or RGB image as input.  ...  More study is needed in this direction. Conclusion Recent methods to produce 3D shape from a single image have used a variety of representation for shape (voxels, octrees, multiple depth maps).  ... 
doi:10.1109/cvpr.2018.00323 dblp:conf/cvpr/ShinFH18 fatcat:nqhebnzh5zfllfo7iggwu2gsqa

PVSeRF: Joint Pixel-, Voxel- and Surface-Aligned Radiance Field for Single-Image Novel View Synthesis [article]

Xianggang Yu, Jiapeng Tang, Yipeng Qin, Chenghong Li, Linchao Bao, Xiaoguang Han, Shuguang Cui
2022 arXiv   pre-print
We present PVSeRF, a learning framework that reconstructs neural radiance fields from single-view RGB images, for novel view synthesis.  ...  To address this challenge, we propose to incorporate explicit geometry reasoning and combine it with pixel-aligned features for radiance field prediction.  ...  Single-view 3D Object Reconstruction Given a single image containing a object, 3D object reconstruction aims to recover the 3D geometry of the object.  ... 
arXiv:2202.04879v1 fatcat:iacxkdhstfgrncxr4wxvux4zrm

Multi-view Supervision for Single-View Reconstruction via Differentiable Ray Consistency

Shubham Tulsiani, Tinghui Zhou, Alexei A. Efros, Jitendra Malik
2017 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)  
for learning single-view 3D prediction.  ...  We study the notion of consistency between a 3D shape and a 2D observation and propose a differentiable formulation which allows computing gradients of the 3D shape given an observation from an arbitrary  ...  We gratefully acknowledge NVIDIA corporation for the donation of Tesla GPUs used for this research.  ... 
doi:10.1109/cvpr.2017.30 dblp:conf/cvpr/TulsianiZEM17 fatcat:kvvkkk75ljdj7fbzlwif4ik3na

Multi-view Supervision for Single-view Reconstruction via Differentiable Ray Consistency [article]

Shubham Tulsiani, Tinghui Zhou, Alexei A. Efros, Jitendra Malik
2017 arXiv   pre-print
for learning single-view 3D prediction.  ...  We study the notion of consistency between a 3D shape and a 2D observation and propose a differentiable formulation which allows computing gradients of the 3D shape given an observation from an arbitrary  ...  We gratefully acknowledge NVIDIA corporation for the donation of Tesla GPUs used for this research.  ... 
arXiv:1704.06254v1 fatcat:xmzflirmizdx7a5iq62yq75k2m

Learning View Priors for Single-view 3D Reconstruction [article]

Hiroharu Kato, Tatsuya Harada
2019 arXiv   pre-print
Because of this ambiguity, although a 3D object reconstructor can be trained using a single view or a few views per object, reconstructed shapes only fit the observed views and appear incorrect from the  ...  There is some ambiguity in the 3D shape of an object when the number of observed views is small.  ...  Acknowledgment This work was partially funded by ImPACT Program of Council for Science, Technology and Innovation (Cabinet Office, Government of Japan) and partially supported by JST CREST Grant Number  ... 
arXiv:1811.10719v2 fatcat:3jnpm3zalbg7la5ples34fgw4a

3D Shape Reconstruction from Sketches via Multi-view Convolutional Networks [article]

Zhaoliang Lun, Matheus Gadelha, Evangelos Kalogerakis, Subhransu Maji, Rui Wang
2017 arXiv   pre-print
We propose a method for reconstructing 3D shapes from 2D sketches in the form of line drawings.  ...  Our method takes as input a single sketch, or multiple sketches, and outputs a dense point cloud representing a 3D reconstruction of the input sketch(es).  ...  We acknowledge the MassTech Collaborative grant for funding the UMass GPU cluster.  ... 
arXiv:1707.06375v3 fatcat:pphspwefmrbuxjlvnd7bxy75tu

Multi-View Silhouette and Depth Decomposition for High Resolution 3D Object Representation [article]

Edward Smith, Scott Fujimoto, David Meger
2018 arXiv   pre-print
Drawing motivation from the canonical view representation of objects, we introduce a novel method for the fast up-sampling of 3D objects in voxel space through networks that perform super-resolution on  ...  We evaluate our work on multiple experiments concerning high-resolution 3D objects, and show our system is capable of accurately predicting novel objects at resolutions as large as 512×512×512 -- the highest  ...  This decoder receives a latent representation of the 3D shape and produces a probability for occupancy at each discrete position in 3D voxel space.  ... 
arXiv:1802.09987v3 fatcat:qj27gfxsxnawrozxhfk77beada

ShapeCodes: Self-Supervised Feature Learning by Lifting Views to Viewgrids [article]

Dinesh Jayaraman, Ruohan Gao, Kristen Grauman
2018 arXiv   pre-print
We introduce an unsupervised feature learning approach that embeds 3D shape information into a single-view image representation.  ...  The main idea is a self-supervised training objective that, given only a single 2D image, requires all unseen views of the object to be predictable from learned features.  ...  In short, the function of the encoder is to lift a 2D view to a single vector representation of the full 3D object shape.  ... 
arXiv:1709.00505v4 fatcat:ccyklhukyfhfboqssezau26uzi

MeshMVS: Multi-View Stereo Guided Mesh Reconstruction [article]

Rakesh Shrestha, Zhiwen Fan, Qingkun Su, Zuozhuo Dai, Siyu Zhu, Ping Tan
2021 arXiv   pre-print
First, our system predicts a coarse 3D volume from the color images by probabilistically merging voxel occupancy grids from the prediction of individual views.  ...  Deep learning based 3D shape generation methods generally utilize latent features extracted from color images to encode the semantics of objects and guide the shape generation process.  ...  Our system can struggle to roughly reconstruct shapes with very complex topology while some fine topology of the mesh is missing.  ... 
arXiv:2010.08682v3 fatcat:mszoiznjofcqllyc77ywdkqogu

Learning a Multi-View Stereo Machine [article]

Abhishek Kar, Christian Häne, Jitendra Malik
2017 arXiv   pre-print
In contrast to recent learning based methods for 3D reconstruction, we leverage the underlying 3D geometry of the problem through feature projection and unprojection along viewing rays.  ...  We present a learnt system for multi-view stereopsis.  ...  The authors would like to thank David Fouhey, Saurabh Gupta and Shubham Tulsiani for valuable discussions and Fyusion Inc. for providing GPU hours for the work.  ... 
arXiv:1708.05375v1 fatcat:o5ism2lpmrabfmpxa5k7ecgfjq

Conditional Single-view Shape Generation for Multi-view Stereo Reconstruction [article]

Yi Wei, Shaohui Liu, Wang Zhao, Jiwen Lu, Jie Zhou
2019 arXiv   pre-print
Then, we formulate the task of multi-view reconstruction as taking the intersection of the predicted shape spaces on each single image.  ...  Experimental results and ablation studies show that our proposed approach outperforms state-of-the-art methods on 3D reconstruction test error and demonstrate its generalization ability on real world data  ...  Acknowledgements This work was supported in part by the National Natural Science Foundation of China under Grant U1813218, Grant 61822603, Grant U1713214, Grant 61672306, and Grant 61572271.  ... 
arXiv:1904.06699v2 fatcat:khnrizpp6vdqloidfnweyrsjwy

Multi-view to Novel View: Synthesizing Novel Views With Self-learned Confidence [chapter]

Shao-Hua Sun, Minyoung Huh, Yuan-Hong Liao, Ning Zhang, Joseph J. Lim
2018 Lecture Notes in Computer Science  
Specifically, our model consists of a flow prediction module and a pixel generation module to directly leverage information presented in source views as well as hallucinate missing pixels from statistical  ...  We evaluate our model on images rendered from 3D object models as well as real and synthesized scenes.  ...  Acknowledgments This project was supported by the center for super intelligence, Kakao Brain, and SKT. This research of Shao-Hua Sun and Minyoung Huh were partially supported by Snap Inc.  ... 
doi:10.1007/978-3-030-01219-9_10 fatcat:x54uve2m4zdnrgcvlnczfjp67i

Multi-view Consistency as Supervisory Signal for Learning Shape and Pose Prediction [article]

Shubham Tulsiani, Alexei A. Efros, Jitendra Malik
2018 arXiv   pre-print
We present a framework for learning single-view shape and pose prediction without using direct supervision for either.  ...  Our proposed training setup enforces geometric consistency between the independently predicted shape and pose from two views of the same instance.  ...  This work was supported in part by Intel/NSF VEC award IIS-1539099 and NSF Award IIS-1212798. We gratefully acknowledge NVIDIA corporation for the donation of GPUs used for this research.  ... 
arXiv:1801.03910v2 fatcat:fj7s4amks5g75feh7ky7356aza

3D Sketching using Multi-View Deep Volumetric Prediction

Johanna Delanoy, Mathieu Aubry, Phillip Isola, Alexei A. Efros, Adrien Bousseau
2018 Proceedings of the ACM on Computer Graphics and Interactive Techniques  
We complement this single-view network with an updater CNN that refines an existing prediction given a new drawing of the shape created from a novel viewpoint.  ...  At the core of our approach is a deep convolutional neural network (CNN) that predicts occupancy of a voxel grid from a line drawing.  ...  ACKNOWLEDGMENTS Many thanks to Yulia Gryaditskaya for sketching several of our results, and for her help on the renderings and video.  ... 
doi:10.1145/3203197 dblp:journals/pacmcgit/DelanoyAIEB18 fatcat:vwt4h6uo5rgejae42wo4s6nqga
« Previous Showing results 1 — 15 out of 8,411 results