Filters








24,643 Hits in 5.9 sec

PRAFlow_RVC: Pyramid Recurrent All-Pairs Field Transforms for Optical Flow Estimation in Robust Vision Challenge 2020 [article]

Zhexiong Wan, Yuxin Mao, Yuchao Dai
<span title="2020-09-14">2020</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
RAFT (Recurrent All Pairs Field Transforms) currently represents the state-of-the-art in optical flow estimation.  ...  Optical flow estimation is an important computer vision task, which aims at estimating the dense correspondences between two frames.  ...  Very recently, Recurrent All-Pairs Field Transforms (RAFT) [1] builds multi-scale 4D correlation volumes for all pairs of pixels, and iteratively updates a flow field through a gated recurrent unit (  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2009.06360v1">arXiv:2009.06360v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/wzblk76ky5h6pp2t2ga2dckfy4">fatcat:wzblk76ky5h6pp2t2ga2dckfy4</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20200921222910/https://arxiv.org/pdf/2009.06360v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2009.06360v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

V-PROM: A Benchmark for Visual Reasoning Using Visual Progressive Matrices

Damien Teney, Peng Wang, Jiewei Cao, Lingqiao Liu, Chunhua Shen, Anton Van den Hengel
<span title="2020-04-03">2020</span> <i title="Association for the Advancement of Artificial Intelligence (AAAI)"> <a target="_blank" rel="noopener" href="https://fatcat.wiki/container/wtjcymhabjantmdtuptkk62mlq" style="color: black;">PROCEEDINGS OF THE THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE AND THE TWENTY-EIGHTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE</a> </i> &nbsp;
We evaluate a range of deep learning architectures, and find that existing models, including those popular for vision-and-language tasks, are unable to solve seemingly-simple instances.  ...  Advances in machine learning have generated increasing enthusiasm for tasks that require high-level reasoning on top of perceptual capabilities, particularly over visual data.  ...  In (Hoshen and Werman 2017) , the authors propose a CNN model to solve problems involving geometric operations such as rotations and reflections. Barrett et al.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1609/aaai.v34i07.6885">doi:10.1609/aaai.v34i07.6885</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/v63l3mjlcbewhpbrr3yabmupwe">fatcat:v63l3mjlcbewhpbrr3yabmupwe</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20201104164406/https://aaai.org/ojs/index.php/AAAI/article/download/6885/6739" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/32/00/320050c091702da78a583869ed5895135e0dcc51.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1609/aaai.v34i07.6885"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="external alternate icon"></i> Publisher / doi.org </button> </a>

LRTA: A Transparent Neural-Symbolic Reasoning Framework with Modular Supervision for Visual Question Answering [article]

Weixin Liang, Feiyang Niu, Aishwarya Reganti, Govind Thattai, Gokhan Tur
<span title="2020-11-21">2020</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
We propose LRTA [Look, Read, Think, Answer], a transparent neural-symbolic reasoning framework for visual question answering that solves the problem step-by-step like humans and provides human-readable  ...  It then executes the reasoning instructions one at a time by traversing the scene graph using a recurrent neural-symbolic execution module.  ...  Acknowledgement We would like to thank Robinson Piramuthu, Dilek Hakkani-Tur, Arindam Mandal, Yanbang Wang and the anonymous reviewers for their insightful feedback and discussions that have notably shaped  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2011.10731v1">arXiv:2011.10731v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/jk7gpqjhvjdpnosoyc6ccpjfde">fatcat:jk7gpqjhvjdpnosoyc6ccpjfde</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20201126043518/https://arxiv.org/pdf/2011.10731v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/86/fd/86fdbc4540d146b0a2d7d61bf9f0109fa1331dac.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2011.10731v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

SASRA: Semantically-aware Spatio-temporal Reasoning Agent for Vision-and-Language Navigation in Continuous Environments [article]

Muhammad Zubair Irshad, Niluthpol Chowdhury Mithun, Zachary Seymour, Han-Pang Chiu, Supun Samarasekera, Rakesh Kumar
<span title="2021-08-26">2021</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
In this regard, we present a hybrid transformer-recurrence model which focuses on combining classical semantic mapping techniques with a learning-based method.  ...  This paper presents a novel approach for the Vision-and-Language Navigation (VLN) task in continuous 3D environments, which requires an autonomous agent to follow natural language instructions in unseen  ...  Semantic-Linguistic Attention Map Transformer (SLAM-T) and RGBD-Linguistic Transformer consistently reason between visual and textual spatial domains.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2108.11945v1">arXiv:2108.11945v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/paep2xj5czhzbokiyyekmof4qq">fatcat:paep2xj5czhzbokiyyekmof4qq</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20210906154211/https://arxiv.org/pdf/2108.11945v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/66/64/6664bfb49a5f3e5a5dd60b5cdafe3a63a53d834d.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2108.11945v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Convolutional neural networks for modeling and forecasting nonlinear nonstationary processes

Andrii Belas, Petro Bidyuk
<span title="2021-06-30">2021</span> <i title="OU Scientific Route"> <a target="_blank" rel="noopener" href="https://fatcat.wiki/container/lc3plfwkdnfnhdzumyi2jdqv2u" style="color: black;">ScienceRise</a> </i> &nbsp;
It indicates that one-dimensional convolutional neural networks can be a quite reasonable choice for solving time series forecasting problems. The area of practical use of the research results.  ...  There are several popular approaches to solving the problems of adequate model constructing and forecasting nonlinear nonstationary processes, such as autoregressive models and recurrent neural networks  ...  It clearly indicates that 1D CNNs can be quite reasonable choice for solving time series forecasting problems.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.21303/2313-8416.2021.001924">doi:10.21303/2313-8416.2021.001924</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/5iqhssi7evanjgcvrb576662be">fatcat:5iqhssi7evanjgcvrb576662be</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20210709053556/http://journal.eu-jr.eu/sciencerise/article/download/1924/1661" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/d1/c5/d1c5784199fc19d08b7d4488324ab697074e08d7.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.21303/2313-8416.2021.001924"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="unlock alternate icon" style="background-color: #fb971f;"></i> Publisher / doi.org </button> </a>

Stable and expressive recurrent vision models [article]

Drew Linsley, Alekh Karkada Ashok, Lakshmi Narasimhan Govindarajan, Rex Liu, Thomas Serre
<span title="2020-10-22">2020</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
Primate vision depends on recurrent processing for reliable perception.  ...  We further show that recurrent vision models trained with C-RBP to solve the large-scale Panoptic Segmentation MS-COCO challenge outperform the leading feedforward approach, with fewer free parameters.  ...  Additional support from the Brown University Carney Institute for Brain Science, Initiative for Computation in Brain and Mind, and Center for Computation and Visualization (CCV).  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2005.11362v2">arXiv:2005.11362v2</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/7hg3cvah3fblfkcpgnhsnbiddu">fatcat:7hg3cvah3fblfkcpgnhsnbiddu</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20201029063902/https://arxiv.org/pdf/2005.11362v2.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/53/34/5334ee7be861cc8f8023316b00743c0e36f55e22.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2005.11362v2" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Hierarchical Cross-Modal Agent for Robotics Vision-and-Language Navigation [article]

Muhammad Zubair Irshad, Chih-Yao Ma, Zsolt Kira
<span title="2021-04-21">2021</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
Deep Learning has revolutionized our ability to solve complex problems such as Vision-and-Language Navigation (VLN).  ...  baselines in all key metrics and sets a new benchmark for Robo-VLN.  ...  Abstract-Deep Learning has revolutionized our ability to solve complex problems such as Vision-and-Language Navigation (VLN).  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2104.10674v1">arXiv:2104.10674v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/zkbxnpilifgxtjg4z4gs6uwdry">fatcat:zkbxnpilifgxtjg4z4gs6uwdry</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20210423182136/https://arxiv.org/pdf/2104.10674v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/f9/91/f991840d345172846d5ea43ad61313f056de1901.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2104.10674v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Understanding the computational demands underlying visual reasoning [article]

Mohit Vaishnav, Remi Cadene, Andrea Alamia, Drew Linsley, Rufin VanRullen, Thomas Serre
<span title="2021-12-09">2021</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
visual reasoning problems.  ...  on the type of visual reasoning problem.  ...  Additional support was provided by the ANR-3IA Artificial and Natural Intelligence Toulouse Institute (ANR-19-PI3A-0004), the Center for Computation and Visualization (CCV) and High Performance Computing  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2108.03603v2">arXiv:2108.03603v2</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/qnylgwdicvgwfm7a6umxrea2ta">fatcat:qnylgwdicvgwfm7a6umxrea2ta</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20211211084557/https://arxiv.org/pdf/2108.03603v2.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/e7/51/e751c865595eed2c073f6709a7ad4350c6dd6be9.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2108.03603v2" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Deep neural networks: a new framework for modelling biological vision and brain information processing [article]

Nikolaus Kriegeskorte
<span title="2015-10-26">2015</span> <i title="Cold Spring Harbor Laboratory"> bioRxiv </i> &nbsp; <span class="release-stage" >pre-print</span>
Convolutional feedforward networks, which now dominate computer vision, take further inspiration from the architecture of the primate visual hierarchy.  ...  Human-level visual recognition abilities are coming within reach of artificial systems.  ...  The section Recurrent neural nets for vision describes networks using recurrent computation.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1101/029876">doi:10.1101/029876</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/lxuwpdhzrvhpdmtyzg33ogwncy">fatcat:lxuwpdhzrvhpdmtyzg33ogwncy</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20190503153604/https://www.biorxiv.org/content/biorxiv/early/2015/10/26/029876.full.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/04/d5/04d539af1874e4e63ee55fe12ae6aeaf77391354.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1101/029876"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="external alternate icon"></i> biorxiv.org </button> </a>

A survey on Information Visualization in light of Vision and Cognitive sciences [article]

Jose Rodrigues-Jr, Luciana Zaina, Maria Oliveira, Bruno Brandoli, and Agma Traina
<span title="2016-05-13">2016</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
In the intent of promoting a better picture, here, we survey concepts on vision, cognition, and Information Visualization organized in a theorization named Visual Expression Process.  ...  Information Visualization techniques are built on a context with many factors related to both vision and cognition, making it difficult to draw a clear picture of how data visually turns into comprehension  ...  (Ware 2005) A top-down (problem-solving) model that states that what we see in a visualization depends on what we are seeking for.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/1505.07079v2">arXiv:1505.07079v2</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/amlp7vteyrctxdhjmcqiwptfgm">fatcat:amlp7vteyrctxdhjmcqiwptfgm</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20191025104834/https://arxiv.org/pdf/1505.07079v2.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/54/ec/54ecdc0e83a57252b722f3d48b887db5305ef900.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/1505.07079v2" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

U-Net and its variants for Medical Image Segmentation : A short review [article]

Vinay Ummadi
<span title="2022-04-17">2022</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
Following how different hybrid architectures have built upon strong techniques from visual recognition tasks.  ...  In the end we will see current challenges and future directions for medical image segmentation(MIS).  ...  General tasks in computer vision has reached reasonable performance on most datasets but this is not the case with the medical vision tasks.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2204.08470v1">arXiv:2204.08470v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/me5mmiu6qvbyfedzlug67mj2ka">fatcat:me5mmiu6qvbyfedzlug67mj2ka</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20220421012557/https://arxiv.org/pdf/2204.08470v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/0d/99/0d99a105749399ee11ec181c5335ceacfe596d60.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2204.08470v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Recurrent processing during object recognition

D. Wyatte, R. O'Reilly
<span title="2011-09-23">2011</span> <i title="Association for Research in Vision and Ophthalmology (ARVO)"> <a target="_blank" rel="noopener" href="https://fatcat.wiki/container/rahwjcylyzhtlni57joehivgqm" style="color: black;">Journal of Vision</a> </i> &nbsp;
of visual learning.  ...  Furthermore, this interaction of recurrent connectivity and learning predicts that high-level visual representations should be shaped by error signals from nearby, associated brain areas over the course  ...  ACKNOWLEDGMENTS The authors would like to thank Michael Tarr, Thomas Palmeri, Garrison Cottrell, Tim Curran, and Nicolas Pinto for their helpful comments and suggestions.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1167/11.11.889">doi:10.1167/11.11.889</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/vggce4s4eze4xbxuuxunutcrfi">fatcat:vggce4s4eze4xbxuuxunutcrfi</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20170811060604/https://grey.colorado.edu/mediawiki/sites/mingus/images/b/b9/OReillyWyatteHerdEtAl13.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/97/a5/97a537a5208ca03054970c66d873d3b7dae0d39a.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1167/11.11.889"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="unlock alternate icon" style="background-color: #fb971f;"></i> Publisher / doi.org </button> </a>

Solving Bongard Problems with a Visual Language and Pragmatic Reasoning [article]

Stefan Depeweg, Constantin A. Rothkopf, Frank Jäkel
<span title="2018-04-12">2018</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
More than 50 years ago Bongard introduced 100 visual concept learning problems as a testbed for intelligent vision systems. These problems are now known as Bongard problems.  ...  While this approach is far from solving all Bongard problems, it solves the biggest fraction yet.  ...  Instead of two separate modules-one for vision and one for cognition, each operating on different principles-there might just be one deep, probably recurrent, neural network.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/1804.04452v1">arXiv:1804.04452v1</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/y7aepqqbbjeyjehoibc2xvjt4m">fatcat:y7aepqqbbjeyjehoibc2xvjt4m</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20191017123800/https://arxiv.org/pdf/1804.04452v1.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/0a/a4/0aa4782a0d37c10850ba69de98e655b77238db3d.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/1804.04452v1" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Going in circles is the way forward: the role of recurrence in visual inference [article]

Ruben S. van Bergen, Nikolaus Kriegeskorte
<span title="2020-11-16">2020</span> <i > arXiv </i> &nbsp; <span class="release-stage" >pre-print</span>
Biological visual systems exhibit abundant recurrent connectivity. State-of-the-art neural network models for visual recognition, by contrast, rely heavily or exclusively on feedforward computation.  ...  This important insight suggests that computational neuroscientists may not need to engage recurrent computation, and that computer-vision engineers may be limiting themselves to a special case of FNN if  ...  ACKNOWLEDGEMENTS We thank Samuel Lippl, Heiko Schütt, Andrew Zaharia, Tal Golan and Benjamin Peters for detailed comments on a draft of this paper.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2003.12128v3">arXiv:2003.12128v3</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/c7vjrebe4bbbnna34s4zlashaa">fatcat:c7vjrebe4bbbnna34s4zlashaa</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20201118004114/https://arxiv.org/pdf/2003.12128v3.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/87/1d/871d8f68791f98ca7c54cf0a5a8712da7298117a.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener" href="https://arxiv.org/abs/2003.12128v3" title="arxiv.org access"> <button class="ui compact blue labeled icon button serp-button"> <i class="file alternate outline icon"></i> arxiv.org </button> </a>

Visual Linguistic Model and Its Applications in Image Captioning

Ravin Kumar
<span title="2020-04-16">2020</span> <i title="Springer Science and Business Media LLC"> <a target="_blank" rel="noopener" href="https://fatcat.wiki/container/yccq43c5yrambibcoue7gd3qma" style="color: black;">SN Computer Science</a> </i> &nbsp;
Research work on this problem statement requires efforts in both computer vision and natural language processing domains to obtain better quality image descriptions.  ...  In this approach, we generate a sequence of visual embeddings for objects and their relationships present in the image.  ...  This article is part of the topical collection "Advances in Computational Approaches for Artificial Intelligence, Image Processing, IoT and Cloud Applications" guest edited by Bhanu Prakash K N and M.  ... 
<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1007/s42979-020-00135-w">doi:10.1007/s42979-020-00135-w</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/kzq7ekgmtfdohhp4x5jl4k3y3i">fatcat:kzq7ekgmtfdohhp4x5jl4k3y3i</a> </span>
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20200711081233/https://link.springer.com/content/pdf/10.1007/s42979-020-00135-w.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/3c/bd/3cbd06475b3367da1d28b60655f57038e3c5a67e.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.1007/s42979-020-00135-w"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="external alternate icon"></i> springer.com </button> </a>
&laquo; Previous Showing results 1 &mdash; 15 out of 24,643 results