Filters








33,563 Hits in 8.3 sec

Measuring the impact of temporal context on video retrieval

Daragh Byrne, Peter Wilkins, Gareth J.F. Jones, Alan F. Smeaton, Noel E. O'Connor
2008 Proceedings of the 2008 international conference on Content-based image and video retrieval - CIVR '08  
We developed two interfaces with identical retrieval functionality in order to measure the effects of such context on user performance.  ...  In this paper we describe the findings from the K-Space interactive video search experiments in TRECVid 2007, which examined the effects of including temporal context in video retrieval.  ...  We also extend our thanks to the participants in our experiments.  ... 
doi:10.1145/1386352.1386393 dblp:conf/civr/ByrneWJSO08 fatcat:gy3jkax2kbeffhgagu7rihqx3e

Interest seam image

Xiao Zhang, Gang Hua, Lei Zhang, Heung-Yeung Shum
2010 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition  
The representation capacity of the proposed interest seam image is demonstrated in a large scale video retrieval task.  ...  The optimal seam is used to extract a seam of pixels from each video frame to form one column of an image, based on which an interest seam image is finally composited.  ...  This indicates that our video retrieval system could be scaled to web video dataset. Impact of Gaussian central prior The impact of the parameter β in Eq. 2 is tested on Qr.  ... 
doi:10.1109/cvpr.2010.5540042 dblp:conf/cvpr/ZhangHZS10 fatcat:n2dhm7f4u5cotgpu4gvp655m6y

User variance and its impact on video retrieval benchmarking

Peter Wilkins, Raphaël Troncy, Martin Halvey, Daragh Byrne, Alia Amin, P. Punitha, Alan F. Smeaton, Robert Villa
2009 Proceeding of the ACM International Conference on Image and Video Retrieval - CIVR '09  
In this paper, we describe one of the largest multi-site interactive video retrieval experiments conducted in a laboratory setting.  ...  Interactive video retrieval performance is difficult to cross-compare as variables exist across users, interfaces and the underlying retrieval engine.  ...  Acknowledgments This paper was supported by the European Commission under contract FP6-027026 (K-Space) and by Science Foundation Ireland under grant 07/CE/I1147 (CLARITY: Centre for Sensor Web Technologies  ... 
doi:10.1145/1646396.1646400 dblp:conf/civr/WilkinsTHBAPSV09 fatcat:u6j5iqwiljhg5kzppoelaqhy34

Exploring the Temporal Cues to Enhance Video Retrieval on Standardized CDVA

Won Jo, Guentaek Lim, Joonsoo Kim, Joungil Yun, Yukyung Choi
2022 IEEE Access  
Therefore, we conduct self-evaluations of CDVA to analyze the impact of each module on the retrieval task.  ...  As the demand for large-scale video analysis increases, video retrieval research is also becoming more active.  ...  In this paper, to make headway with this standardized method, we analyze the impact on the video retrieval task with self-evaluations.  ... 
doi:10.1109/access.2022.3165177 fatcat:wxsq42u46vdsrjez5d4kuzgcxm

Diagnosing Error in Temporal Action Detectors [chapter]

Humam Alwassel, Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem
2018 Lecture Notes in Computer Science  
Our analysis shows that the most impactful areas to work on are: strategies to better handle temporal context around the instances, improving the robustness w.r.t. the instance absolute and relative size  ...  Despite the recent progress in video understanding and the continuous rate of improvement in temporal action localization throughout the years, it is still unclear how far (or close?)  ...  This publication is based upon work supported by the King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research (OSR) under Award No. OSR-CRG2017-3405.  ... 
doi:10.1007/978-3-030-01219-9_16 fatcat:c3umlpr3xnco3lunevuiprkcei

Diagnosing Error in Temporal Action Detectors [article]

Humam Alwassel, Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem
2018 arXiv   pre-print
Our analysis shows that the most impactful areas to work on are: strategies to better handle temporal context around the instances, improving the robustness w.r.t. the instance absolute and relative size  ...  Despite the recent progress in video understanding and the continuous rate of improvement in temporal action localization throughout the years, it is still unclear how far (or close?)  ...  This publication is based upon work supported by the King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research (OSR) under Award No. OSR-CRG2017-3405.  ... 
arXiv:1807.10706v1 fatcat:ozerrprikjhxnlouginxozozwm

Learn from Unlabeled Videos for Near-duplicate Video Retrieval

Xiangteng He, Yulin Pan, Mingqian Tang, Yiliang Lv, Yuxin Peng
2022 Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval  
of video retrieval on accuracy and efficiency.  ...  Near-duplicate video retrieval (NDVR) aims to find the copies or transformations of the query video from a massive video database.  ...  Besides, we further evaluate the impact of each transformation on the retrieval performance of frame-level encoding.  ... 
doi:10.1145/3477495.3532010 fatcat:iafkdqw725egfem7f5gghrvkcu

Event Retrieval in Large Video Collections with Circulant Temporal Encoding

Jerome Revaud, Matthijs Douze, Cordelia Schmid, Herve Jegou
2013 2013 IEEE Conference on Computer Vision and Pattern Recognition  
Given a video clip of a specific event, e.g., the wedding of Prince William and Kate Middleton, the goal is to retrieve other videos representing the same event from a dataset of over 100k videos.  ...  Our approach encodes the frame descriptors of a video to jointly represent their appearance and temporal order.  ...  We thank Jonathan Delhumeau for helping with the annotation of EVVE.  ... 
doi:10.1109/cvpr.2013.318 dblp:conf/cvpr/RevaudDSJ13 fatcat:ojlcocxzfre47e5sc6z6rzhktq

FeEval A Dataset for Evaluation of Spatio-temporal Local Features

Julian Stottinger, Sebastian Zambanini, Rehanullah Khan, Allan Hanbury
2010 2010 20th International Conference on Pattern Recognition  
Similar to prior work on 2D images, this leads to a repeatability and matching measurement in videos for spatiotemporal features estimating the overlap of features under increasing changes in the data.  ...  For the first time, this dataset allows for a systematic measurement of the stability and the invariance of local features in videos.  ...  Acknowledgments This work was partly supported by the Austrian Research Promotion Agency (FFG) project OMOR 815994, MuBisA 819862 and the CogVis 7 Ltd.  ... 
doi:10.1109/icpr.2010.128 dblp:conf/icpr/StottingerZKH10 fatcat:ggaipddt4bfpxdntcj2dehk2jm

ViSiL: Fine-grained Spatio-Temporal Video Similarity Learning [article]

Giorgos Kordopatis-Zilos, Symeon Papadopoulos, Ioannis Patras, Ioannis Kompatsiaris
2019 arXiv   pre-print
We train the proposed network using a triplet loss scheme and evaluate it on five public benchmark datasets on four different video retrieval problems where we demonstrate large improvements in comparison  ...  video retrieval approaches that embed the whole frame or even the whole video into a vector descriptor before the similarity estimation.  ...  Acknowledgments: This work is supported by the WeVerify H2020 project, partially funded by the EU under contract numbers 825297. The work of Ioannis Patras has been supported by EPSRC under grant No.  ... 
arXiv:1908.07410v1 fatcat:64wjijh2ifetbfwibz6njr4rlm

ViSiL: Fine-grained Spatio-Temporal Video Similarity Learning

Giorgos Kordopatis-Zilos, Symeon Papadopoulos, Ioannis Patras, Ioannis Kompatsiaris
2019 Zenodo  
We train the proposed network using a triplet loss scheme and evaluate it on five public benchmark datasets on four different video retrieval problems where we demonstrate large improvements in comparison  ...  video retrieval approaches that embed the whole frame or even the whole video into a vector descriptor before the similarity estimation.  ...  Acknowledgments: This work is supported by the WeVerify H2020 project, partially funded by the EU under contract numbers 825297. The work of Ioannis Patras has been supported by EPSRC under grant No.  ... 
doi:10.5281/zenodo.3615770 fatcat:5gsi5s7h3ngs3fd47vu4nzthai

Multi-scale 2D Representation Learning for weakly-supervised moment retrieval [article]

Ding Li, Rui Wu, Yongqiang Tang, Zhizhong Zhang, Wensheng Zhang
2021 arXiv   pre-print
Video moment retrieval aims to search the moment most relevant to a given language query.  ...  Experiments on two benchmark datasets Charades-STA and ActivityNet Captions demonstrate that our approach achieves superior performance to state-of-the-art results.  ...  Discussion In this section, we mainly discuss the impact of the selection of temporal scales and the impact of the loss weight, some of Impact of Multiple Temporal Scales.  ... 
arXiv:2111.02741v1 fatcat:fmvmp2k3xvcjlp3d6fahgiqkiq

Localizing Moments in Video with Natural Language [article]

Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, Bryan Russell
2017 arXiv   pre-print
We consider retrieving a specific temporal segment, or moment, from a video given a natural language text description.  ...  We demonstrate that MCN outperforms several baseline methods and believe that our initial results together with the release of DiDeMo will inspire further research on localizing video moments with natural  ...  Rows 9-12 of Table 3 demonstrate the importance of temporal context for moment retrieval.  ... 
arXiv:1708.01641v1 fatcat:sgrv3qlhhfaujh6szkoxgwgmqa

Self-supervised Video Retrieval Transformer Network [article]

Xiangteng He, Yulin Pan, Mingqian Tang, Yiliang Lv
2021 arXiv   pre-print
Comprehensive experiments on two challenging video retrieval datasets, namely FIVR-200K and SVD, verify the effectiveness of our proposed SVRTN method, which achieves the best performance of video retrieval  ...  Content-based video retrieval aims to find videos from a large video database that are similar to or even near-duplicate of a given query video.  ...  Impact of clip length on clip-level encoding.  ... 
arXiv:2104.07993v1 fatcat:5lpvrgxrkvci7ntaruzhtlxqau

Video Object Mining: Issues and Perspectives

Jonathan Weber, Sebastien Lefevre, Pierre Gancarski
2010 2010 IEEE Fourth International Conference on Semantic Computing  
Today, video is becoming one of the primary sources of information.  ...  This gap can be bridged by relying on the real objects present in videos because of the semantic meaning of objects.  ...  ACKNOWLEDGEMENTS This work has been supported by Ready Business System, Entzheim, France and the French National Association for Research and Technology (ANRT).  ... 
doi:10.1109/icsc.2010.71 dblp:conf/semco/WeberLG10 fatcat:ylrd66b72jfghpjn4sudvefifa
« Previous Showing results 1 — 15 out of 33,563 results