73,475 Hits in 8.0 sec

Towards Open-World Text-Guided Face Image Generation and Manipulation [article]

Weihao Xia, Yujiu Yang, Jing-Hao Xue, Baoyuan Wu
2021 arXiv   pre-print
More importantly, our method supports open-world scenarios, including both image and text, without any re-training, fine-tuning, or post-processing.  ...  To be specific, we propose a brand new paradigm of text-guided image generation and manipulation based on the superior characteristics of a pretrained GAN model.  ...  Diving Deep into Open-World Images and Texts In this section, we illustrate the results of images and texts with real-world variances.  ... 
arXiv:2104.08910v1 fatcat:aexf7c26ancr7lwja5sonaj4ji

Multimodal Image Synthesis and Editing: A Survey [article]

Fangneng Zhan, Yingchen Yu, Rongliang Wu, Jiahui Zhang, Shijian Lu, Lingjie Liu, Adam Kortylewski, Christian Theobalt, Eric Xing
2022 arXiv   pre-print
We then describe multimodal image synthesis and editing approaches extensively with detailed frameworks including Generative Adversarial Networks (GANs), Auto-regressive models, Diffusion models, Neural  ...  As information exists in various modalities in real world, effective interaction and fusion among multimodal information plays a key role for the creation and perception of multimodal data in computer  ...  ACKNOWLEDGMENTS This study is supported under the RIE2020 Industry Alignment Fund -Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry  ... 
arXiv:2112.13592v3 fatcat:46twjhz3hbe6rpm33k6ilnisga

Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language Model [article]

Zipeng Xu, Tianwei Lin, Hao Tang, Fu Li, Dongliang He, Nicu Sebe, Radu Timofte, Luc Van Gool, Errui Ding
2022 arXiv   pre-print
., Predict, Prevent, and Evaluate (PPE), for disentangled text-driven image manipulation that requires little manual annotation while being applicable to a wide variety of manipulations.  ...  Finally, we propose a new evaluation metric to Evaluate the disentangled image manipulation. We verify the effectiveness of our method on the challenging face editing task.  ...  This work was supported by the PRIN project CREATIVE Prot. 2020ZSL9F9, by the EUREGIO project OLIVER and by the EU H2020 AI4Media project under Grant 951911.  ... 
arXiv:2111.13333v2 fatcat:ewsrexmxmbbrthvhrftgm73k2i

SemanticAdv: Generating Adversarial Examples via Attribute-conditional Image Editing [article]

Haonan Qiu, Chaowei Xiao, Lei Yang, Xinchen Yan, Honglak Lee, Bo Li
2020 arXiv   pre-print
In this paper, we aim to explore the impact of semantic manipulation on DNNs predictions by manipulating the semantic attributes of images and generate "unrestricted adversarial examples".  ...  To further demonstrate the applicability of SemanticAdv beyond face recognition domain, we also generate semantic perturbations on street-view images.  ...  pure image generation [55,31,6], attributeto-image generation [75,12], text-to-image generation [44,56,49,48,78,28], and image-to-image translation [26,81,39,68,24].  ... 
arXiv:1906.07927v4 fatcat:tyduj5qtsjhcbiyeqhybb3xhmm

Covert Ideologies and Sign Manipulation: a Functional Semiotic Analysis of Western vs. Arabic News Coverage of New Zealand 2019 Mosque Carnage

Waleed Alsoufi, Naser N. Albzour
2021 International Journal of Applied Linguistics and English Literature  
The heinous crime has left the world in a state of mourning and utter shock. Media coverage of this attack overwhelmed the Western and the Arab newspapers as well as social media worldwide.  ...  The primary concern of this comparative study is to explore sign manipulation in front page photographs and headlines in both Western and Arabic news reports in a humble attempt to determine how visual  ...  Semiotics can be applied to 'texts' which involve signs such as sounds and visual images.  ... 
doi:10.7575/aiac.ijalel.v.10n.6p.50 fatcat:bmv2c3qccnei7hxwbu6egml4pi

NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion [article]

Chenfei Wu, Jian Liang, Lei Ji, Fan Yang, Yuejian Fang, Daxin Jiang, Nan Duan
2021 arXiv   pre-print
Furthermore, it also shows surprisingly good zero-shot capabilities on text-guided image and video manipulation tasks. Project repo is  ...  Compared to several strong baselines, N\"UWA achieves state-of-the-art results on text-to-image generation, text-to-video generation, video prediction, etc.  ...  Furthermore, N ÜWA shows surprisingly good zero-shot capabilities not only on text-guided image manipulation, but also text-guided video manipulation.  ... 
arXiv:2111.12417v1 fatcat:uqu73laxdjfcdbytxbj5xxk3te

Virtual Life Network: A Body-Centered Networked Virtual Environment

Igor-Sunday Pandzic, Nadia Magnenat Thalmann, Tolga K. Capin, Daniel Thalmann
1997 Presence - Teleoperators and Virtual Environments  
To further enhance the realistic feel of the virtual environment and to simplify object manipulation we provide the facility of defining object behaviors by attaching motor functions to the objects.  ...  Since it is generally not possible to track all degrees of freedom of the human body in order to reproduce the realistic body motion, we introduce the motor functions that generate natural motion for standard  ...  We would like to thank assistants of LIG and MIRALAB for the models and libraries.  ... 
doi:10.1162/pres.1997.6.6.676 fatcat:dczwnle5srbszpjwhlz4oktk2q

Blended Latent Diffusion [article]

Omri Avrahami, Ohad Fried, Dani Lischinski
2022 arXiv   pre-print
The tremendous progress in neural image generation, coupled with the emergence of seemingly omnipotent vision-language models has finally enabled text-based interfaces for creating and editing images.  ...  In this paper, we present an accelerated solution to the task of local text-driven editing of generic images, where the desired edits are confined to a user-provided mask.  ...  ACKNOWLEDGMENTS This work was supported in part by the Israel Science Foundation (grants No. 2492/20 and 1574/21).  ... 
arXiv:2206.02779v1 fatcat:o4pty25yx5gnjkl7u5fax5hwt4

Deep Learning and Synthetic Media [article]

Raphaël Millière
2022 arXiv   pre-print
, and can be indistinguishable from real sounds and images recorded with a sensor.  ...  After reviewing important etiological features of deep learning pipelines for media manipulation and generation, I argue that "deepfakes" and related synthetic media produced with such pipelines do not  ...  It has also become increasingly easy to guide image generation directly with text.  ... 
arXiv:2205.05764v1 fatcat:6th5uy6zifgydj4cbecvs3fw4u

Kapow! Bam! Wham!

Sara Abi Villanueva
2022 Study and Scrutiny Research on Young Adult Literature  
role in their world.  ...  I identified four themes: (a) adults serving as guides, (b) adults and authority, (c) the careless and reckless teen, and (d) adolescents as social change makers.  ...  Marvel awakens a lost generation of teens manipulated by the Inventor.  ... 
doi:10.15763/issn.2376-5275.2022.5.2.220-240 fatcat:iprx7tgvbndsjdxsxaoioq3nga

Reflexivity as a Process for Coming Into Knowing

Karen McGarry
2019 LEARNing Landscapes  
Acting reflexively implies a "self-critical and self-conscious stance" (Glass, 2015, p. 555) of recognizing myself within a research process as an intentional participant-practitioner of generating knowledge  ...  My aim is to affirm the intentionality of my reflexive praxis as a way of knowing and becoming through committed intertextual inquiry and discovery.  ...  By creating reflexive visual texts, I begin to construct parts of an artist/researcher/teacher identity through overlays of guiding information generating potential pathways of knowing.  ... 
doi:10.36510/learnland.v12i1.985 fatcat:h2bzxlpwo5fq3fdcusb6rg2gb4

War, Lies, and the News Photo: Second World War Photographic Propaganda in PM's Weekly (1940–1941)

Carol Payne
2014 RACAR Revue d art canadienne  
Due diligence has been made to locate the copyright holders of the images reproduced in this article.  ...  Acknowledgements Thanks to Kim Sichel and Mitchell Frank for help with earlier versions of this article.  ...  At the same time, the PM captions and texts insert other forms of connotation by directing viewers toward specific pictorial details, imposing alternate readings of the images, and deflating the images  ... 
doi:10.7202/1027747ar fatcat:dflzyr23wrdirizocpzgoitej4

Guiding Visual Attention in Deep Convolutional Neural Networks Based on Human Eye Movements [article]

Leonard E. van Dyck, Sebastian J. Denzler, Walter R. Gruber
2022 arXiv   pre-print
We use human eye tracking data to directly modify training examples and thereby guide the models' visual attention during object recognition in natural images either towards or away from the focus of human  ...  Our results demonstrate that the proposed guided focus manipulation works as intended in the negative direction and non-human-like models focus on significantly dissimilar image parts compared to humans  ...  While data from human participants displayed a strong face bias (Mdn = 2.74), which varied considerably across time, DCNNs generally demonstrated clearly less focused attention towards face regions (STD-nets  ... 
arXiv:2206.10587v1 fatcat:i76vpszsxnglrd4kqxl5mcmfh4

StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators [article]

Rinon Gal, Or Patashnik, Haggai Maron, Gal Chechik, Daniel Cohen-Or
2021 arXiv   pre-print
Can a generative model be trained to produce images from a specific domain, guided by a text prompt only, without seeing any image? In other words: can an image generator be trained "blindly"?  ...  Leveraging the semantic power of large scale Contrastive-Language-Image-Pre-training (CLIP) models, we present a text-driven method that allows shifting a generative model to new domains, without having  ...  Acknowledgments We thank Yuval Alaluf, Ron Mokady and Ethan Fetaya for reviewing early drafts and helpful suggestions.  ... 
arXiv:2108.00946v2 fatcat:lnn4ydsoenauxbpu6ijpm3ccn4

NimbRo@Home: Winning Team of the RoboCup@Home Competition 2012 [chapter]

Jörg Stückler, Ishrat Badami, David Droeschel, Kathrin Gräve, Dirk Holz, Manus McElhone, Matthias Nieuwenhuisen, Michael Schreiber, Max Schwarz, Sven Behnke
2013 Lecture Notes in Computer Science  
This year we improved the gripper design of our robots and further advanced mobile manipulation capabilities such as object perception and manipulation planning.  ...  For human-robot interaction, we propose to complement faceto-face communication between user and robot with a remote user interface for handheld PCs.  ...  We estimate the pose of objects in RGB-D images through real-time registration towards the model.  ... 
doi:10.1007/978-3-642-39250-4_10 fatcat:ypx2qsnwq5hgnayguidwtcfohm
« Previous Showing results 1 — 15 out of 73,475 results