A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2021; you can also visit the original URL.
The file type is application/pdf
.
Filters
Towards Open-World Text-Guided Face Image Generation and Manipulation
[article]
2021
arXiv
pre-print
More importantly, our method supports open-world scenarios, including both image and text, without any re-training, fine-tuning, or post-processing. ...
To be specific, we propose a brand new paradigm of text-guided image generation and manipulation based on the superior characteristics of a pretrained GAN model. ...
Diving Deep into Open-World Images and Texts In this section, we illustrate the results of images and texts with real-world variances. ...
arXiv:2104.08910v1
fatcat:aexf7c26ancr7lwja5sonaj4ji
Multimodal Image Synthesis and Editing: A Survey
[article]
2022
arXiv
pre-print
We then describe multimodal image synthesis and editing approaches extensively with detailed frameworks including Generative Adversarial Networks (GANs), Auto-regressive models, Diffusion models, Neural ...
As information exists in various modalities in real world, effective interaction and fusion among multimodal information plays a key role for the creation and perception of multimodal data in computer ...
ACKNOWLEDGMENTS This study is supported under the RIE2020 Industry Alignment Fund -Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry ...
arXiv:2112.13592v3
fatcat:46twjhz3hbe6rpm33k6ilnisga
Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language Model
[article]
2022
arXiv
pre-print
., Predict, Prevent, and Evaluate (PPE), for disentangled text-driven image manipulation that requires little manual annotation while being applicable to a wide variety of manipulations. ...
Finally, we propose a new evaluation metric to Evaluate the disentangled image manipulation. We verify the effectiveness of our method on the challenging face editing task. ...
This work was supported by the PRIN project CREATIVE Prot. 2020ZSL9F9, by the EUREGIO project OLIVER and by the EU H2020 AI4Media project under Grant 951911. ...
arXiv:2111.13333v2
fatcat:ewsrexmxmbbrthvhrftgm73k2i
SemanticAdv: Generating Adversarial Examples via Attribute-conditional Image Editing
[article]
2020
arXiv
pre-print
In this paper, we aim to explore the impact of semantic manipulation on DNNs predictions by manipulating the semantic attributes of images and generate "unrestricted adversarial examples". ...
To further demonstrate the applicability of SemanticAdv beyond face recognition domain, we also generate semantic perturbations on street-view images. ...
pure image generation [55,31,6], attributeto-image generation [75,12], text-to-image generation [44,56,49,48,78,28], and image-to-image translation [26,81,39,68,24]. ...
arXiv:1906.07927v4
fatcat:tyduj5qtsjhcbiyeqhybb3xhmm
Covert Ideologies and Sign Manipulation: a Functional Semiotic Analysis of Western vs. Arabic News Coverage of New Zealand 2019 Mosque Carnage
2021
International Journal of Applied Linguistics and English Literature
The heinous crime has left the world in a state of mourning and utter shock. Media coverage of this attack overwhelmed the Western and the Arab newspapers as well as social media worldwide. ...
The primary concern of this comparative study is to explore sign manipulation in front page photographs and headlines in both Western and Arabic news reports in a humble attempt to determine how visual ...
Semiotics can be applied to 'texts' which involve signs such as sounds and visual images. ...
doi:10.7575/aiac.ijalel.v.10n.6p.50
fatcat:bmv2c3qccnei7hxwbu6egml4pi
NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion
[article]
2021
arXiv
pre-print
Furthermore, it also shows surprisingly good zero-shot capabilities on text-guided image and video manipulation tasks. Project repo is https://github.com/microsoft/NUWA. ...
Compared to several strong baselines, N\"UWA achieves state-of-the-art results on text-to-image generation, text-to-video generation, video prediction, etc. ...
Furthermore, N ÜWA shows surprisingly good zero-shot capabilities not only on text-guided image manipulation, but also text-guided video manipulation. ...
arXiv:2111.12417v1
fatcat:uqu73laxdjfcdbytxbj5xxk3te
Virtual Life Network: A Body-Centered Networked Virtual Environment
1997
Presence - Teleoperators and Virtual Environments
To further enhance the realistic feel of the virtual environment and to simplify object manipulation we provide the facility of defining object behaviors by attaching motor functions to the objects. ...
Since it is generally not possible to track all degrees of freedom of the human body in order to reproduce the realistic body motion, we introduce the motor functions that generate natural motion for standard ...
We would like to thank assistants of LIG and MIRALAB for the models and libraries. ...
doi:10.1162/pres.1997.6.6.676
fatcat:dczwnle5srbszpjwhlz4oktk2q
Blended Latent Diffusion
[article]
2022
arXiv
pre-print
The tremendous progress in neural image generation, coupled with the emergence of seemingly omnipotent vision-language models has finally enabled text-based interfaces for creating and editing images. ...
In this paper, we present an accelerated solution to the task of local text-driven editing of generic images, where the desired edits are confined to a user-provided mask. ...
ACKNOWLEDGMENTS This work was supported in part by the Israel Science Foundation (grants No. 2492/20 and 1574/21). ...
arXiv:2206.02779v1
fatcat:o4pty25yx5gnjkl7u5fax5hwt4
Deep Learning and Synthetic Media
[article]
2022
arXiv
pre-print
, and can be indistinguishable from real sounds and images recorded with a sensor. ...
After reviewing important etiological features of deep learning pipelines for media manipulation and generation, I argue that "deepfakes" and related synthetic media produced with such pipelines do not ...
It has also become increasingly easy to guide image generation directly with text. ...
arXiv:2205.05764v1
fatcat:6th5uy6zifgydj4cbecvs3fw4u
Kapow! Bam! Wham!
2022
Study and Scrutiny Research on Young Adult Literature
role in their world. ...
I identified four themes: (a) adults serving as guides, (b) adults and authority, (c) the careless and reckless teen, and (d) adolescents as social change makers. ...
Marvel awakens a lost generation of teens manipulated by the Inventor. ...
doi:10.15763/issn.2376-5275.2022.5.2.220-240
fatcat:iprx7tgvbndsjdxsxaoioq3nga
Reflexivity as a Process for Coming Into Knowing
2019
LEARNing Landscapes
Acting reflexively implies a "self-critical and self-conscious stance" (Glass, 2015, p. 555) of recognizing myself within a research process as an intentional participant-practitioner of generating knowledge ...
My aim is to affirm the intentionality of my reflexive praxis as a way of knowing and becoming through committed intertextual inquiry and discovery. ...
By creating reflexive visual texts, I begin to construct parts of an artist/researcher/teacher identity through overlays of guiding information generating potential pathways of knowing. ...
doi:10.36510/learnland.v12i1.985
fatcat:h2bzxlpwo5fq3fdcusb6rg2gb4
War, Lies, and the News Photo: Second World War Photographic Propaganda in PM's Weekly (1940–1941)
2014
RACAR Revue d art canadienne
Due diligence has been made to locate the copyright holders of the images reproduced in this article. ...
Acknowledgements Thanks to Kim Sichel and Mitchell Frank for help with earlier versions of this article. ...
At the same time, the PM captions and texts insert other forms of connotation by directing viewers toward specific pictorial details, imposing alternate readings of the images, and deflating the images ...
doi:10.7202/1027747ar
fatcat:dflzyr23wrdirizocpzgoitej4
Guiding Visual Attention in Deep Convolutional Neural Networks Based on Human Eye Movements
[article]
2022
arXiv
pre-print
We use human eye tracking data to directly modify training examples and thereby guide the models' visual attention during object recognition in natural images either towards or away from the focus of human ...
Our results demonstrate that the proposed guided focus manipulation works as intended in the negative direction and non-human-like models focus on significantly dissimilar image parts compared to humans ...
While data from human participants displayed a strong face bias (Mdn = 2.74), which varied considerably across time, DCNNs generally demonstrated clearly less focused attention towards face regions (STD-nets ...
arXiv:2206.10587v1
fatcat:i76vpszsxnglrd4kqxl5mcmfh4
StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators
[article]
2021
arXiv
pre-print
Can a generative model be trained to produce images from a specific domain, guided by a text prompt only, without seeing any image? In other words: can an image generator be trained "blindly"? ...
Leveraging the semantic power of large scale Contrastive-Language-Image-Pre-training (CLIP) models, we present a text-driven method that allows shifting a generative model to new domains, without having ...
Acknowledgments We thank Yuval Alaluf, Ron Mokady and Ethan Fetaya for reviewing early drafts and helpful suggestions. ...
arXiv:2108.00946v2
fatcat:lnn4ydsoenauxbpu6ijpm3ccn4
NimbRo@Home: Winning Team of the RoboCup@Home Competition 2012
[chapter]
2013
Lecture Notes in Computer Science
This year we improved the gripper design of our robots and further advanced mobile manipulation capabilities such as object perception and manipulation planning. ...
For human-robot interaction, we propose to complement faceto-face communication between user and robot with a remote user interface for handheld PCs. ...
We estimate the pose of objects in RGB-D images through real-time registration towards the model. ...
doi:10.1007/978-3-642-39250-4_10
fatcat:ypx2qsnwq5hgnayguidwtcfohm
« Previous
Showing results 1 — 15 out of 73,475 results