Filters








5,638 Hits in 7.3 sec

Zero-Shot Text-Guided Object Generation with Dream Fields [article]

Ajay Jain, Ben Mildenhall, Jonathan T. Barron, Pieter Abbeel, Ben Poole
2022 arXiv   pre-print
Instead, we guide generation with image-text models pre-trained on large datasets of captioned images from the web.  ...  Our method, Dream Fields, can generate the geometry and color of a wide range of objects without 3D supervision.  ...  Method In this section, we develop Dream Fields: a zero-shot object synthesis method given only a natural language caption.  ... 
arXiv:2112.01455v2 fatcat:z3mktn6omfcg7hhwcfnowg2p7a

AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars [article]

Fangzhou Hong, Mingyuan Zhang, Liang Pan, Zhongang Cai, Lei Yang, Ziwei Liu
2022 arXiv   pre-print
To democratize this technology to a larger audience, we propose AvatarCLIP, a zero-shot text-driven framework for 3D avatar generation and animation.  ...  Remarkably, AvatarCLIP can generate unseen 3D avatars with novel animations, achieving superior zero-shot capability.  ...  Through direct supervision on images, CLIP shows great success in zero-shot text-guided image generation ].  ... 
arXiv:2205.08535v1 fatcat:ybbcbmjs2fckljkkx3pqgymejy

CLIP-Actor: Text-Driven Recommendation and Stylization for Animating Human Meshes [article]

Kim Youwang, Kim Ji-Yeon, Tae-Hyun Oh
2022 arXiv   pre-print
Then, our novel zero-shot neural style optimization detailizes and texturizes the recommended mesh sequence to conform to the prompt in a temporally-consistent and pose-agnostic manner.  ...  We build a text-driven human motion recommendation system by leveraging a large-scale human motion dataset with language labels.  ...  None of these methods can generate diverse color and cloth details of human motion in a zero-shot manner, e.g., with only a text guide.  ... 
arXiv:2206.04382v2 fatcat:txb5tg5fbjdp7jtebhb2cc4e6e

GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models [article]

Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, Mark Chen
2022 arXiv   pre-print
Diffusion models have recently been shown to generate high-quality synthetic images, especially when paired with a guidance technique to trade off diversity for fidelity.  ...  Additionally, we find that our models can be fine-tuned to perform image inpainting, enabling powerful text-driven image editing.  ...  We would also like to thank Yura Burda for providing feedback on an early draft of this paper, and to Mikhail Pavlov for finding difficult prompts for text-conditional generative models.  ... 
arXiv:2112.10741v3 fatcat:gnvshaybn5d4dfu7iyuhd5i4cu

LaTeRF: Label and Text Driven Object Radiance Fields [article]

Ashkan Mirzaei, Yash Kant, Jonathan Kelly, Igor Gilitschenski
2022 arXiv   pre-print
Neural fields have shown their effectiveness in learning a continuous volumetric representation of a scene from 2D images, but acquiring object representations from these models with weak supervision remains  ...  To faithfully extract the object from the scene, LaTeRF extends the NeRF formulation with an additional 'objectness' probability at each 3D point.  ...  The language module of our method is closely related to dream field [13] but we use it to generate 3D objects that are consistent with the ones provided in a scene.  ... 
arXiv:2207.01583v3 fatcat:6k3pcvhlzba35jgusroo2dshsa

Page 156 of Forest and Stream; A Journal of Outdoor Life, Travel, Nature Study, Shooting, Fishing, Yachting Vol. 90, Issue 3 [page]

1920 Forest and Stream; A Journal of Outdoor Life, Travel, Nature Study, Shooting, Fishing, Yachting  
Any seen by the shooter through this sight when trigger is pulled, is DEAD —as such object must be at the time in shot pattern when gun is discharged.  ...  Information, together with illustrated Booklet and Folder, ase forwarded upon application to . E. PITTMAN, General Passenger Agent.  ... 

Unifying Language Learning Paradigms [article]

Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Dara Bahri, Tal Schuster, Huaixiu Steven Zheng, Neil Houlsby, Donald Metzler
2022 arXiv   pre-print
Our model also achieve strong results at in-context learning, outperforming 175B GPT-3 on zero-shot SuperGLUE and tripling the performance of T5-XXL on one-shot summarization.  ...  We begin by disentangling architectural archetypes with pre-training objectives – two concepts that are commonly conflated.  ...  Finally, we conduct zero/few-shot experiments with UL2 and show that UL2 outperforms GPT-3 175B on zero shot SuperGLUE.  ... 
arXiv:2205.05131v1 fatcat:oixewsfpp5brfkmj2sizcznljm

Red Teaming Language Models with Language Models [article]

Ethan Perez, Saffron Huang, Francis Song, Trevor Cai, Roman Ring, John Aslanides, Amelia Glaese, Nat McAleese, Geoffrey Irving
2022 arXiv   pre-print
We explore several methods, from zero-shot generation to reinforcement learning, for generating test cases with varying levels of diversity and difficulty.  ...  and hospital phone numbers generated as the chatbot's own contact info, leakage of private training data in generated text, and harms that occur over the course of a conversation.  ...  Acknowledgments We thank Angeliki Lazaridou for encouraging us to explore question generation.  ... 
arXiv:2202.03286v1 fatcat:ogptxm22d5e37bzpyv7cizarp4

Multimodal Image Synthesis and Editing: A Survey [article]

Fangneng Zhan, Yingchen Yu, Rongliang Wu, Jiahui Zhang, Shijian Lu, Lingjie Liu, Adam Kortylewski, Christian Theobalt, Eric Xing
2022 arXiv   pre-print
Radiance Fields (NeRF) and other methods.  ...  We then describe multimodal image synthesis and editing approaches extensively with detailed frameworks including Generative Adversarial Networks (GANs), Auto-regressive models, Diffusion models, Neural  ...  Similarly, CLIPMesh [209] presents a technique for zero-shot generation of a 3D model using only a target text prompt.  ... 
arXiv:2112.13592v3 fatcat:46twjhz3hbe6rpm33k6ilnisga

Visualization Component for the Scenario Prototype Generator as a Video Game Development Tool

Gulnara F. Sahibgareeva, Oleg Bedrin, Vlada Kugurakova
2020 Conference on Scientific Services & Internet  
In the future, work will continue to develop and expand the functionality of the script prototype generation tool.  ...  This work belongs to the field of development and research of video games. In particular, the narrative part of video games is considered.  ...  This tool can generate shots based on entered text queries.  ... 
dblp:conf/ssi/SahibgareevaBK20 fatcat:53s34x3spjcmja6ntzgrjpel7m

CrossFit: A Few-shot Learning Challenge for Cross-task Generalization in NLP [article]

Qinyuan Ye, Bill Yuchen Lin, Xiang Ren
2021 arXiv   pre-print
NLP datasets and converted to a unified text-to-text format.  ...  In this paper, we explore whether and how such cross-task generalization ability can be acquired, and further applied to build better few-shot learners across diverse NLP tasks.  ...  Neural text generation from structured data with application to the biography domain.  ... 
arXiv:2104.08835v2 fatcat:xnhrmmsmyzb4fjo7ealrw2vnka

The Worldwood of the Cinematic Image

José Manuel Martins
2012 Phainomenon  
A close analysis of the specifically cinematographic procedure in Akira Kurosawa's 'Dream' Crows reveals it as an articulated and insightful philosophical statement, endowed with general relevance conceming  ...  The Chapter 'Le sentir', in Phénoménologie de la perception, will offer us the proper guide to elucidate what we are already perceiving and conceiving in Kurosawa's film, where the ex-static phenomenological  ...  ABSTRACT A close analysis of the specifically cinematographic procedure in Akira Kurosawa's 'Dream' Crows reveals it as an articulated and insightful philosophical statement, endowed with general relevance  ... 
doi:10.2478/phainomenon-2012-0022 fatcat:m24tjc6xnrfwdd4ivzglc5rjn4

Brainish: Formalizing A Multimodal Language for Intelligence and Consciousness [article]

Paul Pu Liang
2022 arXiv   pre-print
Blum (2021), we describe the desiderata of a multimodal language called Brainish, comprising words, images, audio, and sensations combined in representations that the CTM's processors use to communicate with  ...  consciousness in the CTM, and by implementing a simple version of Brainish and evaluating its capability of demonstrating intelligence on multimodal prediction and retrieval tasks on several real-world image, text  ...  While this may seem straightforward for object-based concepts, the generative process becomes more ambiguous when dealing with non-objects.  ... 
arXiv:2205.00001v3 fatcat:smz3rforbvd6fbfz3rktcwhexe

The Quay Brothers: Choreographed Chiaroscuro, Enigmatic and Sublime

Suzanne H. Buchan
1998 Film quarterly  
In their work with anonymous objects, the Quays demand that their machines and objects act as much if not more than the puppets: "[you] accept their very physicalities palpably as objectified dream or  ...  A moving focal plane is implemented in uncanny point-of-view shots and helps the spatial transitions between dreamer and dream world.  ...  Stunning light design endows this film with its ethereal quality.  ... 
doi:10.1525/fq.1998.51.3.04a00020 fatcat:umrjsrz5wney5ec5isob23qiw4

The Quay Brothers: Choreographed Chiaroscuro, Enigmatic and Sublime

Suzanne H. Buchan
1998 Film quarterly  
In their work with anonymous objects, the Quays demand that their machines and objects act as much if not more than the puppets: "[you] accept their very physicalities palpably as objectified dream or  ...  A moving focal plane is implemented in uncanny point-of-view shots and helps the spatial transitions between dreamer and dream world.  ...  Stunning light design endows this film with its ethereal quality.  ... 
doi:10.2307/1213598 fatcat:5te4bgmvyfciboafqzzb553np4
« Previous Showing results 1 — 15 out of 5,638 results