Generative Models

Motion by Queries: Identity-Motion Trade-offs in Text-to-Video Generation

StyleGAN-NADA: CLIP-Guided Domain Adaptation of Image Generators

Can a generative model be trained to produce images from a specific domain, guided by a text prompt only, without seeing any image? In other words: can an image generator be trained blindly?