Hierarchical Feature Warping and Blending for Talking Head Animation
APISR: Anime Production Inspired Real-World Anime Super-Resolution
Re:Draw—Context Aware Translation as a Controllable Method for Artistic Production
MobileDiffusion: Subsecond Text-to-Image Generation on Mobile Devices
UFOGen: You Forward Once Large Scale Text-to-Image Generation via Diffusion GANs
Application of Generative Adversarial Networks in Color Art Image Shadow Generation
FlatGAN: A Holistic Approach for Robust Flat-Coloring in High-Definition with Understanding Line Discontinuity
Consistency Trajectory Models (CTM): Learning Probability Flow ODE Trajectory of Diffusion
The Colorization Based on Self-Attention Mechanism and GAN
Semi-supervised reference-based sketch extraction using a contrastive learning framework
StyleTTS 2: Towards Human-Level Text-to-Speech through Style Diffusion and Adversarial Training with Large Speech Language Models
Vocos: Closing the gap between time-domain and Fourier-based neural vocoders for high-quality audio synthesis
Multi-Label Classification in Anime Illustrations Based on Hierarchical Attribute Relationships
TANGO: Text-to-Audio Generation using Instruction-Tuned LLM and Latent Diffusion Model
Thangka Sketch Colorization Based on Multi-Level Adaptive-Instance-Normalized Color Fusion and Skip Connection Attention
Two-Step Training: Adjustable Sketch Colorization via Reference Image and Text Tag
Approaching an unknown communication system by latent space exploration and causal inference
Enhancing Image Representation in Conditional Image Synthesis
StencilTorch: An Iterative and User-Guided Framework for Anime Lineart Colorization
PMSGAN: Parallel Multistage GANs for Face Image Translation
FAEC-GAN: An unsupervised face-to-anime translation based on edge enhancement and coordinate attention
A survey on text generation using generative adversarial networks
Appearance-preserved Portrait-to-anime Translation via Proxy-guided Domain Adaptation
Null-text Inversion for Editing Real Images using Guided Diffusion Models
An analysis: different methods about line art colorization
Guiding Users to Where to Give Color Hints for Efficient Interactive Sketch Colorization via Unsupervised Region Prioritization
T2CI-GAN: Text to Compressed Image generation using Generative Adversarial Network
GET3D: A Generative Model of High Quality 3D Textured Shapes Learned from Images
Using Generative Adversarial Networks for Conditional Creation of Anime Posters
AnimeSR: Learning Real-World Super-Resolution Models for Animation Videos
StyleTTS: A Style-Based Generative Model for Natural and Diverse Text-to-Speech Synthesis
VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance
Imitating, Fast and Slow: Robust learning from demonstrations via decision-time planning
TATS: Long Video Generation with Time-Agnostic VQGAN and Time-Sensitive Transformer
Truncated Diffusion Probabilistic Models and Diffusion-based Adversarial Autoencoders
Using Constant Learning Rate of Two Time-Scale Update Rule for Training Generative Adversarial Networks
Microdosing: Knowledge Distillation for GAN based Compression
An unsupervised font style transfer model based on generative adversarial networks
Multimodal Conditional Image Synthesis with Product-of-Experts GANs
TnT Attacks! Universal Naturalistic Adversarial Patches Against Deep Neural Network Systems
Tackling the Generative Learning Trilemma with Denoising Diffusion GANs
MSMT-GAN: Multi-Tailed, Multi-Headed, Spatial Dynamic Memory refined Text-to-Image Synthesis
Unpaired font family synthesis using conditional generative adversarial networks
Fake It Till You Make It: Face analysis in the wild using synthetic data alone
MCL-GAN: Generative Adversarial Networks with Multiple Specialized Discriminators
MLP Singer: Towards Rapid Parallel Korean Singing Voice Synthesis
GANs N’ Roses: Stable, Controllable, Diverse Image to Image Translation (works for videos too!)
MixerGAN: An MLP-Based Architecture for Unpaired Image-to-Image Translation
Deep Generative Modeling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models
AniGAN: Style-Guided Generative Adversarial Networks for Unsupervised Anime Face Generation
The Role of AI Attribution Knowledge in the Evaluation of Artwork
XMC-GAN: Cross-Modal Contrastive Learning for Text-to-Image Generation
VQ-GAN: Taming Transformers for High-Resolution Image Synthesis
LDM: Automatic Colorization of Anime Style Illustrations Using a Two-Stage Generator
dStyle-GAN: Generative Adversarial Network based on Writing and Photography Styles for Drug Identification in Darknet Markets
Automatic Colorization of High-resolution Animation Style Line-art based on Frequency Separation and Two-Stage Generator
Image Generators with Conditionally-Independent Pixel Synthesis
RetinaGAN: An Object-aware Approach to Sim-to-Real Transfer
HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis
A Good Image Generator Is What You Need for High-Resolution Video Synthesis
Why Spectral Normalization Stabilizes GANs: Analysis and Improvements
Improving GAN Training with Probability Ratio Clipping and Sample Reweighting
Object Segmentation Without Labels with Large-Scale Generative Models
Generative Adversarial Phonology: Modeling unsupervised phonetic and phonological learning with neural networks
CiwGAN and fiwGAN: Encoding information in acoustic data to model lexical learning with Generative Adversarial Networks
Reference-Based Sketch Image Colorization using Augmented-Self Reference and Dense Semantic Correspondence
Learning to Simulate Dynamic Environments with GameGAN [homepage]
PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models
Do We Need Zero Training Loss After Achieving Zero Training Error?
Smooth markets: A basic mechanism for organizing gradient-based learners
microbatchGAN: Stimulating Diversity with Multi-Adversarial Discrimination
StarGAN Based Facial Expression Transfer for Anime Characters
Deep-Eyes: Fully Automatic Anime Character Colorization with Painting of Details on Empty Pupils
PaintsTorch: a User-Guided Anime Line Art Colorization Tool With Double Generator Conditional Adversarial Network
Interactive Anime Sketch Colorization with Style Consistency via a Deep Residual Neural Network
Parallel WaveGAN: A fast waveform generation model based on generative adversarial networks with multi-resolution spectrogram
Tag2Pix: Line Art Colorization Using Text Tag With SECat and Changing Loss
Anime Sketch Coloring with Swish-gated Residual U-net and Spectrally Normalized GAN (SSN-GAN)
COCO-GAN: Generation by Parts via Conditional Coordinating
InGAN: Capturing and Remapping the "DNA" of a Natural Image
GAN-QP: A Novel GAN Framework without Gradient Vanishing and Lipschitz Constraint
ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks
Twin-GAN: Unpaired Cross-Domain Image Translation with Weight-Sharing GANs
IntroVAE: Introspective Variational Autoencoders for Photographic Image Synthesis
Sem-GAN: Semantically-Consistent Image-to-Image Translation
The relativistic discriminator: a key element missing from standard GAN
An empirical study on evaluation metrics of generative adversarial networks
GANomaly: Semi-Supervised Anomaly Detection via Adversarial Training
Toward Diverse Text Generation with Inverse Reinforcement Learning
Synthesizing Programs for Images using Reinforced Adversarial Learning
A Variational Inequality Perspective on Generative Adversarial Networks
ChatPainter: Improving Text to Image Generation using Dialogue
Spectral Normalization for Generative Adversarial Networks
High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial Networks
StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation
XGAN: Unsupervised Image-to-Image Translation for Many-to-Many Mappings
GraspGAN: Using Simulation and Domain Adaptation to Improve Efficiency of Deep Robotic Grasping
OptionGAN: Learning Joint Reward-Policy Options using Generative Adversarial Inverse Reinforcement Learning
Training Shallow and Thin Networks for Acceleration via Knowledge Distillation with Conditional Adversarial Networks
Towards the Automatic Anime Characters Creation with Generative Adversarial Networks
Learning Universal Adversarial Perturbations with Generative Models
Semi-Supervised Haptic Material Recognition for Robots using Generative Adversarial Networks
Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-To-End Learning from Demonstration
CAN: Creative Adversarial Networks, Generating "Art" by Learning About Styles and Deviating from Style Norms
Language Generation with Recurrent Generative Adversarial Networks without Pre-training
Objective-Reinforced Generative Adversarial Networks (ORGAN) for Sequence Generation Models
Stabilizing Training of Generative Adversarial Networks through Regularization
SD-GAN: Semantically Decomposing the Latent Spaces of Generative Adversarial Networks
Accelerating Science with Generative Adversarial Networks: An Application to 3D Particle Showers in Multi-Layer Calorimeters
CycleGAN: Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks
Mastering Sketching: Adversarial Augmentation for Structured Prediction
I2T2I: Learning Text to Image Synthesis with Textual Data Augmentation
Improving Neural Machine Translation with Conditional Sequence Generative Adversarial Nets
Learning to Discover Cross-Domain Relations with Generative Adversarial Networks
ArtGAN: Artwork Synthesis with Conditional Categorical GANs
Learning from Simulated and Unsupervised Images through Adversarial Training
Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space
Pix2Pix: Image-to-Image Translation with Conditional Adversarial Networks
A Connection between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy-Based Models
Connecting Generative Adversarial Networks and Actor-Critic Methods
Neural Photo Editing with Introspective Adversarial Networks
SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient
Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets
Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks
Meta-Font, Metamathematics, and Metaphysics: Comments on Donald Knuth’s Article ‘The Concept of a Meta-Font’
Introducing AuraSR—An Open Reproduction of the GigaGAN Upscaler
TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up
Akanazawa/vgan: Code for Image Generation of Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow
Akanimax/Variational_Discriminator_Bottleneck: Implementation (with Some Experimentation) of the Paper Titled "Variational Discriminator Bottleneck"
MSG-GAN: Multi-Scale Gradients GAN (Architecture Inspired from ProGAN but Doesn’t Use Layer-Wise Growing)
GAN-QP: A Novel GAN Framework without Gradient Vanishing and Lipschitz Constraint
IntroVAE: A PyTorch Implementation of Paper ‘IntroVAE: Introspective Variational Autoencoders for Photographic Image Synthesis’
Twin-GAN: Unpaired Cross-Domain Image Translation With Weight-Sharing GANs
Junyanz/CycleGAN: Software That Can Generate Photos from Paintings, Turn Horses into Zebras, Perform Style Transfer, and More.
Kevinlyu/DCGAN_Pytorch: DCGAN With Vanilla GAN and Least Square GAN Objective
Nolan-Dev/GANInterface: Tool to Interface With a StyleGAN Model
Learning to Simulate Dynamic Environments With GameGAN (CVPR 2020)
A Good Image Generator Is What You Need for High-Resolution Video Synthesis
Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks
Case Study: Interpreting, Manipulating, and Controlling CLIP With Sparse Autoencoders
Steganography and the CycleGAN—Alignment Failure Case Study
Welcome to Simulation City, the Virtual World Where Waymo Tests Its Autonomous Vehicles
Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow [Homepage]
2023-begus-figure2-causaldisentanglementwithextremevaluesbysamplingextremeganlatentstointerpret.png
2023-xu-figure1-imagesamplesfromufogendiffusionganmodel.png
2023-xu-figure3-ufogenganfinetuningofdiffusionmodelschematictrainingillustration.png
2020-06-11-gwern-danbooru2019-palms-upscaledrealhandsamples.jpg
2020-05-30-gwern-danbooru2019-figures-randomsamples-40.jpg
2020-anokhin-figure2-schematicarchitectureofconditionallyindependentpixelsynthesisgangenerativemodel.png
2019-09-13-gwern-sagantensorflow-asuka-epoch29minibatch3000.jpg
2019-03-23-gwern-danbooru2018-sfw-512px-trainingsamples.jpg
2018-08-18-gwern-sagantensorflow-wholeasuka-epoch26minibatch4500.png
2018-07-18-gwern-128px-sagantensorflow-wholeasuka-trainingmontage.mp4
2018-mccandlish-openai-howaitrainingscales-gradientnoisescale-paretofrontier.svg
https://paperswithcode.com/sota/text-to-image-generation-on-coco
https://research.google/blog/mobilediffusion-rapid-text-to-image-generation-on-device/
https://research.google/blog/toward-generalized-sim-to-real-transfer-for-robot-learning/
https://towardsdatascience.com/african-masks-gans-tpu-9a6b0cf3105c
https://www.maskaravivek.com/post/gan-synthetic-data-generation/
https://x.com/search?f=tweets&vertical=default&q=BigGAN&src=typd
https%253A%252F%252Farxiv.org%252Fabs%252F2409.16211%2523bytedance.html
https%253A%252F%252Farxiv.org%252Fabs%252F2311.17042%2523stability.html
UFOGen: You Forward Once Large Scale Text-to-Image Generation via Diffusion GANs
https%253A%252F%252Farxiv.org%252Fabs%252F2311.09257%2523google.html
Consistency Trajectory Models (CTM): Learning Probability Flow ODE Trajectory of Diffusion
https%253A%252F%252Farxiv.org%252Fabs%252F2310.02279%2523sony.html
StyleTTS 2: Towards Human-Level Text-to-Speech through Style Diffusion and Adversarial Training with Large Speech Language Models
TANGO: Text-to-Audio Generation using Instruction-Tuned LLM and Latent Diffusion Model
Approaching an unknown communication system by latent space exploration and causal inference
https%253A%252F%252Farxiv.org%252Fabs%252F2303.05511%2523adobe.html
A survey on text generation using generative adversarial networks
https%253A%252F%252Farxiv.org%252Fabs%252F2210.13438%2523facebook.html
Using Generative Adversarial Networks for Conditional Creation of Anime Posters
TATS: Long Video Generation with Time-Agnostic VQGAN and Time-Sensitive Transformer
https%253A%252F%252Farxiv.org%252Fabs%252F2204.03638%2523facebook.html
https%253A%252F%252Farxiv.org%252Fabs%252F2110.04627%2523google.html
https%253A%252F%252Farxiv.org%252Fabs%252F2106.07631%2523google.html
https%253A%252F%252Farxiv.org%252Fabs%252F2104.07636%2523google.html
https%253A%252F%252Farxiv.org%252Fabs%252F2102.09672%2523openai.html
XMC-GAN: Cross-Modal Contrastive Learning for Text-to-Image Generation
https%253A%252F%252Farxiv.org%252Fabs%252F2101.04702%2523google.html
Image Generators with Conditionally-Independent Pixel Synthesis
https%253A%252F%252Fgithub.com%252Farfafax%252FE621-Face-Dataset.html
Parallel WaveGAN: A fast waveform generation model based on generative adversarial networks with multi-resolution spectrogram
https%253A%252F%252Farxiv.org%252Fabs%252F1910.11480%2523naver.html
https%253A%252F%252Fopenai.com%252Fresearch%252Fhow-ai-training-scales.html
CycleGAN: Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks
https%253A%252F%252Farxiv.org%252Fabs%252F1703.10593%2523bair.html
https%253A%252F%252Farxiv.org%252Fpdf%252F1606.03498%2523page%253D3%2526org%253Dopenai.html
Meta-Font, Metamathematics, and Metaphysics: Comments on Donald Knuth’s Article ‘The Concept of a Meta-Font’
%252Fdoc%252Fdesign%252Ftypography%252F1982-hofstadter.pdf.html
Wikipedia Bibliography: