Abstract

We present ControlNet, a neural network architecture to add spatial conditioning controls to large, pretrained text-to-image diffusion models. ControlNet locks the production-ready large diffusion models, and reuses their deep and robust encoding layers pretrained with billions of images as a strong backbone to learn a diverse set of conditional controls. The neural architecture is connected with "zero convolutions"(zero-initialized convolution layers) that progressively grow the parameters from zero and ensure that no harmful noise could affect the finetuning. We test various conditioning controls, e.g., edges, depth, segmentation, human pose, etc., with Stable Diffusion, using single or multiple conditions, with or without prompts. We show that the training of ControlNets is robust with small (<50k) and large (>1m) datasets. Extensive results show that ControlNet may facilitate wider applications to control image diffusion models. © 2023 IEEE.

Keywords

Computer scienceConvolution (computer science)DiffusionEncoding (memory)Artificial intelligenceImage (mathematics)Noise (video)SegmentationConvolutional neural networkSet (abstract data type)Pattern recognition (psychology)Zero (linguistics)Artificial neural networkImage segmentationAlgorithmComputer visionPhysics

Affiliated Institutions

Related Publications

Publication Info

Year
2023
Type
article
Pages
3813-3824
Citations
2649
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2649
OpenAlex
881
Influential
2929
CrossRef

Cite This

Lvmin Zhang, Anyi Rao, Maneesh Agrawala (2023). Adding Conditional Control to Text-to-Image Diffusion Models. 2023 IEEE/CVF International Conference on Computer Vision (ICCV) , 3813-3824. https://doi.org/10.1109/iccv51070.2023.00355

Identifiers

DOI
10.1109/iccv51070.2023.00355
arXiv
2302.05543

Data Quality

Data completeness: 88%