This is an automated archive made by the Lemmit Bot.
The original was posted on /r/stablediffusion by /u/ninjasaid13 on 2024-10-15 05:42:07+00:00.
Disclaimer: I am not the author.
Paper:
Code:
Weights:
Abstract
Recently, large-scale diffusion models have made impressive progress in text-to-image (T2I) generation. To further equip these T2I models with fine-grained spatial control, approaches like ControlNet introduce an extra network that learns to follow a condition image. However, for every single condition type, ControlNet requires independent training on millions of data pairs with hundreds of GPU hours, which is quite expensive and makes it challenging for ordinary users to explore and develop new types of conditions. To address this problem, we propose the CtrLoRA framework, which trains a Base ControlNet to learn the common knowledge of image-to-image generation from multiple base conditions, along with condition-specific LoRAs to capture distinct characteristics of each condition. Utilizing our pretrained Base ControlNet, users can easily adapt it to new conditions, requiring as few as 1,000 data pairs and less than one hour of single-GPU training to obtain satisfactory results in most scenarios. Moreover, our CtrLoRA reduces the learnable parameters by 90% compared to ControlNet, significantly lowering the threshold to distribute and deploy the model weights. Extensive experiments on various types of conditions demonstrate the efficiency and effectiveness of our method. Codes and model weights will be released at this https URL.
This paper presents CtrLoRA, a framework for creating controllable image generation models with minimal data and resources. It trains a Base ControlNet with condition-specific LoRAs, then adapts it to new conditions with additional LoRAs, reducing data needs and speeding up training. The models can be easily integrated with community models for multi-condition generation without extra training. A common issue with CtrlLoRA found is that color-related tasks, like Palette and Lineart, converge more slowly than spatial tasks, likely due to network architecture limitations. Future improvements may come from using more advanced transformer diffusion backbones like Stable Diffusion V3 and Flux.1.