ControlNeXt: Powerful and Efficient Control for Image and Video Generation
Manage episode 435200976 series 2954468
内容由Rob提供。所有播客内容(包括剧集、图形和播客描述)均由 Rob 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal。
Diffusion models have demonstrated remarkable and robust abilities in both image and video generation. To achieve greater control over generated results, researchers introduce additional architectures, such as ControlNet, Adapters and ReferenceNet, to integrate conditioning controls. However, current controllable generation methods often require substantial additional computational resources, especially for video generation, and face challenges in training or exhibit weak control. In this paper, we propose ControlNeXt: a powerful and efficient method for controllable image and video generation. We first design a more straightforward and efficient architecture, replacing heavy additional branches with minimal additional cost compared to the base model. Such a concise structure also allows our method to seamlessly integrate with other LoRA weights, enabling style alteration without the need for additional training. As for training, we reduce up to 90% of learnable parameters compared to the alternatives. Furthermore, we propose another method called Cross Normalization (CN) as a replacement for Zero-Convolution' to achieve fast and stable training convergence. We have conducted various experiments with different base models across images and videos, demonstrating the robustness of our method.
2024: Bohao Peng, Jian Wang, Yuechen Zhang, Wenbo Li, Mingchang Yang, Jiaya Jia
https://arxiv.org/pdf/2408.06070
…
continue reading
2024: Bohao Peng, Jian Wang, Yuechen Zhang, Wenbo Li, Mingchang Yang, Jiaya Jia
https://arxiv.org/pdf/2408.06070
298集单集