.

Members-Only

Recent Talks & Demos are for members only

Exclusive feed

You must be an AI Tinkerers active member to view these talks and demos.

January 30, 2025 · Paris

ControlNets for Virtual Staging

Exploring custom controlnet training for virtual staging: methods, architecture variations, and comparative results on converting empty rooms into furnished scenes.

Video
Overview
Links
Tech stack
  • ControlNet
    ControlNet is a neural network architecture that adds precise spatial conditioning (e.g., Canny edges, OpenPose keypoints) to large, pretrained text-to-image diffusion models like Stable Diffusion.
    ControlNet is a game-changer for diffusion models, providing fine-grained control over image composition. It works by creating two copies of the model’s weights: a 'locked' copy to preserve the original Stable Diffusion capabilities and a 'trainable' copy for new conditions. This connection uses 'zero convolutions' to ensure no harmful noise affects the finetuning. This architecture allows training on small datasets (under 50k images) for tasks like pose-to-image or depth-to-image, efficiently steering the output with specific inputs like OpenPose keypoints or Canny edge maps.

Related projects