Dale Decatur's Avatar

Dale Decatur

@daledecatur

CS PhD student @ UChicago https://ddecatur.github.io/

9
Followers
20
Following
7
Posts
21.11.2024
Joined
Posts Following

Latest posts by Dale Decatur @daledecatur

Preview
Reusing Computation in Text-to-Image Diffusion for Efficient Generation of Image Sets Text-to-image diffusion models enable high-quality image generation but are computationally expensive. While prior work optimizes per-inference efficiency, we explore an orthogonal approach: reducing ...

This work was completed in collaboration with @thibaultgroueix.bsky.social, Yifan Wang, @ranahanocka.bsky.social, @vovakim.bsky.social, and @gadelha.bsky.social‬. Check out our #ICCV2025 poster #153 today during Poster Session #4 from 2:45-4:45 HST!

Paper: arxiv.org/abs/2508.21032

7/

22.10.2025 20:22 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

While our method can be applied to general purpose image generation, our method achieves the most dramatic savings (saves >75% vs standard diffusion) when examples are structurally similar. Some applications of this are style variation, subject variation, and virtual try-on.

6/

22.10.2025 20:22 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Interestingly, we observe that models trained using a text-to-image prior (bottom) generate high frequency details much later in the denoising process than without (top). This makes them ideal for sharing compute with our approach!

5/

22.10.2025 20:22 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image Post image

Compared to standard diffusion (left), our method (right) generates images of comparable quality using a fraction of the compute. Exact savings depend on the prompt set, but we show that our method can save up to 74% of the total denoising steps required for standard diffusion!

4/

22.10.2025 20:22 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We construct a tree by hierarchically clustering prompts. We then map each denoising step k to a height in this tree, using the mean embedding of each cluster at this height as the condition. The steps gradually diverge from shared embeddings to individual prompt embeddings.

3/

22.10.2025 20:22 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We take advantage of the coarse-to-fine nature of diffusion generation: early timesteps generate low frequency structure and later timesteps produce high frequency details. Leveraging this, we share intermediate denoising results at early steps between similar examples.

2/

22.10.2025 20:22 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

Excited to share our #ICCV2025 work Reusing Computation in Text-to-Image Diffusion for Efficient Generation of Image Sets!

Our method generates large sets of images using significantly less compute than standard diffusion.

πŸ“Žhttps://ddecatur.github.io/hierarchical-diffusion/

1/

22.10.2025 20:22 πŸ‘ 8 πŸ” 2 πŸ’¬ 1 πŸ“Œ 2