Create butter-smooth transitions between prompts, powered by stable diffusion
Go to file
Johannes Stelzer 465eb642d6 gradio upd 2023-02-23 10:32:53 +01:00
configs Add config for SD 1.5 2023-01-14 21:04:35 +01:00
ldm sd 2.1 2022-12-09 11:06:44 +00:00
.gitignore mp4 files 2022-11-28 08:45:40 +01:00
LICENSE Initial commit 2022-11-19 19:40:58 +01:00
README.md Update README.md 2023-02-22 08:22:10 +01:00
animation.gif new graphics 2023-02-20 11:21:16 +01:00
example1.jpg example imgs 2023-01-09 08:17:07 +01:00
example1_standard.py cleanup 2023-02-22 10:15:03 +01:00
example2_multitrans.py cleanup 2023-02-22 10:15:03 +01:00
example3.jpg fix 2023-02-20 07:22:01 +01:00
example3_upscaling.py cleanup 2023-02-22 10:15:03 +01:00
example4_multitrans_upscaling.py cleanup 2023-02-22 10:15:03 +01:00
gradio_ui.py gradio upd 2023-02-23 10:32:53 +01:00
latent_blending.ipynb byo-model 2023-01-20 16:12:34 +01:00
latent_blending.py cleanup 2023-02-22 10:15:03 +01:00
movie_util.py cleanup 2023-02-22 10:15:03 +01:00
parameters.md helper functions for crossfeeding 2023-02-20 11:44:50 +01:00
requirements.txt typo 2023-02-21 09:34:22 +01:00
stable_diffusion_holder.py cleanup 2023-02-22 10:15:03 +01:00
utils.py cleanup 2023-02-22 10:15:03 +01:00

README.md

Quickstart

Latent blending enables video transitions with incredible smoothness between prompts, computed within seconds. Powered by stable diffusion 2.1, this method involves specific mixing of intermediate latent representations to create a seamless transition with users having the option to fully customize the transition and run high-resolution upscaling.

Open In Colab

fp_ckpt = hf_hub_download(repo_id="stabilityai/stable-diffusion-2-1-base", filename="v2-1_512-ema-pruned.ckpt")

sdh = StableDiffusionHolder(fp_ckpt)
lb = LatentBlending(sdh)

lb.set_prompt1('photo of my first prompt1')
lb.set_prompt2('photo of my second prompt')
depth_strength = 0.6 # How deep the first branching happens
t_compute_max_allowed = 10 # How much compute time we give to the transition
imgs_transition = lb.run_transition(depth_strength=depth_strength, t_compute_max_allowed=t_compute_max_allowed)

Gradio UI

To run the UI on your local machine, run gradio_ui.py If you want to specify the output directory, you can create a .env file in the latentblending git directory. In here, specify:

DIR_OUT="SET_PATH_HERE"

Example 1: Simple transition

To run a simple transition between two prompts, run example1_standard.py

Example 2: Multi transition

To run multiple transition between K prompts, resulting in a stitched video, run example2_multitrans.py. View a longer example video here.

Example 3: High-resolution with upscaling

You can run a high-res transition using the x4 upscaling model in a two-stage procedure, see example3_upscaling.py. View as video here.

Example 4: Multi transition with high-resolution with upscaling

You can run a multi transition movie and upscale it, see example4_multitrans_upscaling.py.

Customization

Most relevant parameters

You can find the most relevant parameters here.

Change the height/width

lb.set_height(512)
lb.set_width(1024)

Change guidance scale

lb.set_guidance_scale(5.0)

run_transition parameters

  • num_inference_steps: number of diffusions steps.Number of diffusion steps. Higher values will take more compute time.
  • depth_strength: The strength of the diffusion iterations determines when the blending process will begin. A value close to zero results in more creative and intricate outcomes, while a value closer to one indicates a simpler alpha blending. However, low values may also bring about the introduction of additional objects and motion.
  • t_compute_max_allowed: maximum time allowed for computation. Higher values give better results but take longer. Either provide t_compute_max_allowed or nmb_max_branches.
  • nmb_max_branches: The maximum number of branches to be computed. Higher values give better results. Use this if you want to have controllable results independent of your hardware. Either provide t_compute_max_allowed or nmb_max_branches.

Crossfeeding to the last image.

Cross-feeding latents is a key feature of latent blending. Here, you can set how much the first image branch influences the very last one. In the animation below, these are the blue arrows.

crossfeed_power = 0.5 # 50% of the latents in the last branch are copied from branch1
crossfeed_range = 0.7 # The crossfeed is active until 70% of num_iteration, then switched off
crossfeed_decay = 0.2 # The power of the crossfeed decreases over diffusion iterations, here it would be 0.5*0.2=0.1 in the end of the range.
lb.set_branch1_crossfeed(crossfeed_power, crossfeed_range, crossfeed_decay)

Crossfeeding to all transition images

Here, you can set how much the parent branches influence the mixed one. In the animation below, these are the yellow arrows.

crossfeed_power = 0.5 # 50% of the latents in the last branch are copied from the parents
crossfeed_range = 0.7 # The crossfeed is active until 70% of num_iteration, then switched off
crossfeed_decay = 0.2 # The power of the crossfeed decreases over diffusion iterations, here it would be 0.5*0.2=0.1 in the end of the range.
lb.set_parental_crossfeed(crossfeed_power, crossfeed_range, crossfeed_decay)

Installation

Packages

pip install -r requirements.txt

With xformers, stable diffusion will run faster with smaller memory inprint. Necessary for higher resolutions / upscaling model.

conda install xformers -c xformers/label/dev

Alternatively, you can build it from source:

# (Optional) Makes the build much faster
pip install ninja
# Set TORCH_CUDA_ARCH_LIST if running and building on different GPU types
pip install -v -U git+https://github.com/facebookresearch/xformers.git@main#egg=xformers
# (this can take dozens of minutes)

How does latent blending work?

Method

In the figure above, a diffusion tree is illustrated. The diffusion steps are represented on the y-axis, with temporal blending on the x-axis. The diffusion trajectory for the first prompt is the most left column, which is always computed first. Next, the the trajectory for the second prompt is computed, which may be influenced by the first branch (blue arrows, for a description see above at Crossfeeding to the last image.). Finally, all transition images in between are computed. For the transition, there can be an influence of the parents, which helps preserving structures (yellow arrows, for a description see above at Crossfeeding to all transition images). Importantly, the place of injection on the x-axis is not hardfixes a priori, but set dynamically using Perceptual Similarity, always adding a branch where it is needed most.

The concrete parameters for the transition above would be:

lb.set_branch1_crossfeed(crossfeed_power=0.8, crossfeed_range=0.6, crossfeed_decay=0.4)
lb.set_parental_crossfeed(crossfeed_power=0.8, crossfeed_range=0.8, crossfeed_decay=0.2)
imgs_transition = lb.run_transition(num_inference_steps=10, depth_strength=0.2, nmb_max_branches=7)

Perceptual aspects

With latent blending, we can create transitions that appear to defy the laws of nature, yet appear completely natural and believable. The key is to surpress processing in our dorsal visual stream, which is achieved by avoiding motion in the transition. Without motion, our visual system has difficulties detecting the transition, leaving viewers with the illusion of a single, continuous image, see change blindness. However, when motion is introduced, the visual system can detect the transition and the viewer becomes aware of the transition, leading to a jarring effect. Therefore, best results will be achieved when optimizing the transition parameters, particularly the crossfeeding parameters and the depth of the first injection.

Changelog

  • New blending engine with cross-feeding capabilities, enabling structure preserving transitions
  • LPIPS image similarity for finding the next best injection branch, resulting in smoother transitions
  • Time-based computation: instead of specifying how many frames your transition has, you can tell your compute budget and get a transition within that budget.
  • New multi-movie engine
  • Simpler and more powerful gradio ui. You can iterate faster and stitch together a multi movie.
  • Inpaint support dropped (as it only makes sense for a single transition)

Coming soon...

  • Huggingface Space
  • More manipulations to the latent (translation, zoom, masking)
  • Transitions with Depth model

Stay tuned on twitter: @j_stelzer

Contact: stelzer@lunar-ring.ai (Johannes Stelzer)