From 6487dd749190580662910148f8946be14c4daa42 Mon Sep 17 00:00:00 2001 From: Anna Date: Wed, 4 Jan 2023 17:37:14 +0100 Subject: [PATCH] typo --- cherry_picknick.py | 2 +- example1_standard.py | 7 +++---- example3_multitrans.py | 4 ++-- 3 files changed, 6 insertions(+), 7 deletions(-) diff --git a/cherry_picknick.py b/cherry_picknick.py index b3b037f..ec2c9f5 100644 --- a/cherry_picknick.py +++ b/cherry_picknick.py @@ -80,7 +80,7 @@ for i in range(10): # seed0 = 629575320 lb = LatentBlending(sdh) - lb.autosetup_branching(quality='medium', deepth_strength=0.65) + lb.autosetup_branching(quality='medium', depth_strength=0.65) prompt1 = "photo of a futuristic alien temple in a desert, mystic, glowing, organic, intricate, sci-fi movie, mesmerizing, scary" lb.set_prompt1(prompt1) diff --git a/example1_standard.py b/example1_standard.py index 83a40aa..ea042e2 100644 --- a/example1_standard.py +++ b/example1_standard.py @@ -22,7 +22,7 @@ import warnings import torch from tqdm.auto import tqdm from PIL import Image -import matplotlib.pyplot as plt +# import matplotlib.pyplot as plt import torch from movie_util import MovieSaver from typing import Callable, List, Optional, Union @@ -40,7 +40,7 @@ sdh = StableDiffusionHolder(fp_ckpt, fp_config, device) #%% Next let's set up all parameters quality = 'medium' -deepth_strength = 0.65 # Specifies how deep (in terms of diffusion iterations the first branching happens) +depth_strength = 0.65 # Specifies how deep (in terms of diffusion iterations the first branching happens) fixed_seeds = [69731932, 504430820] prompt1 = "photo of a beautiful cherry forest covered in white flowers, ambient light, very detailed, magic" @@ -51,14 +51,13 @@ fps = 30 # Spawn latent blending lb = LatentBlending(sdh) -lb.autosetup_branching(quality=quality, deepth_strength=deepth_strength) +lb.autosetup_branching(quality=quality, depth_strength=depth_strength) lb.set_prompt1(prompt1) lb.set_prompt2(prompt2) # Run latent blending imgs_transition = lb.run_transition(fixed_seeds=fixed_seeds) - # Let's get more cheap frames via linear interpolation (duration_transition*fps frames) imgs_transition_ext = add_frames_linear_interp(imgs_transition, duration_transition, fps) diff --git a/example3_multitrans.py b/example3_multitrans.py index c87b3b7..28499b7 100644 --- a/example3_multitrans.py +++ b/example3_multitrans.py @@ -41,7 +41,7 @@ sdh = StableDiffusionHolder(fp_ckpt, fp_config, device) fps = 30 duration_single_trans = 15 quality = 'medium' -deepth_strength = 0.55 #Specifies how deep (in terms of diffusion iterations the first branching happens) +depth_strength = 0.55 #Specifies how deep (in terms of diffusion iterations the first branching happens) # Specify a list of prompts below list_prompts = [] @@ -56,7 +56,7 @@ list_prompts.append("statue of an ancient cybernetic messenger annoucing good ne list_seeds = [954375479, 332539350, 956051013, 408831845, 250009012, 675588737] lb = LatentBlending(sdh) -lb.autosetup_branching(quality=quality, deepth_strength=deepth_strength) +lb.autosetup_branching(quality=quality, depth_strength=depth_strength) fp_movie = "movie_example3.mp4"