diff --git a/gradio_ui.py b/gradio_ui.py index 390ffaf..29c5635 100644 --- a/gradio_ui.py +++ b/gradio_ui.py @@ -33,6 +33,13 @@ import gradio as gr import copy +""" +try this: + button variant 'primary' for main call-to-action, 'secondary' for a more subdued style + gr.Column(scale=1, min_width=600): + +""" + #%% class BlendingFrontend(): @@ -150,10 +157,7 @@ class BlendingFrontend(): return seed - def downscale_imgs(self, list_imgs): - return [l.resize((self.max_size_imgs, self.max_size_imgs)) for l in list_imgs] - - def run(self, x): + def run(self): print("STARTING DIFFUSION!") self.state_prev = self.state_current.copy() self.state_current = self.get_state_dict() @@ -188,55 +192,38 @@ class BlendingFrontend(): fixed_seeds = [self.seed1, self.seed2] imgs_transition = self.lb.run_transition(fixed_seeds=fixed_seeds) - imgs_transition = [Image.fromarray(l) for l in imgs_transition] + print(f"DONE DIFFUSION! Resulted in {len(imgs_transition)} images") assert np.mod((self.nmb_branches_final-self.nmb_imgs_show)/4, 1)==0, 'self.nmb_branches_final illegal value!' idx_list = np.linspace(0, self.nmb_branches_final-1, self.nmb_imgs_show).astype(np.int32) - list_imgs = [] + list_imgs_preview = [] for j in idx_list: - list_imgs.append(imgs_transition[j]) + list_imgs_preview.append(Image.fromarray(imgs_transition[j])) - # list_imgs = self.downscale_imgs(list_imgs) - self.imgs_show_current = copy.deepcopy(list_imgs) - # Save as jpgs on disk so we are not sending umcompressed data around + timestamp = get_time('second') list_fp_imgs = [] - for i in range(len(list_imgs)): - fp_img = f"img_preview_{i}.jpg" - list_imgs[i].save(fp_img) + for i in range(len(list_imgs_preview)): + fp_img = f"img_preview_{i}_{timestamp}.jpg" + list_imgs_preview[i].save(fp_img) list_fp_imgs.append(fp_img) - return list_fp_imgs - + # Save the movie as well + imgs_transition_ext = add_frames_linear_interp(imgs_transition, self.duration, self.fps) - - def save(self): - if self.lb.tree_final_imgs[0] is None: - return - print("save is called!") - imgs_transition = self.lb.tree_final_imgs - - if False: - # skip for now: writing images. - dp_img = "/" - self.lb.write_imgs_transition(dp_img, imgs_transition) - - - fps = self.fps - # Let's get more cheap frames via linear interpolation (duration_transition*fps frames) - imgs_transition_ext = add_frames_linear_interp(imgs_transition, self.duration, fps) - - # Save as MP4 - fp_movie = f"movie_{get_time('second')}.mp4" + # Save as movie + fp_movie = f"movie_{timestamp}.mp4" if os.path.isfile(fp_movie): os.remove(fp_movie) - ms = MovieSaver(fp_movie, fps=fps) + ms = MovieSaver(fp_movie, fps=self.fps) for img in tqdm(imgs_transition_ext): ms.write_frame(img) ms.finalize() - return fp_movie - + print("DONE SAVING MOVIE! SENDING BACK...") + list_return = list_fp_imgs + [fp_movie] + return list_return + def get_state_dict(self): @@ -252,9 +239,10 @@ class BlendingFrontend(): if __name__ == "__main__": - fp_ckpt = "../stable_diffusion_models/ckpt/v2-1_512-ema-pruned.ckpt" - sdh = StableDiffusionHolder(fp_ckpt) - self = BlendingFrontend(sdh) + fp_ckpt = "../stable_diffusion_models/ckpt/v2-1_768-ema-pruned.ckpt" + # fp_ckpt = "../stable_diffusion_models/ckpt/v2-1_512-ema-pruned.ckpt" + # sdh = StableDiffusionHolder(fp_ckpt) + self = BlendingFrontend(None) with gr.Blocks() as demo: @@ -275,17 +263,17 @@ if __name__ == "__main__": with gr.Row(): depth_strength = gr.Slider(0.01, 0.99, self.depth_strength, step=0.01, label='depth_strength', interactive=True) + duration = gr.Slider(0.1, 30, self.duration, step=0.1, label='video duration', interactive=True) guidance_scale_mid_damper = gr.Slider(0.01, 2.0, self.guidance_scale_mid_damper, step=0.01, label='guidance_scale_mid_damper', interactive=True) - mid_compression_scaler = gr.Slider(1.0, 2.0, self.mid_compression_scaler, step=0.01, label='mid_compression_scaler', interactive=True) with gr.Row(): - b_newseed1 = gr.Button("rand seed 1") + b_run = gr.Button('COMPUTE!', variant='primary') seed1 = gr.Number(42, label="seed 1", interactive=True) - b_newseed2 = gr.Button("rand seed 2") seed2 = gr.Number(420, label="seed 2", interactive=True) - with gr.Row(): - b_run = gr.Button('step1: run preview!') + with gr.Column(): + b_newseed1 = gr.Button("randomize \nseed 1", variant='secondary') + b_newseed2 = gr.Button("randomize \nseed 2", variant='secondary') with gr.Row(): img1 = gr.Image(label="1/5") @@ -295,14 +283,8 @@ if __name__ == "__main__": img5 = gr.Image(label="5/5") with gr.Row(): - b_save = gr.Button('step2: render video') vid = gr.Video() - with gr.Row(): - duration = gr.Slider(0.1, 30, self.duration, step=0.1, label='duration', interactive=True) - fps = gr.Slider(1, 120, self.fps, step=1, label='fps', interactive=True) - - # Bind the on-change methods depth_strength.change(fn=self.change_depth_strength, inputs=depth_strength) num_inference_steps.change(fn=self.change_num_inference_steps, inputs=num_inference_steps) @@ -310,7 +292,6 @@ if __name__ == "__main__": guidance_scale.change(fn=self.change_guidance_scale, inputs=guidance_scale) guidance_scale_mid_damper.change(fn=self.change_guidance_scale_mid_damper, inputs=guidance_scale_mid_damper) - mid_compression_scaler.change(fn=self.change_mid_compression_scaler, inputs=mid_compression_scaler) height.change(fn=self.change_height, inputs=height) width.change(fn=self.change_width, inputs=width) @@ -319,13 +300,11 @@ if __name__ == "__main__": negative_prompt.change(fn=self.change_negative_prompt, inputs=negative_prompt) seed1.change(fn=self.change_seed1, inputs=seed1) seed2.change(fn=self.change_seed2, inputs=seed2) - fps.change(fn=self.change_fps, inputs=fps) duration.change(fn=self.change_duration, inputs=duration) branch1_influence.change(fn=self.change_branch1_influence, inputs=branch1_influence) b_newseed1.click(self.randomize_seed1, outputs=seed1) b_newseed2.click(self.randomize_seed2, outputs=seed2) - b_run.click(self.run, outputs=[img1, img2, img3, img4, img5]) - b_save.click(self.save, outputs=vid) + b_run.click(self.run, outputs=[img1, img2, img3, img4, img5, vid]) demo.launch(share=self.share, inbrowser=True, inline=False)