| | |
| | from diffusers import StableDiffusionControlNetPipeline, ControlNetModel |
| | import requests |
| | import torch |
| | from PIL import Image |
| | from io import BytesIO |
| |
|
| | url = "https://raw.githubusercontent.com/CompVis/stable-diffusion/main/assets/stable-samples/img2img/sketch-mountains-input.jpg" |
| |
|
| | response = requests.get(url) |
| | init_image = Image.open(BytesIO(response.content)).convert("RGB") |
| | init_image = init_image.resize((512, 512)) |
| |
|
| | path = "runwayml/stable-diffusion-v1-5" |
| |
|
| | run_compile = True |
| | controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16) |
| | pipe = StableDiffusionControlNetPipeline.from_pretrained( |
| | path, controlnet=controlnet, torch_dtype=torch.float16 |
| | ) |
| |
|
| | pipe = pipe.to("cuda:0") |
| | pipe.unet.to(memory_format=torch.channels_last) |
| | pipe.controlnet.to(memory_format=torch.channels_last) |
| |
|
| | if run_compile: |
| | print("Run torch compile") |
| | pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True) |
| | pipe.controlnet = torch.compile(pipe.controlnet, mode="reduce-overhead", fullgraph=True) |
| |
|
| | prompt = "ghibli style, a fantasy landscape with castles" |
| |
|
| | for _ in range(3): |
| | image = pipe(prompt=prompt, image=init_image).images[0] |
| |
|