1
0
Fork 0
onnx-web/api/onnx_web/chain/blend_img2img.py

120 lines
3.5 KiB
Python
Raw Normal View History

2023-01-28 23:09:19 +00:00
from logging import getLogger
from typing import List, Optional
import numpy as np
2023-02-05 23:55:04 +00:00
import torch
2023-02-05 13:53:26 +00:00
from PIL import Image
2023-03-05 04:25:49 +00:00
from ..diffusers.load import load_pipeline
from ..diffusers.utils import encode_prompt, parse_prompt, slice_prompt
from ..params import ImageParams, SizeChart, StageParams
2023-02-26 05:49:39 +00:00
from ..server import ServerContext
2023-02-26 20:15:30 +00:00
from ..worker import ProgressCallback, WorkerContext
from .stage import BaseStage
2023-01-28 23:09:19 +00:00
logger = getLogger(__name__)
class BlendImg2ImgStage(BaseStage):
max_tile = SizeChart.unlimited
def run(
self,
worker: WorkerContext,
server: ServerContext,
_stage: StageParams,
params: ImageParams,
sources: List[Image.Image],
*,
strength: float,
callback: Optional[ProgressCallback] = None,
stage_source: Optional[Image.Image] = None,
prompt_index: Optional[int] = None,
**kwargs,
) -> List[Image.Image]:
params = params.with_args(**kwargs)
# multi-stage prompting
if prompt_index is not None:
params = params.with_args(prompt=slice_prompt(params.prompt, prompt_index))
logger.info(
"blending image using img2img, %s steps: %s", params.steps, params.prompt
)
prompt_pairs, loras, inversions, (prompt, negative_prompt) = parse_prompt(
params
)
pipe_type = params.get_valid_pipeline("img2img")
pipe = load_pipeline(
server,
params,
pipe_type,
worker.get_device(),
inversions=inversions,
loras=loras,
)
pipe_params = {}
2023-08-26 04:33:41 +00:00
if params.is_pix2pix():
pipe_params["image_guidance_scale"] = strength
elif params.is_control():
pipe_params["controlnet_conditioning_scale"] = strength
2023-08-26 04:33:41 +00:00
else:
pipe_params["strength"] = strength
outputs = []
for source in sources:
2023-08-21 03:28:08 +00:00
if params.is_lpw():
logger.debug("using LPW pipeline for img2img")
rng = torch.manual_seed(params.seed)
result = pipe.img2img(
source,
prompt,
generator=rng,
guidance_scale=params.cfg,
negative_prompt=negative_prompt,
num_inference_steps=params.steps,
callback=callback,
**pipe_params,
)
else:
# encode and record alternative prompts outside of LPW
prompt_embeds = encode_prompt(
pipe, prompt_pairs, params.batch, params.do_cfg()
)
2023-08-21 03:28:08 +00:00
if not params.is_xl():
2023-08-21 03:28:08 +00:00
pipe.unet.set_prompts(prompt_embeds)
rng = np.random.RandomState(params.seed)
result = pipe(
prompt,
generator=rng,
guidance_scale=params.cfg,
image=source,
negative_prompt=negative_prompt,
num_inference_steps=params.steps,
callback=callback,
**pipe_params,
)
outputs.extend(result.images)
return outputs
def steps(
self,
params: ImageParams,
*args,
) -> int:
return params.steps # TODO: multiply by strength
def outputs(
self,
params: ImageParams,
sources: int,
) -> int:
return sources + 1