1
0
Fork 0
onnx-web/api/onnx_web/chain/source_txt2img.py

180 lines
5.5 KiB
Python
Raw Permalink Normal View History

2023-01-28 23:09:19 +00:00
from logging import getLogger
2023-11-19 00:13:13 +00:00
from typing import Optional, Tuple
import numpy as np
2023-02-05 23:55:04 +00:00
import torch
2023-02-05 13:53:26 +00:00
2023-12-03 17:11:23 +00:00
from ..constants import LATENT_FACTOR
from ..diffusers.load import load_pipeline
2023-07-10 22:48:26 +00:00
from ..diffusers.utils import (
encode_prompt,
get_latents_from_seed,
get_tile_latents,
parse_prompt,
2023-11-11 00:37:42 +00:00
parse_reseed,
slice_prompt,
2023-07-10 22:48:26 +00:00
)
from ..params import ImageParams, Size, SizeChart, StageParams
2023-02-26 05:49:39 +00:00
from ..server import ServerContext
2023-02-26 20:15:30 +00:00
from ..worker import ProgressCallback, WorkerContext
from .base import BaseStage
2024-01-03 03:49:22 +00:00
from .result import ImageMetadata, StageResult
2023-01-28 23:09:19 +00:00
logger = getLogger(__name__)
class SourceTxt2ImgStage(BaseStage):
2023-09-15 13:40:56 +00:00
max_tile = SizeChart.max
def run(
self,
worker: WorkerContext,
server: ServerContext,
2023-09-05 02:18:07 +00:00
stage: StageParams,
params: ImageParams,
sources: StageResult,
*,
dims: Tuple[int, int, int] = None,
size: Size,
callback: Optional[ProgressCallback] = None,
latents: Optional[np.ndarray] = None,
prompt_index: Optional[int] = None,
**kwargs,
2023-12-03 18:53:50 +00:00
) -> StageResult:
params = params.with_args(**kwargs)
size = size.with_args(**kwargs)
2023-07-09 05:02:27 +00:00
# multi-stage prompting
if prompt_index is not None:
params = params.with_args(prompt=slice_prompt(params.prompt, prompt_index))
logger.info(
2023-11-06 14:48:35 +00:00
"generating image using txt2img, %s steps of %s: %s",
params.steps,
params.model,
params.prompt,
2023-02-05 13:53:26 +00:00
)
if len(sources):
logger.info(
"source images were passed to a source stage, new images will be appended"
)
prompt_pairs, loras, inversions, (prompt, negative_prompt) = parse_prompt(
params
)
2023-07-09 05:02:27 +00:00
if params.is_panorama() or params.is_xl():
tile_size = max(stage.tile_size, params.unet_tile)
2023-09-05 02:18:07 +00:00
else:
tile_size = params.unet_tile
# this works for panorama as well, because tile_size is already max(tile_size, *size)
latent_size = size.min(tile_size, tile_size)
# generate new latents or slice existing
if latents is None:
latents = get_latents_from_seed(int(params.seed), latent_size, params.batch)
2023-07-09 05:02:27 +00:00
else:
latents = get_tile_latents(latents, int(params.seed), latent_size, dims)
2023-07-09 05:02:27 +00:00
2023-11-11 00:37:42 +00:00
# reseed latents as needed
reseed_rng = np.random.RandomState(params.seed)
2023-11-11 00:37:42 +00:00
prompt, reseed = parse_reseed(prompt)
for top, left, bottom, right, region_seed in reseed:
if region_seed == -1:
region_seed = reseed_rng.random_integers(2**32 - 1)
2023-11-11 00:37:42 +00:00
logger.debug(
"reseed latent region: [:, :, %s:%s, %s:%s] with %s",
top,
left,
bottom,
right,
region_seed,
)
latents[
:,
:,
top // LATENT_FACTOR : bottom // LATENT_FACTOR,
left // LATENT_FACTOR : right // LATENT_FACTOR,
] = get_latents_from_seed(
region_seed, Size(right - left, bottom - top), params.batch
)
2023-11-11 00:37:42 +00:00
pipe_type = params.get_valid_pipeline("txt2img")
pipe = load_pipeline(
server,
params,
pipe_type,
worker.get_device(),
embeddings=inversions,
loras=loras,
)
2023-08-21 03:28:08 +00:00
if params.is_lpw():
logger.debug("using LPW pipeline for txt2img")
rng = torch.manual_seed(params.seed)
2024-01-03 03:49:22 +00:00
output = pipe.text2img(
prompt,
height=latent_size.height,
width=latent_size.width,
generator=rng,
guidance_scale=params.cfg,
latents=latents,
negative_prompt=negative_prompt,
num_images_per_prompt=params.batch,
num_inference_steps=params.steps,
eta=params.eta,
callback=callback,
)
else:
# encode and record alternative prompts outside of LPW
if params.is_panorama() or params.is_xl():
2023-11-19 00:13:13 +00:00
logger.debug(
"prompt alternatives are not supported for panorama or SDXL"
)
else:
prompt_embeds = encode_prompt(
pipe, prompt_pairs, params.batch, params.do_cfg()
)
pipe.unet.set_prompts(prompt_embeds)
rng = np.random.RandomState(params.seed)
2024-01-03 03:49:22 +00:00
output = pipe(
prompt,
height=latent_size.height,
width=latent_size.width,
generator=rng,
guidance_scale=params.cfg,
2023-08-30 01:54:28 +00:00
latents=latents,
negative_prompt=negative_prompt,
num_images_per_prompt=params.batch,
num_inference_steps=params.steps,
eta=params.eta,
callback=callback,
)
2024-01-03 03:49:22 +00:00
result = StageResult(source=sources)
for image in output.images:
2024-01-03 04:14:21 +00:00
result.push_image(
image, ImageMetadata(params, size, inversions=inversions, loras=loras)
)
2024-01-03 03:49:22 +00:00
logger.debug("produced %s outputs", len(result))
return result
def steps(
self,
params: ImageParams,
size: Size,
) -> int:
return params.steps
def outputs(
self,
params: ImageParams,
sources: int,
) -> int:
2023-09-13 00:17:03 +00:00
return sources + 1