fix(api): enable optimizations during model conversion
This commit is contained in:
parent
2edb3c6199
commit
fd013c88ef
|
@ -25,6 +25,8 @@ from diffusers import (
|
|||
from onnx import load, save_model
|
||||
from torch.onnx import export
|
||||
|
||||
from onnx_web.diffusion.load import optimize_pipeline
|
||||
|
||||
from ..diffusion.pipeline_onnx_stable_diffusion_upscale import (
|
||||
OnnxStableDiffusionUpscalePipeline,
|
||||
)
|
||||
|
@ -96,6 +98,8 @@ def convert_diffusion_stable(
|
|||
).to(ctx.training_device)
|
||||
output_path = Path(dest_path)
|
||||
|
||||
optimize_pipeline(ctx, pipeline)
|
||||
|
||||
# TEXT ENCODER
|
||||
num_tokens = pipeline.text_encoder.config.max_position_embeddings
|
||||
text_hidden_size = pipeline.text_encoder.config.hidden_size
|
||||
|
|
Loading…
Reference in New Issue