add note about onnx-fp16 opt
This commit is contained in:
parent
31e841ab4a
commit
d26416465b
|
@ -331,9 +331,15 @@ any new packages.
|
|||
```shell
|
||||
> git clone https://github.com/microsoft/onnxruntime
|
||||
> cd onnxruntime/onnxruntime/python/tools/transformers/models/stable_diffusion
|
||||
> python3 optimize_pipeline.py -i /home/ssube/onnx-web/models/stable-diffusion
|
||||
> python3 optimize_pipeline.py \
|
||||
-i /home/ssube/onnx-web/models/stable-diffusion-onnx-v1-5 \
|
||||
-o /home/ssube/onnx-web/models/stable-diffusion-optimized-v1-5 \
|
||||
--use_external_data_format
|
||||
```
|
||||
|
||||
You can convert the model into ONNX float16 using the `--float16` option. Models converted into ONNX float16 can only
|
||||
be run when using [the `onnx-fp16` optimization](server-admin#pipeeline-optimizations).
|
||||
|
||||
The `optimize_pipeline.py` script should work on any [diffusers directory with ONNX models](#converting-diffusers-models),
|
||||
but you will need to use the `--use_external_data_format` option if you are not using `--float16`. See the `--help` for
|
||||
more details.
|
||||
|
|
Loading…
Reference in New Issue