cancel
Showing results for 
Search instead for 
Did you mean: 

[How-To] Running Optimized Automatic1111 Stable Diffusion WebUI on AMD GPUs

adit_bhutani
Staff
Staff
11 0 112K

[UPDATE]: The Automatic1111-directML branch now supports Microsoft Olive under the Automatic1111 WebUI interface, which allows for generating optimized models and running them all under the Automatic1111 WebUI, without a separate branch needed to optimize for AMD platforms. The updated blog to run Stable Diffusion Automatic1111 with Olive Optimizations is available here - UPDATED HOW-TO GUIDE

Prepared by Hisham Chowdhury (AMD), Lucas Neves (AMD), and Justin Stoecker (Microsoft)

Did you know you can enable Stable Diffusion with Microsoft Olive under Automatic1111(Xformer) to get a significant speedup via Microsoft DirectML on Windows? Microsoft and AMD have been working together to optimize the Olive path on AMD hardware, accelerated via the Microsoft DirectML platform API and the AMD User Mode Driver’s ML (Machine Learning) layer for DirectML allowing users access to the power of the AMD GPU’s AI (Artificial Intelligence) capabilities.

adit_bhutani_0-1692393654030.png

 

1. Prerequisites

 

2. Overview of Microsoft Olive

Microsoft Olive is a Python tool that can be used to convert, optimize, quantize, and auto-tune models for optimal inference performance with ONNX Runtime execution providers like DirectML. Olive greatly simplifies model processing by providing a single toolchain to compose optimization techniques, which is especially important with more complex models like Stable Diffusion that are sensitive to the ordering of optimization techniques. The DirectML sample for Stable Diffusion applies the following techniques:

  • Model conversion: translates the base models from PyTorch to ONNX.
  • Transformer graph optimization: fuses subgraphs into multi-head attention operators and eliminating inefficient from conversion.
  • Quantization: converts most layers from FP32 to FP16 to reduce the model's GPU memory footprint and improve performance.

Combined, the above optimizations enable DirectML to leverage AMD GPUs for greatly improved performance when performing inference with transformer models like Stable Diffusion.

 

3. Generate Optimized Stable Diffusion Models using Microsoft Olive

Create Optimized Model

(Following the instruction from Olive, we can generate optimized Stable Diffusion model using Olive)

  1. Open Anaconda/Miniconda Terminal
  2. Create a new environment by sequentially entering the following commands into the terminal, followed by the enter key. Important to note that Python 3.9 is required.
    • conda create --name olive python=3.9
    • conda activate olive
    • pip install olive-ai[directml]==0.2.1
    • git clone https://github.com/microsoft/olive --branch v0.2.1
    • cd olive\examples\directml\stable_diffusion
    • pip install -r requirements.txt
    • pip install pydantic==1.10.12
  3. Generate an ONNX model and optimize it for run-time. This may take a long time.
    • python stable_diffusion.py --optimize

The optimized model will be stored at the following directory, keep this open for later: olive\examples\directml\stable_diffusion\models\optimized\runwayml. The model folder will be called “stable-diffusion-v1-5”. Use the following command to see what other models are supported: python stable_diffusion.py –help

To Test the Optimized Model

  1. To test the optimized model, run the following command: 
    • python stable_diffusion.py --interactive --num_images 2

adit_bhutani_0-1692383483523.png

 

4. Install and Run Automatic1111 Stable Diffusion WebUI

Following the instructions here, install Automatic1111 Stable Diffusion WebUI without the optimized model. It will be using the default unoptimized PyTorch path. Enter the following commands sequentially into a new terminal window.

  1. Open Anaconda/Miniconda Terminal.
  2. Enter the following commands in the terminal, followed by the enter key, to install Automatic1111 WebUI 
  3. CTRL+CLICK on the URL following "Running on local URL:" to run the WebUI

adit_bhutani_2-1692384152623.png

 

 

5. Enable Olive Optimized Path on AMD Radeon

Integrate the optimized model

  1. Copy generated optimized model (the “stable-diffusion-v1-5” folder) from Optimized Model folder into the directory stable-diffusion-webui-directml\models\ONNX. The ONNX folder may need to be created for some users.

Run the Automatic1111 WebUI with the Optimized Model

  1. Launch a new Anaconda/Miniconda terminal window
  2. Navigate to the directory with the webui.bat and enter the following command to run the WebUI with the ONNX path and DirectML. This will be using the optimized model we created in section 3.
    • webui.bat --onnx --backend directml
  3. CTRL+CLICK on the URL following "Running on local URL:" to run the WebUI

adit_bhutani_8-1692385515318.png

4. Pick “stable-diffusion-v1-5” from dropdown

adit_bhutani_9-1692385572420.png

 

6. Conclusion

Running on the default PyTorch path, the AMD Radeon RX 7900 XTX delivers 1.87 iterations/second.

Running on the optimized model with Microsoft Olive, the AMD Radeon RX 7900 XTX delivers 18.59 iterations/second.

adit_bhutani_1-1692387428847.png

End Result is up to 9.9X improvement to performance on AMD Radeon™ RX 7900 XTX.

 

7. Disclaimers & Footnotes

Links to third-party sites are provided for convenience and unless explicitly stated, AMD is not responsible for the contents of such linked sites, and no endorsement is implied. GD-98

Testing conducted by AMD as of August 15th, 2023, on a test system configured with a Ryzen9 7950X 3D(4.2GHz) CPU, 32GB DDR5, Radeon RX 7900XTX GPU, Windows 11 Pro, with AMD Software: Adrenalin Edition 23.7.2, using the application Stable Diffusion 1.5 with Microsoft Olive under Automatic 1111 vs. Default Automatic 1111. 

Performance may vary. System manufacturers may vary configurations, yielding different results. RS-587

The information presented in this document is for informational purposes only and may contain technical inaccuracies, omissions, and typographical errors. The information contained herein is subject to change and may be rendered inaccurate for many reasons, including but not limited to product and roadmap changes, component and motherboard version changes, new model and/or product releases, product differences between differing manufacturers, software changes, BIOS flashes, firmware upgrades, or the like. Any computer system has risks of security vulnerabilities that cannot be completely prevented or mitigated. AMD assumes no obligation to update or otherwise correct or revise this information. However, AMD reserves the right to revise this information and to make changes from time to time to

the content hereof without obligation of AMD to notify any person of such revisions or changes. GD-18.

 

THIS INFORMATION IS PROVIDED ‘AS IS.” AMD MAKES NO REPRESENTATIONS OR WARRANTIES WITH RESPECT TO THE CONTENTS HEREOF AND ASSUMES NO RESPONSIBILITY FOR ANY INACCURACIES, ERRORS, OR OMISSIONS THAT MAY APPEAR IN THIS INFORMATION. AMD SPECIFICALLY DISCLAIMS ANY IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY, OR FITNESS FOR ANY PARTICULAR PURPOSE. IN NO EVENT WILL AMD BE LIABLE TO ANY PERSON FOR ANY RELIANCE, DIRECT, INDIRECT, SPECIAL, OR OTHER CONSEQUENTIAL DAMAGES ARISING FROM THE USE OF ANY INFORMATION CONTAINED HEREIN, EVEN IF AMD IS EXPRESSLY ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.

Copyright 2023 Advanced Micro Devices, Inc. All rights reserved. AMD, the AMD Arrow logo, [insert all other AMD trademarks used in the material IN ALPHABETICAL ORDER here per AMD's Guidelines on Using Trademark Notice and Attribution] and combinations thereof are trademarks of Advanced Micro Devices, Inc. Microsoft is a registered trademark of Microsoft Corporation in the US and/or other countries. Other product names used in this publication are for identification purposes only and may be trademarks of their respective owners. Other product names used in this publication are for identification purposes only and may be trademarks of their respective owners.