Skip to content

nod-ai/sdxl-scripts

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Oct 8, 2024
4fa7ccb · Oct 8, 2024
Aug 7, 2024
Apr 11, 2024
Jul 12, 2024
May 31, 2024
Jul 24, 2024
Jul 26, 2024
Aug 22, 2024
Jun 7, 2024
Jul 23, 2024
Aug 2, 2024
May 31, 2024
May 28, 2024
Oct 8, 2024
Jul 19, 2024
Oct 8, 2024
Aug 2, 2024

Repository files navigation

SDXL IRs and Scripts

SDXL end-to-end benchmarking

  1. Checkout and compile IREE with release build and export PATH=/path/to/iree/build/release/tools:$PATH
  2. Compile the full SDXL model: ./compile-txt2img.sh gfx942 (where gfx942 is the target for MI300X)
  3. Run the benchmark: ./benchmark-txt2img.sh N /path/to/weights/irpa (where N is the GPU index)

Model IRs and weights

Caution

IRs in the following table might be stale. Use the ones in the base_ir/ directory instead.

Note

SDXL-turbo is only different from SDXL in its usage and training/weights. The model architecture (and therefore the weights-stripped MLIR) are equivalent.

Variant Submodel MLIR (No Weights) (Config A) safetensors Splat IRPA MLIR (No Weights) (Config B)
SDXL1.0 1024x1024 (f16, BS1, len64)
UNet + attn Torch - Linalg - - Azure
UNet + PNDMScheduler Azure
Clip1 Azure - -
Clip2 Azure - -
VAE decode + attn Azure - = Azure
VAE encode + attn [GCloud][sdxl-1-1024x1024-f16-stripped-weight-vae-encode] Same as decode - -
SDXL1.0 1024x1024 (f32, BS1, len64)
UNet + attn Azure Azure Azure Azure
Clip1 Azure Azure Azure -
Clip2 Azure Azure Azure -
VAE decode + attn Azure Azure Azure Azure
SDXL compiled pipeline IRPAs (f16)
UNet scheduled_unet_f16.irpa
Prompt Encoder (CLIP1 + CLIP2) prompt_encoder_f16.irpa
VAE vae_decode_f16.irpa

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages