rocm/pytorch-xdit images support several diffusion model inference workloads on gfx942 and gfx950 series (AMD Instinct™ MI300X, MI308X, MI325X and MI350X, MI355X) GPUs. The image has ROCm preview (based on TheRock) and uses xDiT distributed diffusion model inference framework for high-performance image and video generation.
Use
git clone https://github.com/ROCm/MAD
cd MAD
pip install -r requirements.txtto clone the ROCm Model Automation and Dashboarding (MAD) repository to a local directory and install the required packages on the host machine.
Execute benchmarks with
MAD_SECRETS_HFTOKEN=HFTOKEN madengine run --tags TAG --live-outputwhere HFTOKEN is a valid Hugging Face token (note that some models are gated) and TAG
a supported MAD model tag. See Available models section for more details. The inference
latencies can be found from results.csv once the benchmark runs have finished.
| MAD model TAG | Model repository |
|---|---|
| pyt_xdit_flux | Flux.1 |
| pyt_xdit_flux_kontext | Flux.1 Kontext |
| pyt_xdit_flux_2 | Flux.2 |
| pyt_xdit_flux_2_klein | Flux.2 Klein |
| pyt_xdit_hunyuanvideo | HunyuanVideo |
| pyt_xdit_hunyuanvideo_1_5 | HunyuanVideo 1.5 |
| pyt_xdit_ltx_2 | LTX-2 |
| pyt_xdit_qwen_image | Qwen-Image |
| pyt_xdit_qwen_image_edit | Qwen-Image-Edit |
| pyt_xdit_sd_3_5 | Stable diffusion 3.5 |
| pyt_xdit_wan_2_1 | Wan 2.1 |
| pyt_xdit_wan_2_2 | Wan 2.2 |
| pyt_xdit_z_image | Z-Image |