Repository URL to install this package:
|
Version:
1.23.2 ▾
|
# -------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
# --------------------------------------------------------------------------
import logging
from fusion_layernorm import FusionLayerNormalization
from fusion_mha_mmdit import FusionMultiHeadAttentionMMDit
from fusion_options import FusionOptions
from import_utils import is_installed
from onnx import ModelProto
from onnx_model_bert import BertOnnxModel
logger = logging.getLogger(__name__)
class MmditOnnxModel(BertOnnxModel):
def __init__(self, model: ModelProto, num_heads: int = 0, hidden_size: int = 0):
"""Initialize Multimodal Diffusion Transformer (MMDiT) ONNX Model.
Args:
model (ModelProto): the ONNX model
num_heads (int, optional): number of attention heads. Defaults to 0 (detect the parameter automatically).
hidden_size (int, optional): hidden dimension. Defaults to 0 (detect the parameter automatically).
"""
assert (num_heads == 0 and hidden_size == 0) or (num_heads > 0 and hidden_size % num_heads == 0)
super().__init__(model, num_heads=num_heads, hidden_size=hidden_size)
def postprocess(self):
self.prune_graph()
self.remove_unused_constant()
def fuse_layer_norm(self):
layernorm_support_broadcast = True
logger.warning(
"The optimized model requires LayerNormalization with broadcast support. "
"Please use onnxruntime-gpu>=1.21 for inference."
)
fusion = FusionLayerNormalization(
self, check_constant_and_dimension=not layernorm_support_broadcast, force=True
)
fusion.apply()
def fuse_multi_head_attention(self):
fusion = FusionMultiHeadAttentionMMDit(self)
fusion.apply()
def optimize(self, options: FusionOptions | None = None, add_dynamic_axes: bool = False):
assert not add_dynamic_axes
if is_installed("tqdm"):
import tqdm # noqa: PLC0415
from tqdm.contrib.logging import logging_redirect_tqdm # noqa: PLC0415
with logging_redirect_tqdm():
steps = 5
progress_bar = tqdm.tqdm(range(steps), initial=0, desc="fusion")
self._optimize(options, progress_bar)
else:
logger.info("tqdm is not installed. Run optimization without progress bar")
self._optimize(options, None)
def _optimize(self, options: FusionOptions | None = None, progress_bar=None):
if (options is not None) and not options.enable_shape_inference:
self.disable_shape_inference()
# Remove cast nodes that having same data type of input and output based on symbolic shape inference.
self.utils.remove_useless_cast_nodes()
if progress_bar:
progress_bar.update(1)
if (options is None) or options.enable_layer_norm:
self.fuse_layer_norm()
self.fuse_simplified_layer_norm()
if progress_bar:
progress_bar.update(1)
if (options is None) or options.enable_gelu:
self.fuse_gelu()
if progress_bar:
progress_bar.update(1)
if (options is None) or options.enable_attention:
self.fuse_multi_head_attention()
if progress_bar:
progress_bar.update(1)
self.postprocess()
if progress_bar:
progress_bar.update(1)
logger.info(f"opset version: {self.get_opset_version()}")
def get_fused_operator_statistics(self):
"""
Returns node count of fused operators.
"""
op_count = {}
ops = [
"FastGelu",
"MultiHeadAttention",
"LayerNormalization",
"SimplifiedLayerNormalization",
]
for op in ops:
nodes = self.get_nodes_by_op_type(op)
op_count[op] = len(nodes)
logger.info(f"Optimized operators:{op_count}")
return op_count