Skip to content

🐛 [Bug] group_norm implementation may slows down SD vae #3648

@cehongwang

Description

@cehongwang

Bug Description

We noticed VAE Perf Gap b/w Torch-TRT and ONNX-TRT. ONNX-TRT is about 30-40% faster than Torch-TRT. We found that for group_norm layer, ONNX-TRT is using a workaround provided by TensorRT team, InstanceNormalization. Torch-TensorRT uses add_normalization for direct conversion.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions