Optimizing Transformer-based Diffusion Models for Video Generation with NVIDIA TensorRT
This article was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. State-of-the-art image diffusion models take tens of seconds to process a single image. This makes video diffusion even more challenging, requiring significant computational resources and high costs. By leveraging the latest FP8 quantization features on NVIDIA Hopper GPUs […]
Optimizing Transformer-based Diffusion Models for Video Generation with NVIDIA TensorRT Read More +