Real-time Neural Rendering

Calvin University

Latent diffusion models have high speed image synthesis capabilities that can used for sytlizing existing images and synthetic 3D geometry without the use of textures or materials.

The Idea

Neural rendering is a technology that could greatly revolutionize the field of computer graphics. By removing the need for textures, materials and rasterization to generate images, we can substantially increase the capabilities of prodecural generation algorithm, since a much lower level of detail is necessary. In place of these, we can use generative AI models to fill in the gaps so that the frames retain a high level of intricacy, and allow us to switch themes, styles, and settings instantaneously. In this project, I will explore the usage of Latent Consistency Models as a method of realtime neural rendering, in conjunction with ControlNet and AnimateDiff.

Demo Video

Tech Stack

  • Latent Consistency Models
  • Custom ControlNet using depth + material mask
  • ONNX inference engine + vectorized scheduler
  • Olive/AIMET compute graph optimizations
  • Model quantization + Hardware neural engine
  • Optical Flow estimation + frame interpolation

Current Progress

  • Conversion to ONNX (done)
  • Custom C++/C# pipeline (done)
  • ControlNet mods (done)
  • DirectX 12 support (done)
  • CUDA support (done)
  • TensorRT support (in progress)
  • Model quantization + fine-tuning (in progress)
  • Optical Flow estimation + frame interpolation (not started)
  • Custom ControlNet using Multimodal Material Segmentation (not started)

Related Links

This project is based on a number of recent works and projects: