Back To Schedule

PhilTorch

Accelerating Automatic Differentiation of Digital Filters In PyTorch

00:00 - 00:00 | Friday 31st October 2025 |
Intermediate
Advanced

Infinite impulse response (IIR) filters are essential building blocks in many audio applications, due to their strong modelling capability with low computational cost. However, this efficiency advantage is not immediately apparent when incorporating IIR into common non-compiled deep learning frameworks, such as PyTorch, for end-to-end learning. Since PyTorch does not have a low-level automatic differentiation function for recursion, such as IIR, a naive implementation will create a significant number of function and memory allocation calls, thereby slowing down the process. Tackling this issue is crucial for developing real-time systems that combine neural networks and audio filters.

This talk aims to showcase how PhilTorch, a PyTorch package that facilitates efficient gradient optimisations of filters, implement automatic differentiation for IIR using custom kernels. We will see that automatic differentiations through IIR filters also involve IIR filters. By wrapping IIR filters into custom functions, any low-level realisation of filters outside PyTorch can be used to accelerate both filtering and gradient computations. In addition, we will investigate realisations that can be accelerated significantly on GPUs, including diagonalised state-space models and partial fraction expansions, and benchmark them against direct form realisation.

Chin-Yun Yu

Chin-Yun is a third-year PhD student at the Centre for Digital Music, Queen Mary University of London, working on expressive and controllable voice synthesis. He received his B.S. degree in computer science from the National Yang Ming Chiao Tung University, Taiwan, in 2018. He began conducting independent audio research in 2019, with many of his implementations open-sourced on GitHub. He is the main contributor of the differentiable lfilter function in TorchAudio. He and his team also won the bronze medal in the 2021 Music Demixing Challenge with their source separation model, "Danna-Sep".

His research interests include differentiable signal processing, music information retrieval, deep generative models, and spatial audio. Besides voice synthesis being his central research theme, he also has experience in topics, including multipitch estimation, source separation, neural vocoders, bandwidth extension, audio effects modelling, and time-of-arrival estimation in spatial audio.

VolumetricCondensed