Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations
About
Learning representations that accurately capture long-range dependencies in sequential inputs -- including text, audio, and genomic data -- is a key problem in deep learning. Feed-forward convolutional models capture only feature interactions within finite receptive fields while recurrent architectures can be slow and difficult to train due to vanishing gradients. Here, we propose Temporal Feature-Wise Linear Modulation (TFiLM) -- a novel architectural component inspired by adaptive batch normalization and its extensions -- that uses a recurrent neural network to alter the activations of a convolutional model. This approach expands the receptive field of convolutional sequence models with minimal computational overhead. Empirically, we find that TFiLM significantly improves the learning speed and accuracy of feed-forward neural networks on a range of generative and discriminative learning tasks, including text classification and audio super-resolution
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Audio Super-Resolution | Piano (test) | SNR25.4 | 23 | |
| Speech Enhancement | VCTK Vibration sensor 12-bit, 4-16 kHz upsampling (test) | LSD (Log-Spectral Distance)1.68 | 18 | |
| Speech Enhancement | VCTK Accelerometer 12-bit, 4-16 kHz upsampling (test) | LSD1.82 | 18 | |
| Audio Super-Resolution | VCTK Single-speaker (test) | SNR19.5 | 15 | |
| Audio Super-Resolution | VCTK Multi-speaker (test) | SNR19.8 | 15 | |
| Bandwidth Extension (4-22 kHz upsampling) | MagnaTagATune (test) | LSD1.7 | 15 | |
| Audio Super-Resolution | SingleSpeaker | SNR19.5 | 12 | |
| Audio Super-Resolution | MultiSpeaker | SNR19.8 | 12 | |
| Audio Super-Resolution | Piano | SNR25.4 | 12 | |
| Text Classification | Yelp 2 | Accuracy95.6 | 12 |