Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations

About

Learning representations that accurately capture long-range dependencies in sequential inputs -- including text, audio, and genomic data -- is a key problem in deep learning. Feed-forward convolutional models capture only feature interactions within finite receptive fields while recurrent architectures can be slow and difficult to train due to vanishing gradients. Here, we propose Temporal Feature-Wise Linear Modulation (TFiLM) -- a novel architectural component inspired by adaptive batch normalization and its extensions -- that uses a recurrent neural network to alter the activations of a convolutional model. This approach expands the receptive field of convolutional sequence models with minimal computational overhead. Empirically, we find that TFiLM significantly improves the learning speed and accuracy of feed-forward neural networks on a range of generative and discriminative learning tasks, including text classification and audio super-resolution

Sawyer Birnbaum, Volodymyr Kuleshov, Zayd Enam, Pang Wei Koh, Stefano Ermon• 2019

Related benchmarks

TaskDatasetResultRank
Audio Super-ResolutionPiano (test)
SNR25.4
23
Speech EnhancementVCTK Vibration sensor 12-bit, 4-16 kHz upsampling (test)
LSD (Log-Spectral Distance)1.68
18
Speech EnhancementVCTK Accelerometer 12-bit, 4-16 kHz upsampling (test)
LSD1.82
18
Audio Super-ResolutionVCTK Single-speaker (test)
SNR19.5
15
Audio Super-ResolutionVCTK Multi-speaker (test)
SNR19.8
15
Bandwidth Extension (4-22 kHz upsampling)MagnaTagATune (test)
LSD1.7
15
Audio Super-ResolutionSingleSpeaker
SNR19.5
12
Audio Super-ResolutionMultiSpeaker
SNR19.8
12
Audio Super-ResolutionPiano
SNR25.4
12
Text ClassificationYelp 2
Accuracy95.6
12
Showing 10 of 23 rows

Other info

Code

Follow for update