Background Suppression Network for Weakly-supervised Temporal Action Localization
Background Suppression Network for Weakly-supervised Temporal Action Localization
Pilhyeon Lee Youngjung Uh Hyeran Byun

Abstract
Weakly-supervised temporal action localization is a very challenging problem because frame-wise labels are not given in the training stage while the only hint is video-level labels: whether each video contains action frames of interest. Previous methods aggregate frame-level class scores to produce video-level prediction and learn from video-level action labels. This formulation does not fully model the problem in that background frames are forced to be misclassified as action classes to predict video-level labels accurately. In this paper, we design Background Suppression Network (BaS-Net) which introduces an auxiliary class for background and has a two-branch weight-sharing architecture with an asymmetrical training strategy. This enables BaS-Net to suppress activations from background frames to improve localization performance. Extensive experiments demonstrate the effectiveness of BaS-Net and its superiority over the state-of-the-art methods on the most popular benchmarks - THUMOS'14 and ActivityNet. Our code and the trained model are available at https://github.com/Pilhyeon/BaSNet-pytorch.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| weakly-supervised-action-localization-on | BaS-Net | |
| weakly-supervised-action-localization-on-1 | BaS-Net | [email protected]: 34.5 [email protected]:0.95: 22.2 |
| weakly-supervised-action-localization-on-2 | BaS-Net | [email protected]: 38.5 |
| weakly-supervised-action-localization-on-4 | BasNet | [email protected]: 27.0 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.