Abstract
Despite the recent advances in video classification, progress in spatio-temporal action recognition has lagged behind. A major contributing factor has been the prohibitive cost of annotating videos frame-by-frame. In this paper, we present a spatio-temporal action recognition model that is trained with only video-level labels, which are significantly easier to annotate. Our method leverages per-frame person detectors which have been trained on large image datasets within a Multiple Instance Learning framework. We show how we can apply our method in cases where the standard Multiple Instance Learning assumption, that each bag contains at least one instance with the specified label, is invalid using a novel probabilistic variant of MIL where we estimate the uncertainty of each prediction. Furthermore, we report the first weakly-supervised results on the AVA dataset and state-of-the-art results among weakly-supervised methods on UCF101-24.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Video classification has witnessed great advances recently due to large datasets such as Kinetics [20] and Moments in Time [30] which have enabled training of specialised neural network architectures for video [5, 9]. However, progress in other video understanding tasks, such as spatio-temporal action detection, has lagged behind in comparison. There are fewer datasets for action recognition, which are also significantly smaller than their video-classification counterparts. A reason for this is the exorbitant cost of annotating videos with spatio-temporal labels – each frame of an action has to be manually labelled with a bounding box. Moreover, annotating temporal boundaries of actions is not only arduous, but often ambiguous with annotators failing to reach consensus about the start and end times of an action [6, 39].
In this paper, we propose a method to train spatio-temporal action detectors using only weak, video-level annotations as shown in Fig. 1. To achieve this, we leverage image-based person detectors which have been trained on large image datasets such as Microsoft COCO [26] and are accurate across large variations in appearance, scene and pose. We adopt a Multiple Instance Learning (MIL) framework, where a person tubelet is an instance, and all person tubelets in the video form a bag. An important consideration in our approach is the presence of label noise: this is introduced from using off-the-shelf person detectors which have not been trained on the video-dataset of interest, and also the fact that we have to sample tubelets from large bags in long videos due to memory constraints. In both of these scenarios, the standard Multiple Instance Learning assumption [7], that each bag contains at least one instance with the bag-level label, may be violated. We are not aware of previous work that has explicitly addressed this problem, and we do so with a probabilistic variant of MIL where we estimate the uncertainty of an instance-level prediction.
Using our approach, we obtain state-of-the-art results among weakly-supervised methods on the UCF101-24 dataset. Furthermore, we report, to our knowledge, the first weakly-supervised results on the AVA dataset (the only large-scale dataset for spatio-temporal action recognition), where we also show the accuracy trade-offs when annotating video-clips for time intervals of varying durations.
2 Related Work
Most prior work on spatio-temporal action recognition has been fully-supervised. Initial approaches in the area used 3D sliding window detectors in conjunction with handcrafted, volumetric features [21, 25]. Current state-of-the-art approaches are temporal extensions of object detection architectures [19, 35, 41, 52] such as Faster-RCNN [37] and SSD [27]. These approaches predict bounding boxes around the action in a frame, using as input either a single frame along with optical flow to capture temporal information [38, 41] or multiple frames at the input to provide temporal context [19]. The predicted bounding boxes are then linked over time using an online, greedy algorithm or dynamic programming to create spatio-temporal tracks. Our work builds on these methods by also utilising a detection architecture and spatio-temporal linking. However, these approaches all require bounding box annotations at each frame in the video whilst we only use video-level labels which are significantly cheaper to acquire.
Weakly supervised approaches to spatio-temporal action recognition have also been explored before as they enable a significant reduction in annotation time and cost. Relevant to our approach is the work of [6]. Cheŕon et al. [6] also use person detections, and infer their action labels using a formulation based on discriminative clustering [2]. Although their approach allows them to incorporate different types of supervision, it effectively learns a linear classifier on top of pretrained, deep features. Our method in contrast is learned fully end-to-end. Mettes et al. [29] also employed Multiple Instance Learning (MIL), but used action proposals [47] instead of the human detections used by our work and [6]. However, [29], rely on additional cheap “point” annotations (a single spatial point annotated for a subset of the frames which constitute the action) which also ensures that the standard MIL assumption is not violated. In follow-up work [28], the authors removed the need for “point” annotations by incorporating biases (i.e. the presence of objects in the video, a bias that actions typically occur in the centre of a frame) instead. Finally, Weinzaepfel et al. [50] also used a Multiple Instance Learning framework in conjunction with human detections. The authors, however, assumed that sparse spatial supervision was present (i.e. bounding boxes for a small subset of frames in the action tube), unlike our method which requires video-level labels alone.
We also note that many approaches have addressed temporal action detection (localising actions in time but not space) with only video-level tags as supervision [31, 34, 42, 48]. UntrimmedNets [48] uses a network with two branches, a classification module to perform action classification and a selection module to select relevant frames. Hide-and-Seek [42] obtains more precise temporal boundaries by forcing the network to attend to more discriminative frames by randomly hiding parts of videos. However, these methods are trained and evaluated on datasets such as ActivityNet [4] and THUMOS14 [18], which contain mostly one action per video, and are thus significantly less challenging than datasets such as AVA [14] which we evaluate on.
Finally, we note that another approach to combat the effort of dataset annotation has been various forms of self-supervised learning, where discriminative feature representations can be learned with unlabelled data. Examples in video include cross-modal self-supervised learning by learning correspondences between the audio and image streams readily available in videos [1, 33, 51], transcribed speech [44] or using meta-data such as hashtags [10] as a form of weak labelling. Self-supervised approaches, however, are complementary to our approach, as they still require a limited amount of fully-labelled data for the final task of interest. In our weakly-supervised action detection scenario, we never have access to full, spatio-temporal ground-truth annotations for a single training example.
3 Proposed Approach
As shown in Fig. 1, given a set of video clips, with only clip-level annotations of the actions taking place, our goal is to learn a model to recognise and localise these actions in space and time. Our method is based on Multiple Instance Learning (MIL) which we briefly review in Sect. 3.1. Thereafter, we show how we use it for weakly-supervised spatio-temporal action recognition in Sect. 3.2. We then describe how the standard MIL assumption, is often violated in our scenario and describe a method to mitigate this by leveraging uncertainty estimates by our network in Sect. 3.3. Finally, we discuss implementation details of our network in Sect. 3.4.
3.1 Multiple Instance Learning
In the standard Multiple Instance Learning (MIL) [7] formulation, one is given a bag of N instances, denoted as \(x = \{x_1, x_2, \ldots , x_N\}\). The class labels for each of the instances is unknown, but the label for the entire bag, x, is known. The standard MIL assumption is that a bag is assigned a class label if at least one instance in the bag is associated with this label. More formally, we consider the multi-label classification case, where the label vector for the bag is \(y \in \mathbb {R}^{C}\), and \(y_l = 1\) if there is at least one instance with the \(l^{th}\) label is present in the bag, and \(y_l = 0\) otherwise. Note that each bag can be labelled with multiple of the C class labels.
Our goal is to train an instance-level classifier (parameterised as a neural network), that predicts \(p(y_{l} = 1 | x_j)\), or the label probabilities for the \(j^{th}\) instance. However, as we only have the labels for the entire bag, and not each instance, MIL methods aggregate the set of instance-level probabilities, \(\{p_{ij}\}\) for a bag i, to bag-level probabilities, \(p_i\), using an aggregation function, \(g(\cdot )\), where the probabilities are obtained from a suitable activation function (sigmoid or softmax) on the logits output by the neural network:
Once we have bag-level predictions, we can apply a standard classification loss between the bag-level probabilities and bag-level ground truth, and train a neural network with stochastic gradient descent. Since we consider the multi-label classification case, we use the binary cross-entropy:
Note that we defined \(p_{il}\) as the bag-level probability of the \(i^{th}\) bag taking the \(l^{th}\) label, which is obtained using Eq. 1, and \(N_b\) is the number of bags in the mini-batch.
Aggregation. The aggregation function, \(g(\cdot )\), can naturally be implemented in neural networks as a global pooling function over all outputs of the network. Common, permutation-invariant pooling functions include, max-pooling, generalised mean-pooling and log-sum-exponential (LSE) pooling [3] (a smooth and convex approximation of the maximum function) respectively:
Max-pooling only considers the top-scoring instance in the bag, and thus naturally captures the MIL assumption that at least one instance in the bag has the specified, bag-level label. Moreover, it can also be more robust to instances in the bag that do not have the bag-level label. However, mean and LSE pooling have been employed in applications such as weakly-supervised segmentation [36], object recognition [45] and medical imaging [24] where multiple instances in the bag do typically have the bag-level label. Note that higher values of the r hyperparameter for both these functions increases their “peakiness” and approximates the maximum value. For our scenario, detailed in the next section, we found max-pooling to be the most appropriate.
3.2 Weakly-Supervised Spatio-Temporal Action Recognition as Multiple Instance Learning
Our goal is to learn a model to recognise and localise actions in space and time given only video-level annotations. To facilitate this, we leverage a person detector that has been trained on a large image dataset, i.e. Microsoft COCO [26]. Concretely, we run a person detector on our training videos, and create person tubelets which are person detections over K consecutive frames in the video. Our bag for multiple instance learning thus consists of all the tubelets within a video, and is annotated with the video-level labels that we have as supervision, as illustrated in Fig. 2. Note that the size of the bag varies for every video clip, as the bag size is determined by the length of the video and the number of detected people.
As shown in Fig. 2, our network architecture for this task is a Fast-RCNN [12] style detector that has been extended temporally. Given a video clip of K frames, and proposals which in our case are person detections, the network classifies the action(s) taking place at the centre frame of each proposal, given the temporal context of the \(K-1\) frames around it.
Note that the spatio-temporal localisation task is effectively factorised: the spatial localisation capability of the model depends on the quality of the person detections. Temporal localisation, on the other hand, is performed by linking person tubelets through the video as commonly done in the literature [6, 19, 41, 52], since this method can scale to arbitrarily long videos. We use the same algorithm as Kalogeiton et al. [19] which links together detections within a small temporal window greedily based on the spatial intersection over union (IoU) between bounding boxes on consecutive frames.
Finally, note that for a video consisting of T frames, the bag could consist of \(T - K + 1\) person tubelets if a person is detected on each frame of the video, and a tubelet is started from each frame. Due to memory limitations, it is infeasible to fit an entire bag onto a GPU for training. As a result, we uniformly sample instances from each bag during training, whilst still retaining the original bag-level label. This introduces additional noise into the problem, as detailed next.
3.3 Label Noise and Violation of the Standard MIL Assumption
The standard MIL assumption, that at least one instance in the bag is assigned the bag-level label is often violated in our scenario. There are two primary factors for this: Firstly, due to computational constraints, we cannot process a whole bag at a time, but must instead sample instances from a bag. It is therefore possible to sample a bag that does not contain any tubelets with the labelled action. The likelihood of this occurring is inversely proportional to the ratio of the duration of the labelled action to the total video length. Secondly, in a weakly-supervised scenario, we use person detectors that are not trained on the video dataset of interest. Consequently, there can be failures in the detector, especially when there is a large domain gap between the detector’s training distribution and the video dataset. False negatives (missing detections for people in the scene) are a particular issue because it is possible that we do not have a single person tubelet in the bag that corresponds to the labelled action.
Therefore, there are cases when there is no tubelet which actually has the bag-level label. To handle these cases, inspired by [22, 32], we modify the network to additionally predict the uncertainty \(\sigma \in \ \mathbb {R}^{C}\) for each binary label for all tubelets in the bag. Intuitively, to minimise the training error, the network can predict the bag-level label with low uncertainty or it can predict a high uncertainty to avoid being penalised heavily for noisy bags where the bag-level label is not present in any of the tubelets. The final loss, in conjunction with the original cross entropy, is defined as:
As shown by [23], this corresponds to assuming a Boltzmann distribution on the output of the network with a temperature of \(\sigma ^2\), and approximately minimising its log-likelihood.
The loss surface of this probabilistic loss is visualised in Fig. 3. Note how the loss is the lowest when the predicted label is correct and there is low uncertainty. However, the loss is not excessive if the incorrect label is predicted with a high uncertainty. This is in contrast with the standard cross-entropy loss which penalises incorrect predictions heavily.
3.4 Network Architecture and Implementation
Our action detector is similar to Fast-RCNN [12] using the SlowFast [9] video network architecture based on the ResNet-50 backbone [16] pretrained on Kinetics [20]. As described in Sect. 3.2, we use region proposals obtained from a Faster-RCNN detection model trained with Detectron [13]. Region-of-interest features [12] are extracted from the last feature map of “res5” using RoIAlign [15]. Our choice for this architecture is motivated by the fact that it is simple and has achieved state-of-the-art results on the AVA dataset [14] in a fully-supervised setting [9]. Note that our network does not use additional optical flow inputs (which can be considered as an additional source of supervision) as common in other video architectures [5, 6, 19, 41].
We predict the uncertainty, \(\sigma \in \mathbb {R}^C\) for each of the C binary labels defined by the dataset for each tubelet. As we use max-pooling to aggregate the tubelet predictions, we select the uncertainty prediction corresponding to the selected tubelet for computing the loss. For numerical stability, we predict \(v := \log \sigma ^2\) with our network, using the “softplus”, \(f(x) = \log (1 + \exp (-x))\), activation function to ensure positivity. We then compute \(\frac{1}{\sigma ^2} = \exp (-v)\), and avoid the possibility of dividing by 0 which could be the case if we predicted \(\sigma ^2\) directly with the network.
We train our network with synchronous stochastic gradient descent (SGD), using 8 GPUs and a batch size of 4 on each GPU. In our case, each element of a batch is of a bag from Multiple Instance Learning. Each bag samples a maximum of 4 tubelets. Each tubelet itself consists of 16 frames.
4 Experiments
4.1 Experimental Set-Up
We evaluate our method on UCF101-24 and AVA, described in more detail below. Note that other video datasets such as THUMOS [18] and ActivityNet [4] are not suitable for spatiotemporal localisation, as they lack bounding box annotations.
UCF101-24: UCF101-24 is a subset of the UCF101 [43] dataset, consisting of 24 action classes with spatio-temporal localisation annotation, released as bounding box annotations of humans. Although each video contains only a single action class, it may contain multiple individuals performing the action with different spatial and temporal boundaries. Moreover, there may also be people present in the video who are not performing any labelled action. Following standard practice, we use the corrected annotations of [41] and report the mean average precision at a video level (Video AP) for the first split of the dataset. For evaluating the Video AP, we link tubelets together using the algorithm of [19].
AVA [14]: This dataset consists of 430, 15 min video clips obtained from movies. 80 atomic visual actions are annotated exhaustively for all people in the video, where one person is often simultaneously performing multiple actions. The dataset annotates keyframes at every second in the video. Following standard practice, we report the Frame AP at an IoU threshold of 0.5 using v2.2 annotations.
4.2 Experiments on UCF101-24
We first conduct ablation studies of our model on the UCF101-24 dataset. We discard the spatio-temporal annotations for the whole untrimmed video, and so our bag in multiple instance learning contains tubelets from the whole video.
Ablation Study. Table 1 ablates different variants of our method: The most naïve baseline is to not perform any multiple instance learning, and to simply train in a fully-supervised fashion assuming that the label of a tubelet is the video-level label. As shown in the first row of Table 1, this method performs the worst as the assumed tubelet-level labels are often incorrect. The use of multiple instance learning improves results, with the various aggregation functions performing similarly. Max-pooling, however, performs the best, and we believe this is because the max operation is the most suitable for dealing with the noise present in our tubelets as described in Sect. 3.3. Note that for mean and LSE-pooling, we set \(r = 1\). Finally, introducing our uncertainty-based loss function improves results even further, obtaining a Video mAP of 35.0 at a threshold of 0.5. This is 80% of the performance achieved by our fully-supervised baseline.
Person Detections on UCF101-24. Note that for our weakly-supervised experiments, the person tubelets for training are obtained from a Faster-RCNN [37] person detector that has only been trained on Microsoft COCO [26]. There is a significant domain gap between COCO and UCF, and the annotation protocol of person boxes on UCF is also not consistent (for example, the bounding box for a person riding a horse often includes the horse in UCF) with that of COCO. These discrepancies are reflected by the fact that our person detections used during training only have a recall of 46.9% compared to the ground truth person boxes, when using an IoU threshold of 0.5 to signify a correct match. Furthermore, the precision of our person tubelets on the training set is only 21.1%. A major contributing factor to this is that UCF action annotations are not exhaustive – there may be people in the video who are not labelled at all as they are not performing an annotated action. These people will, however, still be detected by a COCO-trained detector and considered as false positives during this evaluation.
The fact that we are able to train our model with these annotations demonstrates the ability of our multiple instance learning method to handle label noise in the training set. The inconsistencies in the UCF101-24 dataset labelling are detailed further in the supplementary, and has also been noted previously by Chéron et al. [6].
Noise in the person detections are not a problem for the training of our fully-supervised baseline, as it is trained with ground-truth boxes in addition to predicted boxes. As we have box-level supervision in this case, predicted detections which have an IoU of more than 0.5 with a ground-truth detection are assigned the label of the ground-truth box, or the negative label otherwise, during fully-supervised training.
As the goal of this paper is not to develop a better human detector or tracker for building the person tubelets, we use the Faster-RCNN detector released publicly by Chéron et al. [6] for all our evaluations on the UCF101-24 validation set. This detector was originally trained on COCO and then finetuned on the UCF101-24 training set using Detectron [13].
The Effect of Tubelet Sampling. For the tubelets of length \(K =16\) that we use, there is a mean of 33.1 tubelets per video in the UCF101-24 dataset. In computing this, we only consider tubelets which have a spatio-temporal IoU of less than 0.5 with each other. More tubelets would be obtained if we counted one from each frame of the video.
As we can fit a maximum of 16 tubelets onto a 16 GB Nvidia V100 GPU, it is clear that it is necessary to sample the tubelets in each bag. Note that UCF videos often have a high number of tubelets, as there are often many people in the video who are not labelled as performing an action. As described in the previous subsection, this is also a significant source of noise.
Table 2 shows the effect of changing the batch size (number of bags), and the number of tubelets sampled per bag, such that GPU memory usage is maximised. We can see that the uncertainty loss helps in all cases and that accuracy decreases with low batch sizes. We believe this is due to batch normalisation statistics being too correlated when more tubes are from the same video.
Comparison to State-of-the-Art. Table 3 compares our results to the state-of-the-art. The bottom-half of the table shows that we outperform previous weakly-supervised methods by a large margin. The top-half shows that our fully-supervised baseline is also competitive with the fully-supervised state-of-the-art, although that is not the main goal of this work. The fully-supervised methods which outperform our method are based on action detectors which directly predict the person proposals with the network, and are thus able to handle the person annotation peculiarities of the UCF101-24 dataset more effectively. We do not observe any issues with person detections for our experiments on AVA in the next section.
Qualitative Results. Figure 4 presents qualitative results of our method. The first two rows show success cases of our method where the tubelet detection and linking have performed well. The third row shows a failure case, since the basketball player represented by the green track is not actually performing the “Basketball Dunk” action. According to the UCF101-24 annotations, only the player represented with the blue track is performing this action. This video clip is thus an example of a video where there are many people not performing the action annotated for the video, and is especially challenging for our weakly-supervised method. The fourth row shows a different failure case as an error by the online tubelet linking algorithm (we used the same method as [19]) has made the identities of the two cyclists change after they occluded each other.
4.3 Experiments on AVA
In this section, we report what to our knowledge are the first weakly-supervised action detection experiments on AVA [14]. The AVA dataset labels keyframes in a 15 min video clip, where each keyframe is sampled every second (i.e. 1 Hz). The evaluation protocol of the AVA dataset measures the ability of an action detection model to classify the actions occuring in a keyframe given the temporal context around it.
We control the difficulty of the weakly-supervised action recognition problem by combining the annotations from N consecutive keyframes into a single, clip-level annotation. This effectively means that we are obtaining clip-level annotations for sub-clips of N seconds from the original AVA video. The weakly-supervised problem gets more difficult as N increases, as the sub-clips get longer and the number of observed labels within each sub-clip increases. Note that when \(N = 1\), only the spatial localisation ability of the model is being tested, as during training, it is unknown which of the subclip-level labels correspond to each person tubelet in the MIL bag. When \(N > 1\), the subclip-level labels can correspond to zero, one or many of the person tubelets at different keyframes in the clip, and it is thus a more difficult task. As an AVA video clip consists of 900 s, \(N = 900\) represents the most extreme case when spatio-temporal annotations are discarded for the entire 15 min video.
Table 4 shows the results of our model in this setting. As expected, the performance of our method improves the shorter the sub-clip. For \(N = 1\) and \(N = 5\), our method obtains 90% and 72% of fully-supervised performance respectively, suggesting that bounding-box level annotations are not required for training action recognition models if the video clips are annotated over short temporal intervals. Understandably, the results from \(N = 900\), where we use the video-level annotations over the whole 15 min clip are the worst as it is the most difficult setting.
Figure 5 further analyses the per-class results for the different levels of supervision presented in Table 4. As expected, stronger levels of supervision (shorter sub-clip durations) result in better per-class accuracy. However, some action classes are affected more than others by weaker labels (longer sub-clips). Examples of this include “sing to” and “listen to” which show a larger difference to the fully-supervised baseline than other classes. Moreover, some classes such as “watch (a person)”, “get up”, “close (e.g., a door, a box)” and “hand clap” perform reasonably when trained with sub-clips (\(N \le 10\)), but much more poorly when trained with longer sub-clips.
Finally, we compare our fully-supervised baseline to the state-of-the-art in Table 5. Note that our weakly-supervised result from sub-clips of 10 s (Table 4) outperforms the original fully-supervised baseline using introduced by the AVA dataset [14] that uses both RGB and optical flow as inputs. Our model, on the other hand, only uses RGB as its input modality. Our SlowFast model performs similarly to the published results of the original authors [9]. Note that we have not used Non-local [49], test-time augmentation or ensembling which are all complementary methods to improve performance [9]. We can see that in contrast to the UCF dataset in the previous section, our person detector is accurate on AVA, and so a Fast-RCNN-style detector using person tubelets as proposals can achieve state-of-the-art results.
5 Conclusion and Future Work
We have proposed a weakly supervised spatio-temporal action detection method based on Multiple Instance Learning (MIL). Our approach incorporates uncertainty predictions made by the network such that it can better handle noise in our bags and violations of the standard MIL assumption by predicting a high uncertainty for noisy bags which cannot be classified correctly. We achieve state-of-the-art results among weakly supervised methods on the UCF101-24 dataset, and also report the first weakly-supervised results on AVA, which is the only large-scale action recognition dataset. Our analysis of the accuracy trade-offs as the time intervals for which sub-clips of the video are annotated will also aid future dataset annotation efforts.
Future work is to incorporate additional sources of noisy, weakly-labelled data, such as data which can be scraped off internet search engines.
References
Arandjelovic, R., Zisserman, A.: Look, listen and learn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 609–617 (2017)
Bach, F.R., Harchaoui, Z.: DIFFRAC: a discriminative and flexible framework for clustering. In: Advances in Neural Information Processing Systems, pp. 49–56 (2008)
Boyd, S., Boyd, S.P., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)
Caba Heilbron, F., Escorcia, V., Ghanem, B., Carlos Niebles, J.: ActivityNet: a large-scale video benchmark for human activity understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 961–970 (2015)
Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299–6308 (2017)
Chéron, G., Alayrac, J.B., Laptev, I., Schmid, C.: A flexible model for training action localization with varying levels of supervision. In: Advances in Neural Information Processing Systems, pp. 942–953 (2018)
Dietterich, T.G., Lathrop, R.H., Lozano-Pérez, T.: Solving the multiple instance problem with axis-parallel rectangles. Artif. Intell. 89(1–2), 31–71 (1997)
Escorcia, V., Dao, C.D., Jain, M., Ghanem, B., Snoek, C.: Guess where? Actor-supervision for spatiotemporal action localization. arXiv preprint arXiv:1804.01824 (2018)
Feichtenhofer, C., Fan, H., Malik, J., He, K.: SlowFast networks for video recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 6202–6211 (2019)
Ghadiyaram, D., Tran, D., Mahajan, D.: Large-scale weakly-supervised pre-training for video action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 12046–12055 (2019)
Girdhar, R., Carreira, J., Doersch, C., Zisserman, A.: Video action transformer network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 244–253 (2019)
Girshick, R.: Fast R-CNN. In: ICCV (2015)
Girshick, R., Radosavovic, I., Gkioxari, G., Dollár, P., He, K.: Detectron (2018). https://github.com/facebookresearch/detectron
Gu, C., et al.: AVA: a video dataset of spatio-temporally localized atomic visual actions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6047–6056 (2018)
He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: ICCV (2017)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)
Hou, R., Chen, C., Shah, M.: Tube convolutional neural network (T-CNN) for action detection in videos. In: International Conference on Computer Vision, pp. 5822–5831 (2017)
Jiang, Y.G., et al.: THUMOS challenge: action recognition with a large number of classes (2014)
Kalogeiton, V., Weinzaepfel, P., Ferrari, V., Schmid, C.: Joint learning of object and action detectors. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4163–4172 (2017)
Kay, W., et al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)
Ke, Y., Sukthankar, R., Hebert, M.: Efficient visual event detection using volumetric features. In: International Conference on Computer Vision, vol. 1, pp. 166–173. IEEE (2005)
Kendall, A., Gal, Y.: What uncertainties do we need in Bayesian deep learning for computer vision? In: Advances in Neural Information Processing Systems, pp. 5574–5584 (2017)
Kendall, A., Gal, Y., Cipolla, R.: Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: Computer Vision and Pattern Recognition, pp. 7482–7491 (2018)
Kraus, O.Z., Ba, L.J., Frey, B.: Classifying and segmenting microscopy images using convolutional multiple instance learning. arXiv preprint arXiv:1511.05286 (2015)
Laptev, I., Pérez, P.: Retrieving actions in movies. In: International Conference on Computer Vision, pp. 1–8. IEEE (2007)
Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740–755. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10602-1_48
Liu, W., et al.: SSD: single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21–37. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2
Mettes, P., Snoek, C.G., Chang, S.F.: Localizing actions from video labels and pseudo-annotations. In: British Machine Vision Conference (BMVC) (2017)
Mettes, P., van Gemert, J.C., Snoek, C.G.M.: Spot on: action localization from pointly-supervised proposals. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 437–453. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46454-1_27
Monfort, M., et al.: Moments in time dataset: one million videos for event understanding. arXiv preprint arXiv:1801.03150 (2018)
Nguyen, P., Liu, T., Prasad, G., Han, B.: Weakly supervised action localization by sparse temporal pooling network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6752–6761 (2018)
Novotny, D., Albanie, S., Larlus, D., Vedaldi, A.: Self-supervised learning of geometrically stable features through probabilistic introspection. In: Computer Vision and Pattern Recognition, pp. 3637–3645 (2018)
Owens, A., Wu, J., McDermott, J.H., Freeman, W.T., Torralba, A.: Ambient sound provides supervision for visual learning. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 801–816. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_48
Paul, S., Roy, S., Roy-Chowdhury, A.K.: W-TALC: weakly-supervised temporal activity localization and classification. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 563–579 (2018)
Peng, X., Schmid, C.: Multi-region two-stream R-CNN for action detection. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 744–759. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46493-0_45
Pinheiro, P.O., Collobert, R.: From image-level to pixel-level labeling with convolutional networks. In: Computer Vision and Pattern Recognition, pp. 1713–1721 (2015)
Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems, pp. 91–99 (2015)
Saha, S., Singh, G., Sapienza, M., Torr, P.H., Cuzzolin, F.: Deep learning for detecting multiple space-time action tubes in videos. In: BMVC (2016)
Sigurdsson, G.A., Russakovsky, O., Gupta, A.: What actions are needed for understanding human actions in videos? In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2137–2146 (2017)
Singh, G., Saha, S., Cuzzolin, F.: TraMNet - transition matrix network for efficient action tube proposals. In: Jawahar, C.V., Li, H., Mori, G., Schindler, K. (eds.) ACCV 2018. LNCS, vol. 11366, pp. 420–437. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-20876-9_27
Singh, G., Saha, S., Sapienza, M., Torr, P.H., Cuzzolin, F.: Online real-time multiple spatiotemporal action localisation and prediction. In: International Conference on Computer Vision, pp. 3637–3646 (2017)
Singh, K.K., Lee, Y.J.: Hide-and-seek: forcing a network to be meticulous for weakly-supervised object and action localization. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3544–3553. IEEE (2017)
Soomro, K., Zamir, A.R., Shah, M.: UCF101: a dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402 (2012)
Sun, C., Myers, A., Vondrick, C., Murphy, K., Schmid, C.: VideoBERT: a joint model for video and language representation learning. In: International Conference on Computer Vision, pp. 7464–7473 (2019)
Sun, C., Paluri, M., Collobert, R., Nevatia, R., Bourdev, L.: ProNet: learning to propose object-specific boxes for cascaded neural networks. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016
Sun, C., Shrivastava, A., Vondrick, C., Murphy, K., Sukthankar, R., Schmid, C.: Actor-centric relation network. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 318–334 (2018)
Van Gemert, J.C., Jain, M., Gati, E., Snoek, C.G., et al.: APT: action localization proposals from dense trajectories. In: BMVC, vol. 2, p. 4 (2015)
Wang, L., Xiong, Y., Lin, D., Van Gool, L.: UntrimmedNets for weakly supervised action recognition and detection. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 4325–4334 (2017)
Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7794–7803 (2018)
Weinzaepfel, P., Martin, X., Schmid, C.: Towards weakly-supervised action localization, vol. 2. arXiv preprint arXiv:1605.05197 (2016)
Zhao, H., Gan, C., Rouditchenko, A., Vondrick, C., McDermott, J., Torralba, A.: The sound of pixels. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 570–586 (2018)
Zhao, J., Snoek, C.G.: Dance with flow: two-in-one stream action detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9935–9944 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Arnab, A., Sun, C., Nagrani, A., Schmid, C. (2020). Uncertainty-Aware Weakly Supervised Action Detection from Untrimmed Videos. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, JM. (eds) Computer Vision – ECCV 2020. ECCV 2020. Lecture Notes in Computer Science(), vol 12355. Springer, Cham. https://doi.org/10.1007/978-3-030-58607-2_44
Download citation
DOI: https://doi.org/10.1007/978-3-030-58607-2_44
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-58606-5
Online ISBN: 978-3-030-58607-2
eBook Packages: Computer ScienceComputer Science (R0)