Abstract
Predicting human’s gaze from egocentric videos serves as a critical role for human intention understanding in daily activities. In this paper, we present the first transformer-based model to address the challenging problem of egocentric gaze estimation. We observe that the connection between the global scene context and local visual information is vital for localizing the gaze fixation from egocentric video frames. To this end, we design the transformer encoder to embed the global context as one additional visual token and further propose a novel global–local correlation module to explicitly model the correlation of the global token and each local token. We validate our model on two egocentric video datasets – EGTEA Gaze + and Ego4D. Our detailed ablation studies demonstrate the benefits of our method. In addition, our approach exceeds the previous state-of-the-art model by a large margin. We also apply our model to a novel gaze saccade/fixation prediction task and the traditional action recognition problem. The consistent gains suggest the strong generalization capability of our model. We also provide additional visualizations to support our claim that global–local correlation serves a key representation for predicting gaze fixation from egocentric videos. More details can be found in our website (https://bolinlai.github.io/GLC-EgoGazeEst).
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Recent findings in cognitive science have validated the capability of eye movements in reflecting the cognitive processes of human (Yarbus, 2013), which are essential for understanding human intention, modeling interactions across a group of people as well as reasoning daily activities in various scenarios (Hayhoe & Ballard, 2005). Recently, more attention has been paid to egocentric gaze behavior modeling (Huang et al., 2018; Li et al., 2021; Huang et al., 2020; Liu et al., 2020; Thakur et al., 2021; Zhang et al., 2018). Such an understanding of visual attention and intention from the first-person perspective can be valuable for many applications, including Augmented Reality (AR), Virtual Reality (VR), and Human-Robot Interaction (HRI). However, how to measure human’s gaze remains a key challenge in this field.
While wearable eye trackers are a standard way to obtain measurements of gaze behavior, they require calibration, consume significant power, and add substantial cost and complexity to wearable platforms. Alternatively, prior works (Li et al., 2013, 2021; Huang et al., 2018; Soo Park & Shi, 2015; Huang et al., 2020; Tavakoli et al., 2019; Thakur et al., 2021; Al-Naser et al., 2019; Huang et al., 2020; Naas et al., 2020) seek to estimate the visual attention of the camera wearer from videos captured from a first-person perspective. In this paper, we address this challenging task of egocentric gaze estimation. Moreover, we introduce a novel task of predicting whether there is gaze saccade within the given egocentric videos. This novel task serves as a key step for understanding human gaze variation and may promise more power-efficient AR user experience. The problem setting of egocentric gaze estimation and egocentric gaze saccade/fixation prediction are introduced in Fig. 1.
The key challenge in modeling gaze behavior from egocentric videos is to effectively integrate multiple gaze cues into a holistic analysis of visual attention. Cues include the likelihood that different scene objects are gaze targets (i.e. salience), the relative location of gaze targets within the video frame (i.e. center prior), and the patterns of camera movement that are reflective of visual attention (i.e. head motions accompanying a gaze shift). Prior works on visual saliency prediction propose to use two-stream networks (Wang , 2015), dilated convolutional layers (Yang et al., 2019) or pyramid architectures (Hussain et al., 2022) to enlarge the receptive field, yet incorporating a global representation of the input is still missing from the model designs. Recently, the transformer architecture has achieved great success in various vision tasks by modeling the spatio-temporal correlation among local visual tokens (Strudel et al., 2021; Lou et al., 2021; Liu et al., 2021; Fang et al., 2021; Ren et al., 2022; Lee et al., 2022; Patrick et al., 2021; Ma et al., 2022; Li et al., 2022). Vision transformer shows the potential to effectively capture the global representation, since its receptive field can cover the entire input space. However, the pairwise comparisons performed by standard Self-Attention (SA) mechanism is not optimized for interpreting local video features in the context of the global scene. Figure 2 presents the key role of comparisons between local patches and global context - the gaze target is a salient object pointed at by both the camera wearer and another person. Such a salient object can not be easily localized by only modeling the correlation of local patches.
To this end, our paper introduces a novel transformer-based deep model that explicitly embeds global context and calculates spatio-temporal global–local correlation for egocentric gaze estimation. Specifically, we design a transformer encoder that adopts a global visual token embedding strategy to incorporate the global scene context. The single global visual token is handled together with all local visual tokens by standard self-attention layers in the encoder. We then come up with a novel Global–Local Correlation (GLC) module that highlights the connection between global and local visual tokens by masking out all correlations across local tokens. Finally, we adopt a transformer-based decoder to produce gaze prediction output. To the best of our knowledge, this is the first work applying vision transformer to egocentric gaze estimation. As shown in the heatmap from Fig. 2, our model understands the scene in a holistic view and successfully captures the gaze target with the proposed global context embedding and global–local correlation module.
We exhaustively evaluate our approach on two egocentric video datasets – EGTEA Gaze+ (Li et al., 2018) and Ego4D (Grauman et al., 2022). To begin with, we investigate different strategies for global context embedding and show the contribution of involving the global token and the global–local correlation module. We then compare with all prior works of egocentric gaze estimation on the two datasets. We also apply our model to the novel saccade/fixation prediction problem, and traditional action recognition problem to demonstrate the generalization capability of our model. Our proposed model is easy to incorporate into existing transformer-based video analysis architectures, so we implement all experiments with two different transformer backbones. Our method improves the performance of both backbones and finally yields an improvement of +3.9% in F1 score over the previous state-of-the-art method for egocentric gaze estimation. It also boosts the performance on egocentric action recognition and gaze saccade/fixation prediction tasks by a notbale margin. The codes and pretrained models are publicly available to the research community. In summary, this work makes the following contributions:
-
We introduce the first transformer-based approach to address the challenging task of egocentric gaze estimation, and introduce a novel task of gaze saccade/fixation prediction from egocentric video.
-
We utilize a global visual token embedding strategy to incorporate global visual context into self-attention, and further introduce a novel Global–Local Correlation module to explicitly model the correlation between global context and each local visual token.
-
Our novel design obtains consistent improvement on the EGTEA Gaze+ (Li et al., 2018) and Ego4D (Grauman et al., 2022) datasets and outperforms the previous state-of-the-art method by +3.9% on EGTEA and +5.6% on Ego4D in terms of F1 score. Importantly, this is the first work that uses the Ego4D dataset for egocentric gaze estimation, which serves as important benchmark for future research in this direction.
-
We provide more insights of our model by applying it to saccade/fixation prediction and egocentric action recognition tasks. We also visualize correlation weights to show more evidence of the global–local correlation mechanism.
An early version of this paper (Lai et al., 2022) was accepted by BMVC 2022 and then invited this special issue. This paper further extends our previous conference version in several important aspects. First, we introduce a novel task of recognizing saccade gaze movements from the egocentric videos. Second, we conduct additional experiments using recent MotionFormer backbone (Patrick et al., 2021). Our new results suggest that GLC module can be easily plugged into other transformer-based backbones and can produce consistent performance gain on egocentric gaze estimation. Third, we show our model design can also benefit saccade/fixation prediction performance, suggesting global context also contributes to the understanding of rapid gaze movements. Finally, we provide more visualizations of gaze estimation results and correlation weights in GLC module.
This paper is organized in the following order. Section 2 reviews all related works about egocentric gaze estimation, vision transformer and visual saliency modeling, and highlight the difference between prior works and this paper. Section 3 elaborates details of the proposed model. Section 4 presents implementation details for each experiment and the experimental results. Section 5 lists the current limitation of our model and promising future works. Section 6 summarizes all findings of this paper.
2 Related Work
The computational analysis of human gaze behavior is a long-established topic. For example, earlier works consider the problem of eye tracking (Krafka et al., 2016; MacInnes et al., 2018; Ye et al., 2012), which addresses the problem of tracking the gaze movement based on closeup view of human faces or eyes. Moreover, another topic on gaze target prediction(Chong et al., 2020, 2018; Kellnhofer et al., 2019; Nonaka et al., 2022) aims at predicting the gaze target of a subject from the third-person view. In contrast to these prior works, we address the problem of predicting the gaze target directly from egocentric videos captured by wearable cameras. In this paper, we mainly discuss the most relevant prior works on egocentric gaze estimation and related works on transformer-based video representation learning and video saliency prediction.
2.1 Egocentric Gaze Estimation
Previous works focuses on analyzing human daily activities from egocentric videos (Li et al., 2013, 2021; Huang et al., 2018; Liu et al., 2020; Soo Park & Shi, 2015; Huang et al., 2020; Tavakoli et al., 2019; Zhang et al., 2017; Thakur et al., 2021; Al-Naser et al., 2019; Huang et al., 2020; Naas et al., 2020; Jia et al., 2022; Liu et al., 2022). Here, we discuss the most relevant works that develop deep models for egocentric gaze estimation. Zhang et al. (2017) used deep models and an adversarial network to forecast egocentric gaze location in future video frames, which can also be applied to estimate gaze target in current frames by replacing the labels. They further improve this model by adding another branch to incorporate prior information (Zhang et al., 2018). Huang et al. (2018) proposed to explicitly model the temporal attention transition using a LSTM-based architecture and incorporate it into saliency-based models for gaze estimation. Tavakoli et al. (2019) investigated the impact of various factors on egocentric gaze estimation and provided guidance for future work. Another research field is to leverage the relation of human’s action and gaze behavior and model them jointly. Li et al. (2018) sampled a gaze distribution map from the lower layer and used it to selectively pool the features learned by the higher layer. Inspired by this work, Huang et al. (2020) introduced a multi-stream network to enable gaze and action to serve as contexts for each other.
In addition, there exit many works about the variants of egocentric gaze estimation which expand its applications in various scenarios. Soo Park and Shi (2015) introduced the novel problem of predicting joint attention during social interaction using egocentric videos. Huang et al. (2020) collected a new egocentric video dataset and developed a graphical model to detect joint attention. Thakur et al. (2021) proposed a multi-modal network that uses both video and inertial measurement unit data for more accurate egocentric gaze estimation. Naas et al. (2020) developed a tiling scheme for gaze prediction which enables a more efficient VR content delivery.
All of these prior works did not embed global context explicitly or model the connection between local and global visual representations as in our model, which could limit the capability of their models. Additionally, we are the first to develop a transformer-based architecture to address the problem of egocentric gaze estimation.
2.2 Vision Transformer
Transformer architecture is first proposed by Vaswani et al. (2017) and inspires many large language models (Devlin et al., 2018; Liu et al., 2019; Brown et al., 2020). Recently, vision transformers (Dosovitskiy et al., 2022) have demonstrated superior performance on image classification (Dai et al., 2021; Liu et al., 2021; Wang et al., 2021; Ren et al., 2022; Yang et al., 2021; Lee et al., 2022), detection (Dai et al., 2021; Carion et al., 2020; Dai et al., 2021; Fang et al., 2021), segmentation (Strudel et al., 2021; Wang et al., 2021; Zheng et al., 2021; Cheng et al., 2022; Zhang et al., 2022), saliency prediction (Ma et al., 2022; Lou et al., 2021; Liu et al., 2021) and video analysis (Arnab et al., 2021; Neimark et al., 2021; Patrick et al., 2021; Fan et al., 2021; Li et al., 2022; Bertasius, Wang and Torresani, 2021; Wang & Torresani, 2022; Liu et al., 2022). In this section, we focus on reviewing previous works that use vision transformers for pixel-wise visual prediction and video understanding. More related works of saliency prediction is elaborated in Sect. 2.3.
Strudel et al. (2021) developed the first transformer-based architecture for semantic segmentation. Cheng et al. (2022) further unified semantic, instance, and panoptic segmentation in one transformer architecture. Ma et al. (2022) expanded transformers to visual saliency forecasting by using self-attention to capture the correlation between past and future frames. Liu et al. (2021) built a transformer-based model to detect salient objects on RGB-D images. They fused the embeddings of the two modalities by using query from RGB frames and key and value from depth images.
In terms of video transformer, Bertasius et al. (2021) proposed TimeSformer for video action recognition, which is the first transformer-based architecture for video understanding. A similar idea was also explored by Arnab et al. (2021). They downsampled the resolution of input video segment by multiple steps before feeding it into transformer layers. Fan et al. (2021) designed a multiscale video transformer balancing computational cost and action recognition performance. This architecture was further improved by rearranging the layers in each transformer block (Li et al., 2022). Patrick et al. (2021) proposed the trajectory attention mechanism to track the same object in each video frame. Liu et al. (2022) extended the 2D swin-transformer (Liu et al., 2021) to a 3D architecture for action recognition.
Inspired by these successful applications of transformer architectures, we present the first work that uses a vision transformer to address the challenging task of egocentric gaze estimation. In addition, we introduce the novel Global–Local Correlation (GLC) module that provides additional insight into video representation learning with self-attention. We implement this module on two video transformer backbones (Fan et al., 2021; Patrick et al., 2021) and conduct thorough experiments in this paper.
2.3 Visual Saliency
Visual saliency prediction has been well studied in computer vision in recent years (Pan et al., 2017; Wang et al., 2017; Che et al., 2019; Wu et al., 2020; Kroner et al., 2020; Jia & Bruce, 2020; Sun et al., 2022; Lou et al., 2021; Wang et al., 2021; Chen et al., 2021; Khattar et al., 2021; Bellitto et al., 2021; Tsiami et al., 2020; Jiang et al., 2022). Kruthiventi et al. (2017) developed a deep neural network with various kernel sizes to capture saliency features at different scales. Liu et al. (2018) calculated the relation weights between each pixel and the remaining pixels to embed the most relevant contextual features. Zhuge et al. (2022) improved the integrity of detected saliency objects by using integrity channel enhancement mechanism and part-whole verification module. In terms of saliency prediction in videos, Wang et al. (2017) expanded image saliency models to videos by incorporating a new branch to handle temporal information. Wu et al. (2020) proposed SalSAC, which shuffles features of different CNN layers and feeds them to a correlation-based ConvLSTM. Wang et al. (2021) used multiple spatio-temporal self-attention modules to address the limitation of fixed kernel size in 3D models and to model long-range temporal dependencies. Chen et al. (2021) decomposed video saliency prediction into spatial pattern capture and spatio-temporal reasoning. Lou et al. (2021) combined a convolutional network and transformer architecture to model the long-range spatial context. Liu et al. (2023) proposed short-global and long-local attention mechanisms to integrate contexts from neighboring frames.
While visual saliency prediction localizes interesting spatial regions as potential attention targets, egocentric gaze estimation seeks to determine the gaze target of the camera wearers as they interact with a scene. In saliency prediction, cameras typically keep stable and move slowly and salient objects could dominate the view. However, the scene context captured from egocentric video is complex and rapidly changing, which requires a gaze estimation model with the ability of explicitly reasoning about the correlation between local visual features and global scene context. In our experiment section, we demonstrate that our proposed GLC module can significantly benefit gaze estimation performance under this challenging setting.
3 Method
Given an input egocentric video clip with fixed length T and spatial dimension \(H\times W\), our goal is to predict the gaze location in each video frame. Following Li et al. (2018), we consider the gaze prediction as a probabilistic distribution defined on the 2D image plane.
Figure 3 presents an overview of our proposed method. We use the recent multi-scale video transformer (MViT) (Fan et al., 2021) or MotionFormer (Patrick et al., 2021) architecture as the backbone network for video representation learning. We extend the backbone by designing the Visual Token Embedding Module to generate the spatio-temporal tokens of both local visual patches and global visual context and feed them into the standard Multi-Head Self-Attention Module. We then utilize a novel Global–Local Correlation (GLC) Module to explicitly model the correlation between global and local visual tokens for gaze estimation. Finally, we make use of the Decoder Network to predict the gaze distribution based on the learned video representation from the GLC module.
3.1 Transformer Encoder with Global Visual Token Embedding
Visual Token Embedding. We split the input video sequence into non-overlapping patches with size \(s_T\times s_H\times s_W\) and adopt a linear mapping function to project each flattened patch into D-dimension vector space. Following MViT (Fan et al., 2021), this is equivalent to a convolutional layer with a stride of \(s_T\times s_H\times s_W\) and a number of output channels of D. This operation results in N tokens where \(N=\frac{T}{s_T}\times \frac{H}{s_H}\times \frac{W}{s_W}\). In addition, the learnable positional embedding \(\varvec{E}\in \mathbb {R}^{N\times D}\) is added to the local tokens. Our key insight is to further embed global information into a global visual token using convolutional operations, as illustrated in Fig. 3a. Since there is a single global token, it does not require positional embedding.
In our experiments, we examine four global visual embedding strategies as demonstrated in Fig. 4. We (a) implement max pooling on input frames directly, and (b) implement max pooling on unflattened local visual tokens. For (c) and (d), we replace max pooling operations in (a) and (b) with a sequence of convolutional layers. Specifically, for global embedding in (d), we use three additional layers to downsample unflattened local tokens to produce a single global token. In (c), input video frames are first fed into a convolutional layer that is identical to the layer used for local token embedding. Then, the output is passed to a sequence of convolutional layers identical to (d). The experimental results of the four strategies are reported in Sect. 4.2.1 and Table 1. The strategy (d) provides the best gaze estimation performance in our experiments and we thereby use (d) in the final version of our model.
Multi-Head Self-Attention Module. The N local tokens and one global token are fed into a transformer encoder consisting of multiple self-attention blocks. The number of local tokens is downsampled after each self-attention block, while the number of global tokens remains 1. Suppose the input of the j-th layer of encoder is \(\varvec{X}_e^{(j)}=[\varvec{x}_i^{(j)}]_{i=1}^{N_j+1}\in \mathbb {R}^{(N_j+1)\times D_j}\), where \(N_j\) is the number of local tokens, \(D_j\) is the vector length of each token and \(\varvec{x}_i^{(j)}\) is the i-th row of \(\varvec{X}_e^{(j)}\) denoting the i-th token of size \(1\times D_j\). For simplicity, we omit subscript and superscript of j and multi-head operations in the following equations. In each self-attention layer, correlations are calculated in each token pair as shown in Fig. 3b. They are used to reweight values of each token after softmax. Formally, we denote the query, key and value matrices of each self-attention layer in an encoder block as \(\varvec{Q}_e^{(N+1)\times D}=[\varvec{q}_i]_{i=1}^{N+1}\), \(\varvec{K}_e^{(N+1)\times D}=[\varvec{k}_i]_{i=1}^{N+1}\) and \(\varvec{V}_e^{(N+1)\times D}=[\varvec{v}_i]_{i=1}^{N+1}\). The self-attention in transformer encoder is formulated as
Finally, we attach a standard linear layer after the self-attention operation.
3.2 Global–Local Correlation
Even though global information has been explicitly embedded into the global visual token in our model, the transformer encoder treats the global and local tokens equivalently as shown in Eq. 1 and Fig. 3b. In this case, global–local correlation is diluted by correlations among the local tokens, limiting its impact on gaze estimation. In order to address this problem, we propose to increase the available capacity to model global–local token interactions. Our solution is a novel Global-Local Correlation module described in Fig. 3c.
Formally, we denote the global token as the first row vector of \(\varvec{X}_e\), \(i.e. , \varvec{x}_1\). Thus \(\varvec{q}_1\), \(\varvec{k}_1\) and \(\varvec{v}_1\) are the query, key and value projected from the global token, respectively. To explicitly model the connection between global and local visual features, we only calculate the correlation between each local token and the global token, i.e., \(Correlation(\varvec{x}_i, \varvec{x}_1)\), as well as its self-correlation, i.e., \(Correlation(\varvec{x}_i, \varvec{x}_i)\). Then correlation scores are normalized by softmax to further re-weight the values. We exploit a suppression matrix (Liu et al., 2021) \(\varvec{S}^{(N+1)\times (N+1)}\) to suppress the correlation of other tokens, where
We assign zeros to the diagonal and the first column in \(\varvec{S}\) and set a large value \(\lambda \) for the other elements. We follow the empirical choice from the implementation of Liu et al. (2021) and set \(\lambda =10^8\) in our experiments. Formally, the proposed GLC can be formulated as
In this way, we keep the values on the first column and the diagonal, and map them into probability distributions, while values in other positions are nearly “masked out” after the softmax. Residual connections and linear layers are also used in the GLC module as in the regular self-attention block. Finally, the output tokens from the GLC are concatenated with those from the transformer encoder in the channel dimension. We denote outputs of the GLC and the last encoder block as \(\varvec{X}_e^{GLC}\in \mathbb {R}^{(N+1)\times D}\) and \(\varvec{X}_e^{SA}\in \mathbb {R}^{(N+1)\times D}\). The concatenation can then be formulated as \(\varvec{X}_e = \varvec{X}_e^{SA} \oplus \varvec{X}_e^{GLC} \in \mathbb {R}^{(N+1)\times 2D}\). The fused tokens \(\varvec{X}_e\) are subsequently fed into the transformer decoder for gaze estimation.
3.3 Transformer Decoder
To produce the gaze distribution with the desired spatio-temporal resolution, we adopt a decoder to upsample the encoded features. We utilize a transformer decoder based on the multiscale self-attention block of MViT (Fan et al., 2021). Suppose each decoder layer takes visual features \(\varvec{X}_d\in \mathbb {R}^{T'H'W'\times D'}\) as inputs and the corresponding query, key and value matrices are \(\varvec{Q}_d^{T'H'W'\times D'}\), \(\varvec{K}_d^{T'H'W'\times D'}\) and \(\varvec{V}_d^{T'H'W'\times D'}\). As shown in Fig. 3d, we replace the original pooling operation for the query matrix with an upsampling operation implemented with trilinear interpolation and keep the pooling for the key and value matrices. Following Fan et al. (2021), \(\widehat{\varvec{Q}}_d\) is obtained by applying a deconvolutional operation on \(\varvec{Q}_d\), while \(\widehat{\varvec{K}}_d\) and \(\widehat{\varvec{V}}_d\) are obtained by applying convolutional operations on \(\varvec{K}_d\) and \(\varvec{V}_d\). Then, the output of self-attention is calculated in the same way as Eq. 1. In addition, we keep the skip connection in the self-attention layers and replace the pooling operation in skip connections with trilinear interpolation, which produces the upsampled output with dimension \(\widehat{T}\widehat{H}\widehat{W}\times D'\). Our decoder is composed of 4 decoding blocks. Skip connections are used to combine intermediate features of the encoder with corresponding decoder features. Finally, another linear mapping function is used to output the final gaze prediction.
3.4 Network Architecture and Model Training
We adopt MViT (Fan et al., 2021) and MotionFormer (Patrick et al., 2021) as the backbones, with weights initialized from Kinetics-400 pretraining (Kay et al., 2017). The GLC module and decoder are initialized with Xavier initialization (Glorot and Bengio, 2010). For MViT, the token embedding stride is set as \(s_T=2\), \(s_H=4\) and \(s_W=4\) and the embedding dimension is \(D=96\). The encoder is composed of 16 self-attention layers that are divided into 4 blocks. The number of tokens is downsampled at the transition between two blocks. For MotionFormer, the token embedding stride is set as \(s_T=2\), \(s_H=16\), \(s_W=16\) and the embedding dimension is \(D=768\). The encoder consists of 12 layers with trajectory self-attention. The number of tokens doesn’t change in the encoder. We build the decoder with 4 decoder blocks corresponding to the 4 blocks in the encoder. After getting raw output from decoder, softmax is applied on each frame with a temperature \(\tau \). This can be formally written as \(\hat{p}_{ij} = \frac{\exp (\hat{y}_{ij} / \tau )}{\sum _{i,j}\exp ({\hat{y}_{ij} / \tau )}}\) where \(\hat{y}_{ij}\) is the logit at location (i, j) from the model and \(\hat{p}_{ij}\) is probability after softmax. In experiments, \(\tau \) is empirically set as 2. We use KL-divergence loss to capture the difference between labels and predictions. The model is trained using AdamW (Loshchilov & Hutter, xxxx) optimizer with a batch size of 16. We adopt a warm-up training strategy that increases learning rate from \(10^{-6}\) to \(10^{-4}\). Then the learning rate decreases in compliance with cosine annealing scheme (Loshchilov & Hutter, 2016).
4 Experiment
In this section, we show the experimental setup and detailed results. We first elaborate the two datasets used in our experiments, evaluation metrics and data processing details. Second we show exhaustive ablation studies for egocentric gaze estimation and compare with prior works. Third, we validate the generalization capability of our model by applying it to gaze saccade/fixation prediction and egocentric action recognition. Finally, we visualize the predictions and correlation weights in GLC module to provide more insights.
4.1 Datasets and Metrics
Datasets. We conducted our experiments on two egocentric video datasets with gaze tracking data serving as ground truth – EGTEA Gaze+ (Li et al., 2018) and Ego4D (Grauman et al., 2022). The EGTEA Gaze+ dataset is captured under the meal preparation setting, which involves a great deal of hand-object interactions. We used the first train/test split from EGTEA Gaze+ in our experiments (8299 clips for training and 2022 clips for testing). The Ego4D dataset includes 27 videos of 80 participants totaling 31 h with gaze tracking data captured under the social setting. We split the long videos into 5-second video clips and pick clips containing gaze fixation. We used 20 videos (15,310 clips) for training and the other 7 videos (5202 clips) for testing. Note that we keep using the same train/test split for all the three tasks – egocentric gaze estimation, gaze saccade/fixation prediction and action recognition. Importantly, this is the first work that uses the Ego4D dataset for egocentric gaze estimation, and we have made our split publicly available to drive future research on this topic.
Evaluation Metrics. Following Li et al. (2018, 2021) (the source of the EGTEA Gaze+ dataset), we adopt F1 score, recall, and precision as the evaluation metrics for gaze estimation. Note that we do not consider AUC score as our main metrics, since AUC performance can become saturated due to the long-tailed nature of the distribution of gaze in a single frame. In terms of saccade/fixation prediction, we primarily measure the performance by average F1 (average of F1 scores of the two categories) and mean class average (following Li et al. (2021)) because of the imbalance of saccade and fixation, but regular accuracy metric is also provided for reference. For action recognition, we directly follow prior works (Li et al., 2021; Hao et al., 2022) and adopt top-1 accuracy, top-5 accuracy and mean class accuracy.
Data Processing. At training time of egocentric gaze estimation, we randomly sample 8 frames from each video with a sampling interval of 8 as input (i.e. selecting 8 frames from a 72-frame window with equal spacing). All videos are spatially downsampled to 256 in height while keeping the original aspect ratio. We further implement multiple data augmentations including random flipping, shifting, and resizing. We then randomly crop each frame to get an input with dimensions \(8\times 256\times 256\). The output from the decoder is a downsampled heatmap with dimension \(8\times 56\times 56\). For visualization, the output heatmap is upsampled to match the input size by trilinear interpolation. At inference time, the input clip is center-cropped. For gaze labels, we generate a gaussian kernel centered at the gaze location in each input frame with a kernel size of 19 following Chong et al. (2020). We use a uniform distribution for frames where gaze is not tracked in training and only calculate metrics on frames with fixated gaze in testing as in the work of Li et al. (2018). For the EGTEA Gaze+ (Li et al., 2018) dataset, we determine which frames to calculate metrics on by using the provided label of gaze fixations and saccades. On the Ego4D (Grauman et al., 2022) dataset, no label of gaze type is available. We calculate the euclidean spatial distance of gaze between adjacent frames and consider the tracked gaze to be a saccade if the distance is above a threshold, and treat it as fixation otherwise. We adopt an empirical threshold of 40.
In terms of gaze saccade/fixation prediction, we adopt the same data processing settings as gaze estimation. We aggregate the frame-level labels of gaze type to get the label for each video segment. Specifically, the percentages of saccade frames account for 27% and 17% on EGTEA Gaze+ and Ego4D, respectively. The video segment is labeled as saccade if any sampled frame is annotated as saccade. Otherwise, it’s labeled as gaze fixation. Consequently, the ratio of saccade and fixation is 4:1 on EGTEA Gaze+ and 2:1 on Ego4D.
As for egocentric action recognition, we only implement experiments on EGTEA Gaze+ because Ego4D doesn’t provide action labels. The data processing procedures are identical to gaze estimation except that we set the input dimensions as \(8\times 224\times 224\) during training. In testing, the input dimension is \(8\times 256\times 256\) following Li et al. (2021). We also adopt more data augmentation including MixUp (Zhang et al., 2017), color jittering and random erasing.
4.2 Experimental Results on Egocentric Gaze Estimation
4.2.1 The Design Choice of Global Visual Embedding
We introduce four global context embedding strategies in Sect. 3.1 and Fig. 4. We investigate the performance of these strategies on MViT model (Fan et al., 2021). As shown in Table 1, all four global embedding strategies improve the performance of vanilla MViT model on both the EGTEA dataset and the Ego4D dataset. This result supports our claim that global context is essential for gaze estimation. Among the four embedding strategies, (d) achieves the largest performance improvement on both datasets (+0.9% on EGTEA and +0.8% on Ego4D). This indicates that convolutional layers and the embedded local tokens can facilitate the learning of global context. Thus, we use this strategy in the following experiments. Note that all baseline methods use the same transformer decoder.
4.2.2 Evaluation of Global–Local Correlation
We also evaluate the Global–Local Correlation (GLC) module of our model. As presented in Table 1, our full model – MViT+(d)+GLC outperforms the baseline MViT by +1.8% on EGTEA dataset and +2.2% on Ego4D dataset. Specifically, the GLC module contributes to a performance gain of +0.9% on EGTEA and +1.4% on Ego4D (comparing to MViT+(d)). This result suggests that the GLC can break down the mathematical equivalence of global and local tokens in regular self-attention, thereby “highlighting” the global–local connection in the learned representation.
Does the Performance Improvement Come from Additional Parameters? It is possible that the performance of our model benefits from additional parameters in the GLC module. In Table 1, we report the results of another baseline model — MViT+(d)+SA., where we remove the GLC module and add a regular self-attention (SA) layer at the same location. Interestingly, the additional SA layer has minor influence on the overall performance (+0.2% on EGTEA and +0.4% on Ego4D). In contrast, our model outperforms this baseline by +0.7% on EGTEA and +1.0% on Ego4D. This result indicates that the performance boost of our method does not simply come from the additional parameters of GLC. Instead, the explicit modeling of the connection between global and local visual features is the key factor in the performance gain. On the other hand, the regular SA layer includes both global–local correlations and local correlations while the proposed GLC module only calculates global–local correlations. The results suggest that local correlations may dilute the global context and thus limit the performance.
4.2.3 Comparison with Previous State-of-the-Art
In addition to these studies to evaluate the components of our model, we compare our approach with prior works. Apart from MViT (Fan et al., 2021), we also plug the global embedding and GLC modules in another transformer-based architecture – MotionFormer (Patrick et al., 2021). Results are presented in Table 2 and Table 3. Note that, for Attention Transition (Huang et al., 2018), I3D-R50 (Feichtenhofer et al., 2019), MotionFormer (Patrick et al., 2021) and MViT (Fan et al., 2021) from Table 2 and all baselines from Table 3, we initialize the model parameters using pretrained checkpoints from Kinetics (Kay et al., 2017) and finetune the models using the same training set as our method. Interestingly, the baseline MViT and MotionFormer easily outperform all previous works that use CNN-based architectures on both the EGTEA dataset and the Ego4D dataset. In addition, our method implemented on MotionFormer (GLC-MotionFormer) outperforms the best CNN model by +2.3% on F1, +1.8% on recall and +2.3% on precision for EGTEA, and +3.5% on F1, +4.3% on recall and +2.8% on precision for Ego4D. The improvement is more prominent with MViT as backbone (GLC-MViT). It surpasses the best CNN model by +3.9% on F1, +4.0% on recall and +3.5% on precision for EGTEA, and +5.6% on F1, +4.5% on recall and +5.5% on precision for Ego4D. These results demonstrate the superiority of using a transformer-based architecture for egocentric gaze estimation as well as the effectiveness and robustness of our proposed method.
We can also observe MotionFormer lags behind MViT by a large margin. This is because MotionFormer directly downsamples the spatial resolution of video frames by 16 in the visual token embedding module, while MViT adopts a multi-scale downsample strategy. The aggressive reduction in spatial dimension keeps high-level semantic information but loses low-level spatial features. Nonetheless, our method can still boost the performance of MotionFormer prominently on the two datasets (+1.1% on EGTEA and +2.5% on Ego4D). It suggests the proposed method can work as an easy-to-use plug-in for other transformer-based models and brings notable gains.
Moreover, We note that the improvement of our model is more prominent on Ego4D than EGTEA. We speculate that this is because the Ego4D videos with gaze tracking data are captured under social interaction scenarios that contain interactions with both people and objects, and thus require the model to more heavily consider the global–local connections (e.g. the visual information about a social partner’s gesture to an object) to predict the gaze. Another possible reason is that the Ego4D dataset has more samples to train the transformer-based model.
4.2.4 Remarks
Visualization of Predictions. We visualize predictions of our model and other previous methods in Fig. 5. Attention transition (Huang et al., 2018) tends to overestimate gaze area which includes more uncertainty and ambiguity. I3D-R50 (Feichtenhofer et al., 2019), vanilla MotionFormer (Patrick et al., 2021) and vanilla MViT (Fan et al., 2021) architectures run into failure modes when there are multiple objects and people in the scene. In contrast, our model, by explicitly modeling the connection between the global and local visual tokens, more robustly predicts the egocentric gaze distribution from the input video clip. We also illustrate examples of failure cases of our model in Fig. 6. Predicting gaze target near the boundary of the frame or in a scene without enough evidence to infer the gaze target remains a challenging problem for our model as well as prior methods.
What has been learned by the Global–Local Correlation module? We additionally empirically analyze our proposed GLC module. We first calculate the correlation of the global token and each local token, and then normalize the calculated weights into a probabilistic distribution. A higher score suggests that the GLC captures a stronger connection between the particular local token and the global context. We reshape and upsample these weight distributions to form a heatmap, which we overlay with the original input. Since the GLC module applies a multi-head operation, we visualize the results from different heads in Fig. 7. Interestingly, the correlations captured by the GLC heads are quite diverse. Specifically, on the EGTEA dataset, the maps produced by heads 1, 4, 5, and 8 highlight pixels around the gaze point with different uncertainty (which is illustrated by the size of highlighted area). The other four heads focus on surrounding objects and leave gaze areas unattended. As for the Ego4D data, only head 3 captures the wearers’ attention, while the other heads fully focus on the backgrounds in different aspects. This suggests that our GLC module does learn to model human attention by setting different weights from local to global tokens, capturing many facets of scene information (both around the gaze target and in the background) in the multi-headed attention mechanism. Another important finding is that some heads learn to attend to background pixels to prevent the model from omitting important scene context.
4.3 Experimental Results on Gaze Saccade/Fixation Prediction
Apart from gaze estimation, we demonstrate the capability of our model in capturing the feature of gaze fixation and saccade. We use the same backbone as egocentric gaze estimation but replace the decoder with a linear layer for binary classification. Binary cross-entropy loss is adopted in training and the loss weight for saccade and fixation is set as 1:2 to balance the two categories. Results are presented in Table 4.
Vanilla MotionFormer and MViT can both capture the feature of gaze movement on the two datasets. After embedding the global context into a global token, MotionFormer is boosted by +0.6% on EGTEA and +0.6% on Ego4D while MViT is boosted by +0.8% on EGTEA and +1.1% on Ego4D in terms of average F1. The GLC module further improves the performance by a notable margin. Consequently, our method leads to an overall improvement over the MotionFormer baseline by +1.1% on average F1, +1.1% on mean class accuracy and +0.3% on accuracy for EGTEA as well as +1.8% on average F1, +0.9% on mean class accuracy and +2.5% on accuracy for Ego4D. Furthermore, it also improves MViT by +1.6% on average F1, +1.3% on mean class accuracy, +1.7% on accuracy for EGTEA, as well as +3.7% on average F1, +4.4% on mean class accuracy, +3.2% on accuracy for Ego4D. Obviously, gains on Ego4D are more prominent than EGTEA which is consistent with the phenomenon we observed in gaze estimation and can be explained with the similar reasons (see Sect. 4.2.3). We also note that the improvement of MotionFormer is much smaller than MViT. The possible reason is that MotionFormer calculates correlations along the trajectory of each pixel which might ignore some indicators for saccade (e.g. blurry background). The overall improvements further validate the capability of our method in egocentric gaze behavior modeling.
4.4 Experimental Results on Action Recognition
In addition to gaze behavior modeling, we also examine the application of our GLC module to the egocentric video action recognition task, and find that our method performs competitively with methods designed specifically for this task on EGTEA Gaze+. Similar to saccade/fixation prediction, we remove the decoder in the gaze estimation model and keep only the visual token embedding, transformer encoder, and GLC modules. However, we further investigate two different ways to obtain class categories for action recognition: adding a class embedding token at the first layer of transformer, or using pooling across all local tokens to obtain a final embedding. Then a fully-connected layer followed by softmax is used to predict probabilities for each category. We implement both strategies and compare our approaches with previous works in Table 5. We conduct these experiments with two backbones only on EGTEA Gaze+ (Li et al., 2018) using the same split as gaze estimation. Note that the Ego4D (Grauman et al., 2022) social benchmark does not contain action labels.
For vanilla MotionFormer (Patrick et al., 2021) and MViT (Fan et al., 2021), class token embedding performs better than or comparably with the pooling operation. For both strategies, simply adding global embedding to MotionFormer only results in minor gains in the performance. Likewise, adding global embedding to MViT has a minor influence on the overall performance (\(-\)0.2% on top1 accuracy, \(-\)0.5% on top5 accuracy and +1.3% on mean class accuracy while using the class token, and \(-\)0.4%, on top1 accuracy, \(-\)0.2% on top5 accuracy and \(-\)1.1% on mean class accuracy while using pooling layer). This result suggests that simply embedding global context into an additional token has minor influence on the action recognition performance.
In addition, adding our GLC module can only improve the model performance by a small margin when using class token embedding to predict action classes. We hypothesize that this is because only the class token is input into the linear layer for final prediction and re-weighted tokens from GLC are left unused. In contrast, when applying global average pooling on all local tokens, GLC improves top1, top5 and mean class accuracy over the counterpart that doesn’t use GLC (MotionFormer/MViT+Global Token) by +1.7%, +1.4% and +1.7% respectively for MotionFormer and +2.2%, +0.6% and +3.1% respectively for MViT. Gains over corresponding MotionFormer baseline are +1.6%, +3.6% and +3.4% on the tree metrics while gains over MViT baseline are +1.8%, +0.4% and +2.0%. These results indicate our proposed GLC module is a robust and general design that also improves the action recognition performance. However, the impact on action recognition is smaller compared with egocentric gaze estimation because our model doesn’t have any specific design for actoin recognition task.
We note that our model achieves a competitive performance for action recognition on EGTEA without additional design for this specific task. Our top1 accuracy of 65.3% exceeds the model from Wang et al. (2020) by +2.2%, and is only a \(-\)0.2% difference from the recent state-of-the-art method (Hao et al., 2022) for this benchmark of 66.5%. We also want to emphasize that we conduct these action recognition experiments to demonstrate the generalization capability of our proposed GLC module rather than aim to produce SOTA results on action recognition.
Additionally, we visualize the global–local correlation weights of the GLC in Fig. 8. Importantly, the learned global–local correlation is vastly different from the gaze distribution when the model is trained for action recognition; in contrast, a stronger connection between the learned global–local correlation and gaze distribution can be observed when the model is trained for gaze estimation (see Fig. 7). How to design a weakly-supervised model for egocentric gaze estimation remains an open question.
5 Limitations and Future Work
Limitations. Despite the notable gains from global–local correlation, there are still some limitations in our transformer-based method. The model requires larger computational cost, and therefore may not be feasible for on-device computing (e.g. AR/VR). We note that some recent works on network architecture research (Chen et al., 2021) and knowledge distillation(Lin et al., 2022) seek to reduce the computational cost of transformer architecture. These works actually demand a dense model with strong performance as a starting point. Therefore, our work may provide a foundational step for designing light-weight models for the egocentric gaze estimation setting in the future.
Future Work. In this paper, we studied the explicit integration of global scene context for egocentric gaze estimation and proposed a novel modeling approach for this problem. We also showe the results of our proposed architecture on gaze saccade/fixation prediction and egocentric action recognition to demonstrate our model’s generalization cability. Our findings also point to several exciting future research directions:
-
Our proposed GLC module has the potential to address other video understanding tasks including visual saliency prediction in third-person video, active object detection, and future forecasting. We plan to study the effect of our method on those tasks in our future work.
-
Our proposed GLC fails to learn the gaze distribution when the model is trained to predict the action labels. How to design a weakly supervised model for egocentric gaze estimation using action labels is an interesting problem.
-
Our transformer-based model requires larger computational cost, and hence may not be applied for on-device computing. We will continue to study how to combine it with some recent works on network architecture research (Chen et al., 2021) and knowledge distillation (Lin et al., 2022) to reduce the computational cost of transformer architecture.
6 Conclusion
In this paper, we develop a transformer-based architecture to address the task of estimating the camera wear’s gaze fixation based only on egocentric video frames. Our key insight is that our global visual token embedding strategy, which encodes global visual information into the self-attention mechanism, and our global–local correlation (GLC) module, which explicitly reasons about the connection between global and local visual tokens, facilitate strong representation learning for egocentric gaze estimation. Our experiments on the EGTEA Gaze+ and Ego4D datasets demonstrate the effectiveness of our approach. We additionally apply our method to a novel gaze saccade/fixation prediction task and the traditional action recognition problem. The proposed method can improve the performance prominently which shows its strong generalization capability. We also implement the global token embedding strategy and GLC module in two backbones to show it can serve as an easy-to-use plug-in to other transformer-based architecture. We believe our work serves as an essential step in analyzing gaze behavior from egocentric videos and provides valuable insight into learning video representations with transformer architectures.
Data Availability
The data used in the experiments are publicly available online. EGTEA Gaze+ dataset is available via https://cbs.ic.gatech.edu/fpv/. Ego4D dataset is available via https://ego4d-data.org/.
Code Availability
Codes and train/test split for Ego4D are available on https://bolinlai.github.io/GLC-EgoGazeEst/.
References
Al-Naser, M., Siddiqui, S.A., Ohashi, H., Ahmed, S., Katsuyki, N., Takuto, S., & Dengel, A. (2019). Ogaze: Gaze prediction in egocentric videos for attentional object selection. 2019 digital image computing: Techniques and applications (dicta) (pp. 1–8).
Arnab, A., Dehghani, M., Heigold, G., Sun, C., Lučić, M., & Schmid, C. (2021). Vivit: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 6836–6846).
Bellitto, G., Proietto Salanitri, F., Palazzo, S., Rundo, F., Giordano, D., & Spampinato, C. (2021). Hierarchical domain-adapted feature learning for video saliency prediction. International Journal of Computer Vision, 129(12), 3216–3232.
Bertasius, G., Wang, H., & Torresani, L. (2021). Is space-time attention all you need for video understanding?. In International Conference on Machine Learning.
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., & Dhariwal, P. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877–1901.
Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., & Zagoruyko, S. (2020). End-to-end object detection with transformers. In European Conference on Computer Vision (pp. 213–229).
Che, Z., Borji, A., Zhai, G., Min, X., Guo, G., & Le Callet, P. (2019). How is gaze influenced by image transformations? dataset and model. IEEE Transactions on Image Processing, 29, 2287–2300.
Chen, B., Li, P., Li, C., Li, B., Bai, L., Lin, C., & Ouyang, W. (2021). Glit: Neural architecture search for global and local image transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 12–21).
Chen, J., Li, Z., Jin, Y., Ren, D., & Ling, H. (2021). Video saliency prediction via spatio-temporal reasoning. Neurocomputing, 462, 59–68.
Cheng, B., Misra, I., Schwing, A.G., Kirillov, A., & Girdhar, R. (2022). Masked-attention mask transformer for universal image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 1290–1299).
Chong, E., Ruiz, N., Wang, Y., Zhang, Y., Rozga, A., & Rehg, J.M. (2018). Connecting gaze, scene, and attention: Generalized attention estimation via joint modeling of gaze and scene saliency. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 383–398).
Chong, E., Wang, Y., Ruiz, N., & Rehg, J.M. (2020). Detecting attended visual targets in video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 5396–5406).
Dai, X., Chen, Y., Yang, J., Zhang, P., Yuan, L., & Zhang, L. (2021). Dynamic detr: End-to-end object detection with dynamic attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 2988–2997).
Dai, Z., Cai, B., Lin, Y., & Chen, J. (2021). Up-detr: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 1601–1610).
Dai, Z., Liu, H., Le, Q. V., & Tan, M. (2021). Coatnet: Marrying convolution and attention for all data sizes. Advances in Neural Information Processing Systems, 34, 3965–3977.
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. (2022). An image is worth 16x16 words: Transformers for image recognition at scale. Iclr.
Fan, H., Xiong, B., Mangalam, K., Li, Y., Yan, Z., Malik, J., & Feichtenhofer, C. (2021). Multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 6824–6835).
Fang, Y., Liao, B., Wang, X., Fang, J., Qi, J., Wu, R., & Liu, W. (2021). You only look at one sequence: Rethinking transformer in vision through object detection. Advances in Neural Information Processing Systems, 34, 26183–97.
Feichtenhofer, C., Fan, H., Malik, J., & He, K. (2019). Slowfast networks for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 6202–6211).
Glorot, X., & Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (pp. 249–256).
Grauman, K., Westbury, A., Byrne, E., Chavis, Z., Furnari, A., Girdhar, R., et al. (2022). Ego4d: Around the world in 3000 hours of egocentric video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 18995–19012).
Hao, Y., Zhang, H., Ngo, C.-W., & He, X. (2022). Group contextualization for video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 928–938).
Harel, J., Koch, C. & Perona, P. (2006). Graph-based visual saliency. Advances in neural information processing systems. 19.
Hayhoe, M., & Ballard, D. (2005). Eye movements in natural behavior. Trends in Cognitive Sciences, 9(4), 188–194.
Huang, Y., Cai, M., Li, Z., Lu, F., & Sato, Y. (2020). Mutual context network for jointly estimating egocentric gaze and action. IEEE Transactions on Image Processing, 29, 7795–7806.
Huang, Y., Cai, M., Li, Z. & Sato, Y. (2018). Predicting gaze in egocentric video by learning task-dependent attention transition. In Proceedings of the European conference on computer vision (eccv) (pp. 754–769).
Huang, Y., Cai, M., & Sato, Y. (2020). An ego-vision system for discovering human joint attention. IEEE Transactions on Human-Machine Systems, 50(4), 306–316.
Hussain, T., Anwar, A., Anwar, S., Petersson, L., & Baik, S.W. (2022). Pyramidal attention for saliency detection. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (pp. 2877–2887).
Jia, S., & Bruce, N. D. (2020). Eml-net: An expandable multi-layer network for saliency prediction. Image and Vision Computing, 95, 103887.
Jia, W., Liu, M. & Rehg, J.M. (2022). Generative adversarial network for future hand segmentation from egocentric video. In Proceedings of the European Conference on Computer Vision (ECCV).
Jiang, L., Li, Y., Li, S., Xu, M., Lei, S., Guo, Y. & Huang, B. (2022). Does text attract attention on e-commerce images: A novel saliency prediction dataset and method. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 2088–2097).
Kay, W., Carreira, J., Simonyan, K., Zhang, B., Hillier, C., Vijayanarasimhan, S., et al. (2017). The kinetics human action video dataset. arXiv preprint arXiv:1705.06950
Kellnhofer, P., Recasens, A., Stent, S., Matusik, W. & Torralba, A. (2019). Gaze360: Physically unconstrained gaze estimation in the wild. In IEEE International Conference on Computer Vision (ICCV).
Khattar, A., Hegde, S. & Hebbalaguppe, R. (2021). Cross-domain multi-task learning for object detection and saliency estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 3639–3648).
Krafka, K., Khosla, A., Kellnhofer, P., Kannan, H., Bhandarkar, S., Matusik, W., & Torralba, A. (2016). Eye tracking for everyone. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2176–2184).
Kroner, A., Senden, M., Driessens, K., & Goebel, R. (2020). Contextual encoder-decoder network for visual saliency prediction. Neural Networks, 129, 261–270.
Kruthiventi, S. S., Ayush, K., & Babu, R. V. (2017). Deepfix: A fully convolutional neural network for predicting human eye fixations. IEEE Transactions on Image Processing, 26(9), 4446–4456.
Lai, B., Liu, M., Ryan, F., & Rehg, J. (2022). In the eye of transformer: Global-local correlation for egocentric gaze estimation. In British Machine Vision Conference.
Lee, Y., Kim, J., Willette, J., & Hwang, S.J. (2022). Mpvit: Multi-path vision transformer for dense prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.
Li, Y., Fathi, A., & Rehg, J.M. (2013). Learning to predict gaze in egocentric video. In Proceedings of the IEEE International Conference on Computer Vision (pp. 3216–3223).
Li, Y., Liu, M., & Rehg, J. (2021). In the eye of the beholder: Gaze and actions in first person video. In IEEE Transactions on Pattern Analysis and Machine Intelligence.
Li, Y., Liu, M., & Rehg, J.M. (2018). In the eye of beholder: Joint learning of gaze and actions in first person video. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 619–635).
Li, Y., Wu, C.-Y., Fan, H., Mangalam, K., Xiong, B., Malik, J., & Feichtenhofer, C. (2022). Mvitv2: Improved multiscale vision transformers for classification and detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4804–4814).
Lin, S., Xie, H., Wang, B., Yu, K., Chang, X., Liang, X., & Wang, G. (2022). Knowledge distillation via the target-aware transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10915–10924).
Liu, M., Ma, L., Somasundaram, K., Li, Y., Grauman, K., Rehg, J.M., & Li, C. (2022). Egocentric activity recognition and localization on a 3d map. In Proceedings of the European Conference on Computer Vision (ECCV).
Liu, M., Tang, S., Li, Y., & Rehg, J.M. (2020). Forecasting human-object interaction: joint prediction of motor attention and actions in first person video. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 704–721).
Liu, N., Han, J., & Yang, M.-H. (2018). Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 3089–3098).
Liu, N., Nan, K., Zhao, W., Yao, X., & Han, J. (2023). Learning complementary spatial–temporal transformer for video salient object detection. IEEE Transactions on Neural Networks and Learning Systems.
Liu, N., Zhang, N., Wan, K., Shao, L., & Han, J. (2021). Visual saliency transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 4722–4732).
Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., & Stoyanov, V. (2019). Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692
Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., & Guo, B. (2021). Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 10012–10022).
Liu, Z., Ning, J., Cao, Y., Wei, Y., Zhang, Z., Lin, S., & Hu, H. (2022). Video swin transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 3202–3211).
Loshchilov, I., & Hutter, F. (xxxx). Decoupled weight decay regularization. In International Conference on Learning Representations.
Loshchilov, I., & Hutter, F. (2016). Sgdr: Stochastic gradient descent with warm restarts. In International Conference on Learning Representations.
Lou, J., Lin, H., Marshall, D., Saupe, D., & Liu, H. (2021). Transalnet: Visual saliency prediction using transformers. arXiv preprint arXiv:2110.03593
Ma, C., Sun, H., Rao, Y., Zhou, J., & Lu, J. (2022). Video saliency forecasting transformer. In IEEE Transactions on Circuits and Systems for Video Technology.
MacInnes, J.J., Iqbal, S., Pearson, J., & Johnson, E.N. (2018). Wearable eye-tracking for research: Automated dynamic gaze mapping and accuracy/precision comparisons across devices. BioRxiv. 299925
Naas, S.-A., Jiang, X., Sigg, S., & Ji, Y. (2020). Functional gaze prediction in egocentric video. In Proceedings of the 18th International Conference on Advances in Mobile Computing & Multimedia (pp. 40–47).
Neimark, D., Bar, O., Zohar, M., & Asselmann, D. (2021). Video transformer network. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 3163–3172).
Nonaka, S., Nobuhara, S., & Nishino, K. (2022). Dynamic 3d gaze from afar: Deep gaze estimation from temporal eye-head-body coordination. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (cvpr) (p. 2192-2201).
Pan, J., Ferrer, C.C., McGuinness, K., O’Connor, N.E., Torres, J., Sayrol, E., & Giro-i Nieto, X. (2017). Salgan: Visual saliency prediction with generative adversarial networks. arXiv preprint arXiv:1701.01081
Patrick, M., Campbell, D., Asano, Y., Misra, I., Metze, F., Feichtenhofer, C., Henriques, J. F., et al. (2021). Keeping your eye on the ball: Trajectory attention in video transformers. Advances in Neural Information Processing Systems, 34, 12493–12506.
Ren, S., Zhou, D., He, S., Feng, J., & Wang, X. (2022). Shunted self-attention via multi-scale token aggregation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
Soo Park, H., & Shi, J. (2015). Social saliency prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 4777–4785).
Strudel, R., Garcia, R., Laptev, I., & Schmid, C. (2021). Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 7262–7272).
Sun, Y., Zhao, M., Hu, K., & Fan, S. (2022). Visual saliency prediction using multi-scale attention gated network. Multimedia Systems, 28(1), 131–139.
Tavakoli, H.R., Rahtu, E., Kannala, J., & Borji, A. (2019). Digging deeper into egocentric gaze prediction. In 2019 IEEE Winter Conference on Applications of Computer Vision (WACV) (pp. 273–282).
Thakur, S.K., Beyan, C., Morerio, P., & Del Bue, A. (2021). Predicting gaze from egocentric social interaction videos and imu data. In Proceedings of the 2021 International Conference on Multimodal Interaction (pp. 717–722).
Tsiami, A., Koutras, P., & Maragos, P. (2020). Stavis: Spatio-temporal audiovisual saliency network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4766–4776).
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., & Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems. 30
Wang, H., Zhu, Y., Adam, H., Yuille, A., & Chen, L.-C. (2021). Max-deeplab: End-to-end panoptic segmentation with mask transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 5463–5474).
Wang, J., & Torresani, L. (2022). Deformable video transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 14053–14062).
Wang, L., Lu, H., Ruan, X., & Yang, M.-H. (2015). Deep networks for saliency detection via local estimation and global search. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 3183–3192).
Wang, W., Shen, J., & Shao, L. (2017). Video salient object detection via fully convolutional networks. IEEE Transactions on Image Processing, 27(1), 38–49.
Wang, W., Xie, E., Li, X., Fan, D.-P., Song, K., Liang, D., & Shao, L. (2021). Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 568–578).
Wang, X., Wu, Y., Zhu, L., & Yang, Y. (2020). Symbiotic attention with privileged information for egocentric action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, pp. 12249–12256).
Wang, Z., Liu, Z., Li, G., Wang, Y., Zhang, T., Xu, L., & Wang, J. (2021). Spatio-temporal self-attention network for video saliency prediction. IEEE Transactions on Multimedia.
Wu, X., Wu, Z., Zhang, J., Ju, L., & Wang, S. (2020). Salsac: A video saliency prediction model with shuffled attentions and correlation-based convlstm. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, pp. 12410–12417).
Yang, J., Li, C., Zhang, P., Dai, X., Xiao, B., Yuan, L., & Gao, J. (2021). Focal self-attention for local-global interactions in vision transformers. arXiv preprint arXiv:2107.00641
Yang, S., Lin, G., Jiang, Q., & Lin, W. (2019). A dilated inception network for visual saliency prediction. IEEE Transactions on Multimedia, 22(8), 2163–2176.
Yarbus, A. L. (2013). Eye Movements and Vision. Springer.
Ye, Z., Li, Y., Fathi, A., Han, Y., Rozga, A., Abowd, G.D., & Rehg, J.M. (2012). Detecting eye contact using wearable eye-tracking glasses. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing (pp. 699–704).
Zhang, H., Cisse, M., Dauphin, Y.N., & Lopez-Paz, D. (2017). mixup: Beyond empirical risk minimization. In International Conference on Learning Representations.
Zhang, M., Ma, K. T., Lim, J. H., Zhao, Q., & Feng, J. (2018). Anticipating where people will look using adversarial networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(8), 1783–1796.
Zhang, M., Teck Ma, K., Hwee Lim, J., Zhao, Q., & Feng, J. (2017). Deep future gaze: Gaze anticipation on egocentric videos using adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 4372–4381).
Zhang, W., Huang, Z., Luo, G., Chen, T., Wang, X., Liu, W., & Shen, C. (2022). Topformer: Token pyramid transformer for mobile semantic segmentation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12083–12093).
Zheng, S., Lu, J., Zhao, H., Zhu, X., Luo, Z., Wang, Y., et al. (2021). Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the ieee/cvf conference on computer vision and pattern recognition (pp. 6881–6890).
Zhuge, M., Fan, D.-P., Liu, N., Zhang, D., Xu, D., & Shao, L. (2022). Salient object detection via integrity learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(3), 3738–52.
Acknowledgements
Portions of this project were supported in part by NIH R01MH114999 and by a gift from Meta.
Funding
Partial financial support of this project was provided by NIH R01MH114999 and by a gift from Meta.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study design. Data preprocessing, model implementation and experiment design were performed by BL and ML. The first draft of the manuscript was written by BL and it was reviewed and edited by ML, FR and JMR. All authors read and approved the final manuscript.
Corresponding author
Additional information
Communicated by Guang Yang.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lai, B., Liu, M., Ryan, F. et al. In the Eye of Transformer: Global–Local Correlation for Egocentric Gaze Estimation and Beyond. Int J Comput Vis 132, 854–871 (2024). https://doi.org/10.1007/s11263-023-01879-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11263-023-01879-7