Introduction

COVID-19 holds the world in its grip. December 31st 2020, a year after the first statement about a new virus reached the WHO, over 80 million confirmed COVID-19 cases and 1.78 million deaths in 222 countries have been reported1. The virus continues to spread, and many countries suffer from a second wave of infections. As long as the world was awaiting the development, production, and distribution of effective vaccines, we only had one weapon at our disposal in our battle against the Coronavirus: our behavior2.

The World Health Organization (WHO) increasingly and repeatedly stresses the importance of engaging in ‘safety behaviors’, which include washing hands, wearing face masks, and physical distancing. Ultimately, physical distancing is among the most important of the recommended safety behaviors3 because, as complex as the virus may be, transmission of the virus requires an infectious person to be in physical proximity to other people. Therefore, in absence of a vaccine, it is pivotal to reduce the number of contacts between people that are in such physical proximity4.

Despite its universally acclaimed importance, it is unclear how we can optimally stimulate physical distancing through behavioral interventions. Whereas the effectiveness of policy measures to reduce mobility has been investigated (e.g.,5), the effects of more subtle behavioral interventions largely remain a matter of speculation and debate. One of the most striking examples of this debate concerns the behavioral effects of wearing face masks. Although early on in the pandemic extensive experimental research was conducted on the protective function of face masks (e.g., see6 for a review), studies investigating the behavioral effect of wearing face masks on physical distancing remained scarce.

In absence of such experiments, a variety of positions on the behavioral effects of face masks was defended. Policy makers and the World Health Organization expressed their concern that face masks could give people a false sense of security7,8, potentially leading to reduced adherence to other preventive regulations such as physical distancing. The phenomenon that people adjust their behavior in response to the perceived risk is more generally known as the Peltzman effect9, which could explain such compensation behavior. In contrast, others argued that face masks remind people to implement safety behaviors, thereby increasing compliance with COVID-19 regulations (e.g.10). Notably, even after evidence on the medical effects of face masks had accumulated, the debate around the implementation persisted due to the uncertainty of the behavioral effects of face mask policies (e.g.11).

Over the last year some studies have investigated the effect of face masks on physical distancing. Using an experimental set-up it was shown that generally people keep more distance from a confederate when the confederate was wearing a face mask compared to wearing no mask12,13,14, although some evidence was also found for the reverse pattern15. In observational studies it was moreover shown that mandatory wearing of face masks is positively associated to physical distancing10. However, through observational studies it is difficult to disentangle these findings from other confounding factors, such as incidence rates and public health messaging. The current paper adds to the literature on the relation between wearing face masks and physical distancing behavior, specifically by experimentally investigating the effect of wearing a face mask, as opposed to the effect of seeing someone else wear a face mask, on physical distancing.

The debate on mandatory face masks exemplifies how existing scientific work in psychology and behavioral science can provide support for opposing accounts, confusing and occasionally paralyzing policy. At the same time COVID-19 challenges our behavior in ways that have not been anticipated or assessed in previous research, thereby questioning the relevance of existing and behavioral science literature to this novel crisis2,16. This has spurred several investigations of how psychological factors can facilitate a better adherence to safety behaviors, such as the role of emotions17,18, reasoning19, and the effect of public health messaging20,21,22,23. However, it is important to note that most of these studies focus on self-reports and intentions rather than on actual behavior. Additional experiments are especially important as the external validity and generalizability of any one particular experiment is limited, requiring to build up an evidence base consisting of different experiments. The novelty of the COVID-19 pandemic thus warrant only one way forward: to assess the effect of behavioral interventions on physical distancing behavior experimentally.

Investigation of the effect of behavioral interventions on physical distancing is hampered, however, by the lack of suitable methodologies that connect behavior to relevant epidemiological properties. In the current paper we describe, implement, and evaluate a novel methodological framework suited to directly test the effectiveness of different behavioral interventions on physical distancing. The framework is shown in Fig. 1 and involves (1) an experimental design; (2) objective measures of physical distancing; and (3) the representation of people (nodes) and their contacts (links) in a contact network. The proposed methodological framework enables us to evaluate the effectiveness of behavioral interventions on physical distancing through network representations24.

Figure 1
figure 1

Representation and application of the novel methodological framework. The framework consists of (1) an experimental design; (2) objective physical distance measures; and (3) a contact network representation. Here, we show the framework applied at the art fair, where we (1) varied walking directions and supplementary interventions; (2) measured physical distance using a Social Distancing Sensor with ultra-wideband technology; and (3) constructed networks for each condition.

We implemented this novel framework in a first experimental study during a 3-day art fair in Amsterdam. As shown in Fig. 1, in the experimental design we implemented walking direction interventions (bidirectional; unidirectional; no direction) and supplementary interventions (face masks; buzzer feedback when within 1.5 m of another visitor; no supplementary intervention) in 2-h time slots. Note that at the time of the art fair in August 2020, face masks were not mandatory in The Netherlands, allowing us to incorporate face masks as an experimental condition. See “Methods” section for details. We objectively measured physical proximity between visitors using a Social Distancing Sensor (SDS) which utilizes ultra-wideband technology. Whenever two visitors came within 1.5 m from each other, the contact was logged. Subsequently, we used the logged contacts to construct the contact networks of visitors for each experimental condition.

In constructing the networks, we collapsed the contacts over time without losing important information, because only primary infections can be transmitted within a 2-h time frame, and therefore only direct contacts are virologically relevant. In other words, if a visitor is infected during the event, this visitor cannot subsequently infect others due to the incubation period, and as a result the order in which people make contacts can be disregarded. In accordance, here we only focus on the direct paths between nodes in the network, which specify the number of unique contacts between visitors and are encoded in the degree distribution. Finally, we use innovative statistical network methodology, based on an adaptation of Van Duijn25 to undirected networks, which allows us to statistically test differences in the degree distribution (i.e., the distribution of the number of contacts) across experimental conditions.

In short, the proposed methodology enables us to assess the effects of interventions directly at the level of behavior. The methodology is straightforward to implement and provides society with the means to directly evaluate interventions, so that policy can be based on evidence rather than conjecture.

Results

Descriptives

Visitors

Over the course of 3 days 787 people visited the art fair, of whom 639 (81.2%) participated in our study and wore a Social Distancing Sensor, see Table 1. We note that on Saturday 29th and Sunday 30th the art fair opened at 11:00 before the experiment started at 13:30. 74 and 132 visitors, respectively, had entered the art fair between 11:00 and 13:30, of whom some had already left before the experimental condition started. Therefore, some of the 787 visitors did not participate in our studies, and thus the actual percentage of visitors that wore the SDS is higher than the reported 81%. For 392 (61.3%) visitors demographic information was available as they completed a questionnaire upon purchasing their ticket. These visitors were 42 ± 16 (mean ± SD) years old (14–89 years), 225 (57.3%) were female, and 335 (85.5%) of them completed a higher education.

Table 1 Descriptives per condition.

Interventions

From the conditions shown in Table 1, we only selected the six conditions that lasted 2 h (visualized in Fig. 1, see “Methods” for a more elaborate explanation). To evaluate the effectiveness of each behavioral intervention in isolation (face masks, walking directions, buzzer) we select two conditions that differ only in whether the behavioral intervention of interest is implemented, allowing us to estimate the effect of that intervention.

Face masks

To examine the effect of wearing face masks on physical distancing, we compared the contact networks for the face mask condition with the no supplementary intervention condition on the first day (condition 1 vs. 2 in Fig. 1). As can be seen from Table 1, visitors had a median of 8 unique contacts (IQR = 3–14, range = 0–44) in the face mask condition, whereas visitors in the no supplementary intervention condition had a median of 7 unique contacts (IQR = 3–15, range = 0–42). Detailed analysis of the probability of visitors forming contacts in these two conditions—taking the network structure of the data into account (see “Methods”)—did not provide evidence for an effect of wearing face masks on distancing (OR = 1.05, 95% Credible Interval [0.81, 1.33]). These results indicate that, if there are differences in the number of contacts when wearing face masks compared to when not wearing them, these differences are likely to be small. In other words, face masks do not appear to facilitate or inhibit physical distancing.

Walking direction conditions

We performed two comparisons. First, we evaluated the effect of unidirectional walking directions versus no walking directions by comparing the buzzer conditions on day two and three (condition 5 vs. 6). On day 2 with unidirectional walking directions, visitors had a median of 4 unique contacts (IQR = 1–11; range = 0–31), whereas on day 3 without walking directions, visitors had a median of 8 unique contacts (IQR = 5–15; range = 0–38). The probability of visitors forming contacts was substantially higher in the condition with no walking directions compared to the condition with unidirectional walking directions (OR = 1.66, 95% CI [1.25, 2.17]).

Second, we evaluated the effect of bidirectional versus unidirectional walking directions, by comparing the no supplementary intervention conditions on days 1 and 2 respectively (condition 2 vs. 4). During bidirectional walking directions, visitors had a median of 7 unique contacts (IQR = 3–15; range = 0–42), whereas during unidirectional walking directions, visitors had a median of 8 unique contacts (IQR = 4–14; range = 0–48). The probability of visitors forming contacts was about the same in these two conditions (OR = 0.99, 95% CI [0.75, 1.26]). This indicates that, if there are differences in the number of unique contacts visitors had when given bidirectional versus unidirectional walking directions, these are likely to be small. Thus, the implementation of walking directions appears to improve physical distancing, but the differential effects of unidirectional and bidirectional implementations were not statistically significant.

Buzzer

We investigated whether notifying visitors of violations of the 1.5 m rule through a buzzer reduces the number of contacts by comparing the buzzer condition to the no supplementary intervention condition (condition 2 vs. 3). The median number of unique contacts was 10 (IQR = 3–15; range = 0–42) for visitors in the buzzer condition and 7 (IQR = 4–17; range = 0–49) for visitors in the no supplementary intervention condition. The probability of visitors forming contacts was slightly higher in the buzzer condition than in the no supplementary intervention condition (OR = 1.24, 95% CI [0.95, 1.55]). The credible interval around this estimate suggests that buzzers likely do not promote physical distancing, and may actually make it worse. In interpreting this effect, it is however important to note that on this day visitors did not receive an instruction on how the buzzer worked, and feedback occurred only 2 s after the contact was made.

After improvements were made to the buzzer intervention (i.e., by implementing a short demonstration and removing the delay between registration of distance violation and feedback) the buzzer did promote distancing (conditions 5 vs. 4). With adequate instructions and immediate feedback, the median number of unique contacts of visitors in the buzzer condition was 4 (IQR = 1–11; range = 0–31) compared with a median of 8 (IQR = 4–14; range = 0–48) unique contacts in the no supplementary condition. The probability of visitors forming contacts was markedly higher in the no supplementary intervention condition, as compared to the buzzer conditions (OR = 1.43, 95% CI [1.06, 1.91]). This suggests that buzzer feedback can promote physical distancing, but that its effect is dependent on the instructions and immediate feedback.

Discussion

In this paper we introduced and demonstrated a practical methodology to assess the effect of behavioral interventions on physical distancing. The methodology combines systematic variation of interventions in an experimental design with contact network assessment through wearable sensors. Analyzing experimentally induced changes in the contact network allows researchers to systematically track the effect of behavioral interventions and, as such, brings epidemiologically relevant human behavior within reach of the scientific method. Because it uses assessment of intervention effects at the level of behaviors that are known to affect virus transmission, the introduced methodology can be implemented to bridge the gap between behavioral science and epidemiology. Therefore, this type of research can directly inform policy concerning the mitigation of virus spread.

The implementation of the proposed methodology in the current experimental design may have several policy implications. First, consistent with recent findings from observational research26, we observed no significant effects of wearing face masks on physical distancing. This suggests that risk compensation, also known as the Peltzman effect9, is not operative in the case of face masks. Interestingly, however, previous studies showed that seeing someone else wear a face mask does influence physical distancing behavior12,13,14,15. Together, these findings also underscore the need to execute dedicated empirical research in different contexts to evaluate behavioral hypotheses, as it illustrates how prima facie plausible claims about human behavior may miss the mark16. At the same time, it is important to note that while the Peltzman effect may not play a role in mandatory wearing of face masks, it could very well have a profound effect on the course of the pandemic, e.g., through its effects on risk taking after vaccination27.

Second, our analysis clearly shows that implementing walking directions lowers the probability of contacts between visitors in the relevant conditions. We hypothesize that this benefit is counteracted by the fact that the movement patterns of other visitors are relatively unpredictable in a free movement scheme, so that distancing requires more effort and attention, and accidental collisions occur more frequently. The latter hypothesis could be investigated by subjecting movement patterns isolated from camera data28 to quantitative pedestrian modelling29. In comparing different walking directions, we did not observe differential effects of uni- versus bidirectional schemes. Given the dynamics of viral transmission, unidirectional walking directions should be preferred because they mitigate the probability of people facing each other directly; however, in situations where a unidirectional scheme is not available, a bidirectional scheme appears to be a good alternative. The implementation of walking directions is feasible in almost all public spaces and the current study directly supports this policy.

Third, our study opens the possibility of utilizing schemes that use sensor data information to provide people with feedback on their behavior. The current study used a style of feedback most closely related to an alarm: a pattern of light and a buzzer emitted when potential danger is encountered. While this style of feedback does seem to have a positive effect on adherence to distancing guidelines in the current study, its implementation is potentially challenging. Specifically, effective use of sensors seems to depend on adequate instruction and immediate feedback. In the absence of this guidance, on the first day of our study, visitors tended to try out the sensors and as such inadvertently came within 1.5 m of another visitor. This may have obscured the positive effects of feedback on subsequent distancing. After the first day, we countered this effect through a quite elaborate instruction where visitors were invited to try out the sensor on a trial sensor that was not included in the data analysis. However, in many cases such an instruction would be impractical, as it would take time and may lead to queue formation. One last consideration is that wearing a sensor, even without active feedback, could make people aware that they are being measured, thereby influencing their behavior, also known as the mere-measurement effect30.

Interestingly, in the current study the alarm scheme setup warns, but does not reward individuals for keeping distance. Other more rewarding ways to organize feedback may be more engaging (e.g., using gamification by letting people win various awards if they keep their distance) and informative for the user (e.g., many people asked for feedback on how many contacts they made). Clearly, there is a great potential for learning theory (e.g., reinforcement learning) to inform this process31; next to encouraging distancing upon receiving feedback, a feedback system could promote learning to distance effectively and automatically/efficiently.

The present study has yielded experimental support for policy advise concerning physical distancing. However, some limitations warrant attention. First, as we conducted the experiment in a naturalistic setting, we were not in a position to randomize the visitors to time slots. Since participants were not aware of the interventions that were implemented in each time slot, and as the exhibition in the art-fair was the same for all days, a priori differences between visitors are not expected; yet, in future research it is important to extend the design to experimental setups that utilize random assignment to bolster causal conclusions. Second, as we were unable to counterbalance all conditions, the design may yield some confounds; for example, based on the time of day a condition was measured. Third, as visitors could stay as long as they liked, some visitors were part of multiple conditions creating some dependency in the data. Fourth, as the experiment was conducted at an art fair, the visitors are not a random sample from the population and the results may not generalize to the general population or other settings. Similarly, the effectiveness of certain measures implemented at an art fair, may not readily translate to other contexts, such as for example a supermarket. At the same time, a naturalistic design has the advantage of high ecological validity. Ultimately, as different contexts draw different crowds and behavior is affected by one’s environment, we should replicate the interventions across various contexts (e.g., in a supermarket, or at a football stadium) to determine which interventions work best in which contexts.

In conclusion, the presented research methodology has a broad range of applications and can be used to assess behavioral interventions quickly and effectively. The current study provides a template for how such research can be organized. Our analysis code and data (reported separately in Tanis et al.28) are available for use in future studies (https://osf.io/u9avy/). We hope that this framework will stimulate the behavioral science community to engage with the problem of mitigating virus spread through behavioral interventions.

Methods

Participants

The participants in our naturalistic study were visitors of Smart Distance Lab: The Art Fair (www.smartdistancelab.nl). Upon purchasing their ticket, participants were informed that a study would take place during the art fair and that different interventions would be implemented. It was not told, however, which interventions would be implemented when. Participants could select a 30 min time slot themselves (see procedure), which allowed us to ensure that the visitors were sufficiently distributed across experimental conditions. As participants could select their own time slot, the design is a natural experiment rather than a randomized controlled trial.

Design

The experiment took place at an art fair in De Kromhouthal in Amsterdam, see Tanis et al.28 for specifics on the layout. Across experimental conditions, we varied two sets of interventions: walking directions and supplementary interventions. Walking directions varied across the 3 days: bidirectional, unidirectional, and no walking directions respectively, see also Fig. 1. Within days, we varied supplementary interventions: face masks, a buzzer that indicated if one came within 1.5 m of another visitor, no supplementary intervention. It is important to note that when the art fair took place in August 2020, wearing face masks was not obligatory in The Netherlands, and only very few people wore face masks at the time. Of course, if visitors wanted to wear a face mask they were allowed to at all times. While we planned to cross all levels, we deviated from this plan, see “Practical and technological problems”.

Materials

Questionnaires

We administered a pre- and post-questionnaire. The pre-questionnaire included items on demographics, potential coronavirus infection, worries about consequences of the virus, physical distancing and face masks. The post-questionnaire focused on physical distancing behavior during the art fair, and the experience of the event. In this paper, we only use visitors’ demographics such as age and gender, see Tanis et al.28 for an elaborate description of both questionnaires.

Social distancing sensor

In response to the COVID-19 pandemic Focus Technologies B.V. (www.findfocus.nl) together with Sentech B.V. (www.sentech.nl) have developed the Social Distancing Sensor. This sensor is able to determine the distance between itself and other Social Distancing Sensors using ultra-wideband technology with an accuracy of up to ten centimeters. Every sensor has a unique ID and registers how often a violation has occurred with every other sensor. An access point and a computer application are used to manage the system and set settings. Every time a sensor is within reach of the access point, it communicates all distance violations to the access point after which it is written to a central database. A distance violation between two sensors is logged on both sensors and is therefore written to the database twice. The sensor is able to provide direct feedback to the user using a flashing light, a buzzing sensation or a beeping sound. In this research only the light and buzzer are used. To accommodate visitors from the same household, multiple sensors can be paired after which violations of the 1.5 m distance between these specific sensors are not logged to the database and no feedback is given.

Procedure

Via social media, participants were recruited for the event Smart Distance Lab: The Art Fair. Visitors were asked to purchase their ticket in advance. Upon purchasing a ticket, visitors were asked to participate in our study by filling out an informed consent form and a questionnaire on demographics and attitudes towards COVID-19. At the end of the questionnaire a code was provided that allowed participants to select a 30-min time slot to enter the event. In case people showed up at the art fair without having bought their ticket in advance, they could purchase a ticket directly without completing the questionnaire, and they were only asked to provide informed consent. If someone declined to participate in our study, they immediately received this code, and participation in our study was thus not required to visit the art fair. There was a separate code that allowed people to receive a free ticket, but the rest of the procedure was the same.

Visitors were allowed to enter the art fair during their selected time slot. Upon entering, it was explained that research was being carried out in the hall and a health check was conducted. At the Social Distance Sensor (SDS) handout station, visitors were asked whether they had signed the informed consent and completed the corresponding questionnaire. Visitors for whom this was not the case were asked whether they wanted to participate in the study and, if so, to sign the informed consent form. When visitors belonged to the same household, this was registered and the sensors were activated at the same time so that they would not report any incidents with each other. In the face mask condition, after handing out the sensors, face masks were handed out with the request to wear them until they were told that the masks could be removed. Visitors could then enter the art fair and stay inside for as long as they wanted, possibly taking part in multiple experimental conditions. On the way out, visitors handed in their sensors. In addition, they were asked to fill out a post questionnaire about their experience at the art fair.

The study was approved by the ethics review board of the University of Amsterdam (2020-CP-12488), and all participants provided informed consent before participating. All methods were performed in accordance with the relevant guidelines and regulations.

Practical and technological problems

First, the face mask intervention was only executed on day 1, which allows us to only compare the face mask condition to the no supplementary intervention condition on day 1. Second, on day 3, two conditions were cut short (60 min instead of 120 min), see Table 1, and were excluded from analyses. Third, the buzzer notified visitors when they were within 1.5 m from another visitor, but the settings differed across days. Specifically, on the first day the buzzer only notified people after 2 s, which led to a lot of confusion. In addition, we did not actively demonstrate the buzzing feature to the visitors, which resulted in people trying out the feature inside the art fair. On day two and three, visitors were able to test out the buzzer before entering the art fair. In addition, on day two and three the buzzer notified the visitor immediately when they were within 1.5 m of another visitor, without the 2 s delay.

Statistical analyses

Pre-processing and quality control of sensor data

The SDS output contained the reporting and opposing ID for each contact, the number of times they were within 1.5 m, and the time this contact was registered by the access point. Note that a single contact between two visitors is in this database twice, each visitor once as the ‘reporting ID’ and once as the ‘opposing ID’. First, we separated the output into the different experimental conditions and kept records that were between the start and end of each experimental condition (see Table 1 for start and finish times). Second, for each reporting ID in a condition, we checked whether the sensor was handed out during that condition as registered at the handout station. In case we could not link a sensor to a visitor, we removed this contact from the output. These records could occur when an SDS was activated but not handed out, e.g., when the SDS automatically activated when removed from the charger. Third, as the SDS output only contains records of contacts and the reporting and opposing IDs, we had to add visitors to the output who made zero contacts. More details on the processing of the data can be found in the Data Descriptor paper by Tanis et al.28.

Models

The contact networks of visitors were analyzed with a Bayesian logistic regression model including cross-nested random effects for the number of contacts of participants (or actors). For relation Ykij, between actors i and j in network k, the fitted model is:

$${\text{P}}\left( {Y_{kij} = 1} \right) = {{\exp \left( {\mu_{kij} } \right)} \mathord{\left/ {\vphantom {{\exp \left( {\mu_{kij} } \right)} {\left( {1 + \exp \left( {\mu_{kij} } \right)} \right)}}} \right. \kern-\nulldelimiterspace} {\left( {1 + \exp \left( {\mu_{kij} } \right)} \right)}}.$$

The log-odds \(\mu_{kij}\) are constituted of an average mkij and actor-specific random effects Aki and Bkj which are assumed to a have normal distribution with a zero mean:

$$\mu_{kij} = m_{kij} + A_{ki} + B_{kj} ,\quad i \ne j.$$

The average mkij, consists of an overall mean m, and actor-level predictors Xakij with coefficient \(\beta_{a}\):

$$m_{kij} = m + Xa_{kij} \beta_{a} .$$

This is a reduced version of the multilevel p2 model32 for non-directed networks, omitting reciprocity parameters, dyadic predictors, and random effects at the network level, and with identical random sender and receiver effects. In the current application, actor-level dummy variables were applied to model differences in the number of contacts between the two networks observed in different conditions. To distinguish this model from others, it was called b2 (b for bidirectional). Estimation was performed with Markov Chain Monte Carlo simulations, similar to the estimation of the j2 model33,34, with similar prior distributions. Code for this model is available in the R-package dyads35. We report posterior means as point estimates for the odds ratios in the main text together with their associated 95% credible interval. The credible interval gives the range within which the true odds ratio falls with 95% certainty.