Introduction

The introduction of the potential outcome framework has provided a formal set of conditions by which epidemiologists can infer causal relationships from observational data. Using this framework, a number of methods have arisen to estimate causal effects of exposures or treatments in observational data. These methods are often used in longitudinal settings in which exposures, covariates, and outcomes may vary over time. One approach that is fundamental to causal inference is the g-computation algorithm formula (g-formula), which was introduced by Robins, demonstrating that causal inference in complex longitudinal data was possible [1] and which was later formally generalized to estimate effects of arbitrary (e.g., dynamic) treatment regimens [2], competing risks [3], and formal generalization of non-experimental study findings [4]. One of the innovative heuristics that has emerged from this literature has been the re-casting of causal effect estimation in terms of interventions [5]. For epidemiologists, the g-formula has proven to be an essential, if perhaps underutilized tool for estimating human health impacts of exposures and interventions on exposures.

One frequent use for the g-formula is estimating impacts of exposure on the timing of some health event, often referred to as survival analysis. Aside from analytic methods, survival analysis requires three basic tools: a clear definition of failure, a time origin, and a scale for measuring time (time scale) [6]. Such characteristics are often clearly defined in randomized experiments in terms of “natural” time scales like time-on-study, but they may not be in analysis of observational studies with the g-formula where time scale may vary according to substantive interests or for reasons related to policies or interventions of interest; further, multiple time scales may be used as a way to emulate the results of alternative experimental designs [7]. The numerical results and interpretation of a given analysis depend crucially on time scale [8].

In the current manuscript, we explore the choice of time scale in causal effect estimation. Specifically, we discuss issues of time scale for causal effect estimation using the g-formula, illustrating with examples from the literature when possible to demonstrate tradeoffs between etiologic or policy considerations and the potential for bias. We discuss issues of interpretation that arise when standard complications of survival analysis arise, namely left truncation (late entry) and right censoring (loss to follow-up or end of follow-up). We point out when such issues have arisen in our own and others’ work in hopes of spurring stronger consideration of the choice of time scale in causal effect estimation as well as the refinement of methodologic approaches to late entry.

Throughout, we provide examples from a survey of the literature with applications of the g-formula, we performed a MEDLINE search on 9 Oct. 2017 using the search terms given in the Appendix. This search yielded 94 results, which we narrowed to 56 articles after excluding manuscripts that did not include analysis of real data with the g-formula. One of us (AK) classified the articles meeting inclusion criteria according to the time scale of analysis. We focus mainly on the 43 analyses of cohort studies ([9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51], shown in Fig. 1), but we also include selected examples of studies published after the MEDLINE search was conducted. We identified four major types of time scales, given as “study” (time on study), “age,” “landmark” (time since some eligibility defining event such as pregnancy), and “treatment” (time since treatment/exposure).

Fig. 1
figure 1

Cumulative publications of the g-formula identified in a MEDLINE search by time scale of interest. Included studies represent analyses of cohort data (not shown for clarity: reference [9], published in 2001)

Potential Outcomes and the g-Formula

The g-formula is a powerful tool for causal effect estimation developed by Robins as a way to translate Rubin’s potential outcomes framework [52] to the longitudinal setting. Rubin defined potential outcomes with respect to point exposures, or exposures that occur at a specific point in time or represent a summary of exposures measured at a specific point in time. In the point exposure setting, a potential outcome, denoted by yx, is the value of some outcome (y) we would have observed had exposure been set to the value x. For every individual who, in fact, experienced exposure at the level X = x, their potential outcome is observed (i.e., everyone has only one observed potential outcome for a given Y, but all other potential outcomes are unobserved, or counterfactual. While causal inference does not require the existence of counterfactuals [53], they are a useful (possible) fiction that greatly facilitates statistical analysis of causal questions [54,55,56].

In longitudinal settings with time-varying exposures (and confounders), exposure or treatment can be expressed in terms of regimens. An exposure regimen can be expressed in terms of a vector \( {\overline{x}}_k=\left({x}_1,\dots, {x}_k\right) \), where the subscript refers to a time index. For example, on the time scale “age,” the number of television hours watched at age 10 could be denoted by x10, while the lifetime history of watching television by age 10 could be expressed by \( {\overline{x}}_{10} \). We denote potential outcomes as \( {\overline{y}}_k^{{\overline{x}}_k} \), which refers to the set of outcomes we would have observed up to time k, had exposure followed the regimen \( {\overline{x}}_k=\left({x}_1,\dots, {x}_k\right) \). Survival potential outcomes, such as time to death, are denoted by \( {t}^{{\overline{x}}_k} \). For purposes of data analysis with the g-formula, survival outcomes are often discretized such that \( {\overline{y}}_k \) is a vector of binary indicator variables which are 0 except when k is equal to a rounded version of t. This approximation is improved by making the time interval from k to k + 1 be small relative to t [44]. We denote survival by time k as \( {\overline{y}}_k=0 \).

Assuming no competing risks, the survival function of a potential time to event outcome can be expressed by

$$ {S}^{{\overline{x}}_k}(k)=\Pr \left({\overline{Y_k}}^{{\overline{x}}_k}=0\right)=\prod \limits_{m=1}^k\Pr \left({Y}_m^{{\overline{x}}_m}=0|{Y}_{m-1}^{{\overline{x}}_{m-1}}=0\right) $$
(1)

where Π is the product integral and \( \Pr \left({Y}_0^{{\overline{x}}_0}=0\right) \) is defined to be 1. The cumulative distribution function, given as \( {F}^{{\overline{x}}_k}(k)=1-{S}^{{\overline{x}}_k}(k) \), is often used for data analysis. In a closed cohort, the cumulative distribution function corresponds to the epidemiologic measure of risk. Implicitly, there is a fixed target population of interest such that the \( {\overline{Y_k}}^{{\overline{x}}_k} \) refer to potential outcomes in the target population.

The g-formula is often used to estimate risk (sometimes called cumulative risk), demonstrated by the numerous examples, which is useful for inference in public health settings [57•]. The basic unit of risk is given on the right side of (1), \( \Pr \left({Y}_m^{{\overline{x}}_m}=0|{Y}_{m-1}^{{\overline{x}}_{m-1}}=0\right) \), which is the discrete time hazard at time m under the regimen \( {\overline{x}}_k \). Given data from a closed cohort, the discrete time hazard for potential outcomes under the regime \( {\overline{x}}_m \) can be estimated under the causal identification and modeling assumptions of the (parametric) g-formula using probability rules and two regression models:

$$ \Pr \left({Y}_m=0\ |\ {\overline{X}}_m={\overline{x}}_m,{\overline{W}}_m={\overline{w}}_m,{Y}_{m-1}=0;\beta \right) $$
(2)
$$ \Pr \left({W}_m={w}_m\ |\ {\overline{X}}_{m-1}={\overline{x}}_{m-1},{\overline{W}}_{m-1}={\overline{w}}_{m-1},{Y}_{m-1}=0;\alpha \right) $$
(3)

where Wm is the set of confounders (assumed to be discrete for convenience) for the effect of Xm on Ym and β and α are vectors of parameters. Unlike standard approaches to baseline confounding, a model for Wm is necessary in the g-formula. Modeling Wm is typically required to estimate \( \Pr \left({Y}_m^{{\overline{x}}_m}=0|{Y}_{m-1}^{{\overline{x}}_{m-1}}=0\right) \) because interventions to set \( {\overline{X}}_m \) to \( {\overline{x}}_m \) may impact the distribution of subsequent confounders. Some of the effect of Xm on subsequent outcomes may be through future values of W, and modeling this relationship allows that time-varying confounders may be affected by interventions on exposure. Some analyses target the discrete time hazard ratio, in which case analysis may stop at this point [44, 58].

Using the parametric g-formula to estimate risk involves combining the probability rules of (1) and the parametric models in (2) and (3) to estimate risk under the scenario in which we could set exposure to the regime \( {\overline{x}}_m \). Notably, (2) and (3) can be estimated using standard regression software. The specific model forms used depend on context, but often include familiar model forms such as pooled logistic regression (e.g., [46]), pooled linear regression (e.g., [59, 30]), multi-state models (e.g., [14, 60]), multinomial regression (e.g., [61••]), Cox proportional hazards regression (e.g., [1, 50]), and Bayesian regression (e.g., [62••, 63]). In the case of a point exposure, the g-formula can be simplified to a single model [64]. The point exposure case has become a relatively common example of the g-formula (e.g., [10, 18, 28, 30,31,32, 34•, 39, 43••, 65,66,67,68,69,70,71,72,73]); likely because it is a straightforward extension to common modeling frameworks. We focus on the more general case of using the g-formula for survival analysis with time-varying exposure and confounders. However, we cite other examples in which non-survival outcomes (e.g., body mass [26]) are analyzed with the g-formula, but for which issues of time scale are still relevant.

Time and Time Scales in the g-Formula

Time is explicitly considered in the g-formula as the index for time-specific exposures, confounders, and outcomes. The consequence of this is that the value of \( {\overline{Y}}_k^{{\overline{x}}_k} \) will depend on what the time scale to which k refers. The index k could represent age, calendar period, time on treatment, time on study, time since a landmark event, or any other number of time scales. The origin of any time scale occurs at k = 0.

Loss to Follow-up, Censoring, Left Truncation, and Late Entry

In a cohort study, an individual is considered at risk if she or he could (a) experience the outcome of interest and (b) have that outcome recorded in the data. An individual is no longer at risk if he or she experiences the outcome of interest or a competing risk, or if he or she is lost to follow-up. Data are subject to “right censoring” if outcomes might occur among observed individuals but not be recorded. Individuals who are lost to follow-up are often considered censored in that that they are assumed to experience the event of interest after the last time they are observed.

Similarly, individuals can experience outcomes before follow-up starts. Data where some individuals enter the at-risk set after the origin are subject to “left truncation.” For non-repeatable survival outcomes, right censoring and left truncation result in outcomes that we do not observe because the individuals experiencing the events are not under observation. Truncation and censoring are distinct, however. Taking mortality outcomes as an example, if an individual has a right-censored death, then he or she may be observed in the study. However, left truncation results in some deaths that occur before an individual could come under observation; then, he or she is prevented from ever being under observation for the study. We take it as a given that the individuals experiencing those events are of interest for estimating causal effects. Thus, when estimating risk, right censoring and left truncation result in under-counting the number of events, while left truncation also results in under-counting the size of the population. Jointly, we refer to left truncation and right censoring as “non-observation.” While other types of censoring and truncation may occur, we restrict the remainder of the manuscript to these two processes in order to simplify the exposition.

Non-informative Non-observation and Causal Parameters

Causal inference with the g-formula is possible under a set of sufficient identifying assumptions, namely exchangeability. A simplified version of the exchangeability assumption given by [2] is that

$$ {\overline{Y_k}}^{{\overline{x}}_k}\coprod \left({X}_k,{C}_k\right)\mid {\overline{\mathrm{W}}}_k={\overline{w}}_k,{\overline{X}}_{k-1}={\overline{x}}_{k-1},{C}_{k-1}={Y}_{k-1}=0 $$
(4)

where ∐ implies independence between the left and the right sides and Ck is censoring at time k. Notably, no provision for late entry is given in this oft-used definition of exchangeability. However, in his original paper on the g-formula [1], Robins anticipated the problem of late entry, noting that “the only safe option is to match on exposure and [confounder] history until time of entry” (p. 1436). A rigorous derivation of the stronger assumption of exchangeability necessary under late entry is beyond the scope of this paper. Roughly, however, for causal inference to be possible in survival analysis subject to left truncation, exchangeability must hold conditional on the measured past values of exposure and confounders, among those who have entered into the study by time k and remain uncensored at time k. This strengthened exchangeability assumption implies that underlying health status should not depend on entry time, conditional on past values of exposure and confounders. While we do not discuss them here, estimating effects of interventions using the g-formula also requires the causal identification conditions of positivity and consistency and the statistical assumption of correct model specification [1].

From Hazards to Risks with Right Censoring

Provided that exchangeability (4) holds, the conditional, discrete time hazard given in (2) can be consistently estimated given a correct statistical model. However, the survival or cumulative distribution functions may be of greater interest than hazard functions, and we may be more interested in marginal, or population average parameters. Taubman’s influential paper on estimating effects of interventions on risk factors for coronary heart disease (e.g., smoking, body mass, and exercise) is a widely cited example that demonstrated how to estimate population average risk using the g-formula [3]. Notably, Taubman et al. used data from the Nurses’ Health Study, a cohort study which began follow-up in 1976. The authors used a subset of the person-time from 1982 onwards, but included covariate data from 1980. This method of cohort selection resulted in a cohort that was closed on the left, thus avoiding issues of late entry but allowing for right censoring due to loss to follow-up. Taubman et al. used a version of formula (1), modified to account for competing risks, to estimate the 30-year risk of coronary heart disease in the population under a set of hypothetical interventions that started in 1982.

Taubman’s approach to estimating (population average) risk leveraged the absence of late entry. This approach relies on the mathematical result that the population average risk is equal to the average of the individual risks, given by \( {F}^{{\overline{x}}_k}(k)={n}^{-1}{\sum}_i^n{F}_i^{{\overline{x}}_k}(k) \), where i refers to an individual. In the Monte Carlo algorithm often used to estimate parameters of the g-formula, this corresponds to predicting the individual risk at each time k under the intervention “Set \( {\overline{X}}_k={\overline{x}}_k \)” for all cohort members and then averaging those risks across the population. Naively, this could be problematic in the case of right censoring because censoring results in some participants not being present in the data for some time points (which would result in selection bias from informative censoring due to averaging over potentially increasingly healthy subsets of the population as k grows). Taubman et al. solved this problem by assuming exchangeability and simulating each participants’ outcomes beyond the times when they were, in fact, censored (though Taubman et al. also performed a sensitivity analysis in which censoring was not prevented by modeling censoring as a function of prior exposure and covariate histories). Thus, Taubman et al.’s primary analyses encode joint interventions of the type “Set \( {\overline{X}}_k={\overline{x}}_k \) and prevent censoring from loss to follow-up.” Administrative censoring from the end of follow-up is addressed by simply evaluating risk up to the end of follow-up, presuming it occurs at the same time for all participants on the time scale of interest.

Taubman’s approach for loss to follow-up can be easily extended for other forms of censoring, at a possible cost of interpretability. For example, Jain et al. estimate the effect of smoking cessation on 20-year weight gain, which can only be observed among those who survive 20 years [26]. The authors use an implicit intervention to prevent mortality over the risk period. While controversial [74], this approach to competing risks has been shown by Tchetgen Tchetgen et al. to correspond to a combination of principle strata effects and yield a valid test of the sharp null hypothesis between exposure and the outcome of interest [75].

From Hazards to Risks with Left Truncation

For right censoring, the g-formula can account for missing outcomes by simply simulating participant data as though they had not been censored. The situation is not as straightforward for left truncation. Recall that, with right censoring, we are missing outcome data on some individuals who are observed. However, with left truncation, we are missing outcome data on individuals from a target population who were never observed because they experienced some event that prevented them from entering the study to begin with (but are nonetheless of interest). For example, in a target population of all workers who were employed at a Montana copper smelter, explored in work by Robins [1] and later by Keil et al. [76], data were available only for workers who were employed on or after 1 January 1938.

Assuming exchangeability, the missing individuals create no problem for estimating the conditional hazard at time k, which is conditional on covariates and on survival to time k and does not depend on the missing data. However, for cumulative average risk, left truncation is more problematic than right censoring because it implies we do not know the size of the target population to begin with. Note that this problem results when the target population is subject to late entry, even when exchangeability holds in the study sample.

Under the definition of exchangeability extended to account for left truncation, Taubman et al.’s algorithm cannot be used to estimate risk in target population data subject to left truncation. Frequently, the study population (reimagined as a closed cohort) is the target population of interest, so the issue may arise in any data subject to left truncation. In order to account for these missing individuals, we would have to simulate extra members of the population, but it is not clear how these pseudo-individuals would be assigned baseline or time-fixed covariates, given that they were never observed. Unfortunately, there is no apparent solution given in the methodologic literature to this exact problem.

To more explicitly define the unique issue caused by late entry, it is helpful to more concretely describe issues with the more tractable problem of right censoring. Consider a possible cause of loss to follow-up in the Nurses’ Health Study such as stroke, which could prevent participants from filling out study surveys. Assume that stroke and coronary heart disease (Taubman’s outcome of interest) are related through common etiologic pathways, such that they are associated in the population. Under Taubman et al.’s approach, if a woman were lost to follow-up due to stroke-related non-response, her subsequent data would be simulated assuming exchangeability. Thus, in Taubman et al.’s approach, the average target population hazard in older ages would be based on a group of women that includes all of those who survived with stroke. Under an alternative approach, her outcome would be censored after the last completed survey, and risk would be calculated using the hazards estimated among those who were not loss to follow-up due to stroke. If stroke is, in fact, related to underlying CHD risk, then the average hazard among older women will have been underestimated due to selection bias caused by censoring those women, even if the conditional hazard ratio is unbiased; Taubman’s approach avoids censoring-related selection bias by imputing the covariate values after loss to follow-up, thus eliminating censoring. The alternative approach assumes (with respect to censoring) conditional exchangeability in the study sample, but it assumes unconditional exchangeability in the target population. This phenomenon occurs because the target population average risk is calculated from the population average hazards, rather than conditional hazard, and the averaging occurs with respect to the covariate distributions in the observed (and imputed) data. The observed plus imputed data used in Taubman et al.’s approach represent the hypothetical study sample we would have observed, had loss to follow-up been prevented. Notably, this only means that covariates among observed individuals need to be imputed.

The key barrier to progress with reducing selection bias due to late entry in the g-formula is that left truncation means that some members of the target population may never be observed, implying that entire covariate histories of new individuals would need to be imputed. Unfortunately, the theory underlying such an approach is underdeveloped, but addressing potential bias from informative late entry is an area of active research. The analogy with right censoring (and solutions in the g-formula) illuminates this issue and highlights potential solutions. The issues of late entry and exposure occurring prior to baseline are not new, however, and care must be taken when considering both when the time scale starts and when the intervention starts.

Additional Considerations for Interventions in Light of Late Entry

A second problem occurs with the “intervention” aspect of the g-formula with late entry (as well as with causal effect estimation, in general, when exposure occurs before study entry [77]). For the intervention “Set \( {\overline{X}}_k={\overline{x}}_k \),” we implicitly assume that exposure is modifiable at every time past the origin. Consider, however, an individual who enters an occupational cohort study at time k = 5 and, through work records, we discover that individual has been at work, but unexposed from time k = 0 through study entry. Also consider the intervention “If at work, be exposed to high levels of silica.” Suppose that, if exposed to such levels, this individual would have developed silicosis at time k = 3 and left employment. Thus, under the supposed intervention, this individual would not have been in the study sample. Consequently, exposure may influence whether or not an individual enters the study and estimating effects of “intervening” on exposure prior to study entry is subject to selection bias. Such selection bias occurs unless exposure prior to the start of follow-up is known not to affect study outcomes conditional on measured covariates [77], an assumption that is empirically testable if some cohort members do not enter late and exchangeability holds with respect to exposure.

One approach to this issue is to consider exposure accrued prior to study entry to be a baseline confounder, and, as such, cannot be intervened upon. Such issues are common in occupational studies. For example, Neophytou et al. used the g-formula to estimate the effect of interventions on respirable elemental carbon on lung cancer mortality, but only implemented the interventions among the exposure that occurred during observed person-time, while adjusting for exposure that occurred prior to study entry [12]. Such an approach is akin to the “treatment decision design” in which interventions are considered for relevant treatment decisions among prevalent and incident users of drugs [78•]. In other analytic approaches, bias reduction is achieved by creating study samples in which individuals are known to be unexposed/untreated prior to study entry; this set of approaches has been given the name “new user designs” in observational studies of pharmaceuticals [79]. For some etiologic questions, new user designs reduce bias by eliminating “prevalent” exposure that cannot be intervened upon. However, excluding individuals who enter with prior exposure may not be desirable when dealing with small study samples or when individuals exposed before baseline may represent a scientifically interesting group [80].

Potential for Bias and Time Scale Relevance in the g-Formula

Many examples of cohort data analysis with the g-formula involve questions in which a specific time scale is natural to the study design, reflecting interest in the study population as the implied target population. For example, several authors estimate effects of hypothetical interventions to reduce the population burden of cardiovascular and related diseases (e.g., [20, 29, 35, 36, 42•]). In each of these cases, the proposed interventions are expressed in terms of actions that occur at the beginning of follow-up in a closed (on the left) cohort. For relevance to interventions in the study sample, a natural time scale for analysis is “time-on-study,” which was employed by 29/43 of the cohort studies we identified. As shown in Taubman et al.’s example, such analyses can readily address informative loss to follow-up by measured covariates.

In other cases, landmark events such as HIV seroconversion [50], pregnancy [30, 31], or hospitalization [32] define the start of the time scale. In the case of seroconversion and pregnancy, the precise start of the time scale may be unknown, and current approaches to this problem with the g-formula involve imputing the start of the time scale (e.g., [50]). However, in the case of early loss and pregnancy, in particular, such studies have the potential for late entry and possible misidentification of the target population if the time scale might start and stop before measurement of the landmark event [81].

Often, subject matter considerations make it desirable to estimate quantities on time scales that are almost certain to be subjected to late entry in many study designs. Occupational studies are one such example. In the USA, the Occupational Safety and Health Administration (OSHA) sets personal exposure limits for many workplace hazards, such as asbestos [46, 82], radon [23], diesel exhaust [12], arsenic [1, 11], and silica [83]. Exposure limits are based on reducing worker risk below a risk difference of 1/1000 comparing no exposure with exposure at the occupational limit for 45 years (e.g., [84]) where risk is evaluated from ages 20 to 85. Thus, for relevance to policy interventions in the USA, the natural time scale for analysis is “age,” even if it does not correspond to many occupational cohort studies that recruit workers based on calendar time; many workers are older than 20 at baseline and some do not even start work by age 20 and so would not be captured. Age was the second most used time scale in the identified analyses (9/43 studies). Of the six studies we identified via MEDLINE with possible late entry [9, 11, 12, 23, 46, 50], five of them utilized age as the time scale of analysis.

Notably, in the occupational studies we identified that employ the g-formula, all [1, 11, 12, 23, 46, 82, 83] but two [14, 22] used age as the time scale of analysis. Consequently, these studies were subject to left truncation because not all workers started employment (and follow-up) at the origin of the time scale, often age 20. Further, workers who were hired at younger ages in each of these studies will be subjected to administrative censoring due to the end of follow-up without reaching the maximum age on the time scale, meaning that not all workers have the same administrative end to the time scale. Typically, such variation in administrative censoring is addressed by considering such censoring to be equivalent to loss to follow-up and requiring that exchangeability must also hold for administrative censoring. Each of these analyses accounted for loss to follow-up using variations of Taubman’s approach. All but one of the identified occupational studies with late entry used the implicit assumption that marginal exchangeability held for late entry. The notable exception was the original paper that described the g-formula, which included a data analytic example of estimating lung cancer mortality risk among a cohort of copper smelters exposed to arsenic [1]. In this paper, Robins estimated risk in “the subset of the observed study population hired at age 32 in 1935 remaining on work at high exposure until start of follow-up in 1938” (p. 1503). Thus, in Robins’ approach, late entry and variation in administrative censoring times (and the need for additional exchangeability assumptions) were eliminated by considering the target population to be a subset of the study population in which late entry did not occur and who would all reach the end of follow-up on the time scale of interest. This approach trades external validity for internal validity. That is, by conditioning (restricting) on age and date of hire, we lose the ability to easily generalize results to the general smelter worker population. However, through this conditioning, we need only assume that workers in the target population are marginally exchangeable with other workers who could have been in the study (but died or quit work) who were of a similar age, hired at a similar time, and had similar pre-study exposures. Another way to state this idea is that we expect marginal exchangeability to hold within this group of workers who are similar in age, date of hire, and exposure at baseline. An alternative approach might consider completely synthetic worker populations, such as those used by the OSHA to estimate lifetime risk for regulatory purposes.

Conclusions

The choice of time scale in analyses with the g-formula requires careful thought that weighs substantive concerns with the plausibility of the assumptions needed to identify effects on each time scale. Loss to follow-up may not result in bias, provided that exchangeability assumptions are met. However, problems with late entry and variation in administrative censoring times require additional consideration. For the g-formula, in contrast with other survival analysis methods [85, 86], left truncation and right censoring are not automatically addressed by the same method when risk is the estimand of interest and the study population is the target population. We hope this manuscript encourages further consideration of the assumptions necessary to make inference with the g-formula when using a time scale subject to late entry. Further, we hope to spur methodologic innovations to address informative late entry while still allowing inference on biologically or policy relevant time scales.