Keywords

Introduction

Probability models for churn prediction in contractual settings, e.g., Beta-Geometric (BG) [8] and Beta-Weibull [13], have been used in customer retention across business sectors, e.g., online platforms [19], video/music streaming [17], and insurance [27]. The use of such models goes beyond churn prediction and extends to predicting time to purchase [14], patient revisit [20], conversion rate of online advertising [17], and customer-based valuation of business [5, 21]. The models are robust to incomplete information [7], able to leverage interval-censored and non-censored data [10], and provides managerially relevant statistics, e.g., residual lifetime, lifetime value, and their distributions [5, 6, 12]. Despite the proven efficacy, some criticize those models for being covariate-free and hence unable to fully capture group-level and individual-level heterogeneity [22, 23]. Hubbard et al. [17] introduce the Beta-Logistic (BL) model—an extended BG with time-invariant covariates—and empirically improve churn prediction performance in streaming subscribers. While relaxing the restrictive assumption that all individuals share an identical distribution of latent churn rate, BL assumes the heterogeneous distributions to be time-invariant (stationary). That is, BL does not allow for time-varying covariates such as seasonality and tactical marketing activities that are common in many business sectors and result in nonstationary distributions of latent attrition/churn rate [1, 21, 26].

To fill the gap, Fader and Hardie [9] propose a Grassia(II)-Geometric (G2G) model with time-varying covariates, which captures heterogeneous and non-stationary latent churn process. Notwithstanding the theoretical flexibility, G2G’s efficacy has not been empirically examined, and it remains unclear whether G2G effectively improves the less generic BG (assuming a common churn rate distribution) and BL (assuming heterogeneous yet stationary churn rate distributions). We collaborate with a leading electronic manufacturing services (EMS) company and assess the G2G for predicting employee churn in its production plants. Due to high turnover rates, individual heterogeneities, and seasonal variations in employee churn risks (e.g., there are more competing offers in peak seasons), the focal EMS requires a flexible yet interpretable model for predicting churn proportion and remaining lifetime for labor planning and hiring in advance. Motivated by the conceptual similarity between customer and employee churn [24, 25], our analysis shows that stochastic churn modeling can be useful in not only consumer services but also workforce operations.

The contribution of this paper is multi-facet. First, we apply Bayesian estimation to BL and G2G, going beyond prior studies’ focal interest in churn prediction and instill posterior inference in such models. Second, we empirically show that the more generic and flexible G2G significantly outperforms both BG and BL models. Third, we find that covariate-free BG precisely projects the global survival curves and job durations, whereas G2G effectively captures individual heterogeneity and satisfactorily projects local curves that BG cannot capture. Last, unlike some machine learning approaches that are short of rationales and difficult to obtain test statistics, the proposed Bayesian G2G is highly interpretable with readily available credible regions of model estimates, such that managers can gauge the significance of predictors and their impact on the metrics of managerial relevance.

General Problem and Models

Contractual Churn and Probability Modeling

Per [11], the problem of churn modeling in general can be categorized into four quadrants by two dimensions: transaction opportunity (continuous/discrete) and relationship type (noncontractual/contractual). This paper focuses on contractual and discrete-time setting that matches the EMS operations, in which employees have an opportunity to decide on whether quitting from production plants at the end of each period. An employee is considered “survival” if he/she stays on-job in the beginning of the next period. Otherwise, he/she is considered “churn”.

Throughout the paper, we use index \(i\) to represent individuals, and \(T_i \in Z^+\), \(C_i \in \left\{ {0,{ }1} \right\}\), and \(X_i \in R^d\) refer to one’s duration of survival subject to right-censoring, censor indicator (“1” censored survival, “0” observed churn instance), and d-dimensional covariates. For notational brevity, we use \(D = \left\{ {T_i ,C_i ,X_i } \right\}_{i = 1}^N\) to denote data with \(N\) individuals and \(\left\{ \cdot \right\}\) to represent all model parameters. Mathematically, the objective of stochastic churn modeling is to optimize \(\left\{ \cdot \right\}\), including the coefficients on covariate effects and distributional parameters, such that the overall likelihood function is maximized,

$$ P\left( {D| \cdot } \right) = \left\{ { \prod \limits_{i:C_i = 0} P\left( {T_i | \cdot } \right)} \right\} \cdot \left\{ { \prod \limits_{i:C_i = 1} S\left( {T_i | \cdot } \right)} \right\} $$
(6.1)

where the first part is the probability mass function \(P\left( {T| \cdot } \right)\) for employee churned (\(C_i = 0\)) and the second part is the survival function \(S\left( {T| \cdot } \right)\) for employee survival (\(C_i = 1\)). Given the estimated model parameters, one can easily apply the survival function \(S\left( {T_i | \cdot } \right)\) to infer an individual’s discounted expected residual lifetime (DERL) [5]:

$$ DERL\left( {d| \cdot , n} \right) = \mathop \sum \limits_{t = n}^\infty \frac{S(t| \cdot )}{{S(n - 1| \cdot )}} \cdot \left( {\frac{1}{1 + d}} \right)^{t - n} $$
(6.2)

where \(n\) stands for the number of periods an individual has survived and \(d\) denotes a discount rate. The DERL can be adapted to predict an unseen individual’s expected lifetime (EL) by setting \(n = 1.\) That is,

$$ EL\left( {d| \cdot } \right) = \mathop \sum \limits_{t = 1}^\infty S(t| \cdot ) \cdot \left( {\frac{1}{1 + d}} \right)^{t - 1} $$
(6.3)

Probability Churn Models from Marketing

Beta-Geometric (BG)

Beta-Geometric (BG) is built on a simple idea assuming that at each period, an individual determines whether to stay active by flipping a two-sided coin with a latent attrition parameter \(\theta\) (i.e., “head” one ends the contract, “tail” one renews it). The number of trials before showing up of the first head follows a geometric distribution, whereas the probability of attrition \(\theta\) is heterogeneous across individuals and static over time. Assuming heterogeneous \(\theta\) follows a beta distribution, the geometric churn process results in the BG model, i.e.,

$$ P\left( {T_i |\theta } \right) = \theta \left( {1 - \theta } \right)^{T_i - 1} ,\;S\left( {T_i |\theta } \right) = \left( {1 - \theta } \right)^{T_i } ,\;f\left( {\theta |\alpha ,\beta } \right) = \frac{{\theta^{\alpha - 1} \left( {1 - \theta } \right)^{\beta - 1} }}{{{\text{B}}\left( {\alpha ,\beta } \right)}} $$
(6.4)

Integrating out the latent (θ), the probability mass and survival functions can be re-written as:

$$ P\left( {T_i |\alpha ,\beta } \right) = \frac{{{\text{B}}\left( {\alpha + 1,\beta + T_i - 1} \right)}}{{{\text{B}}\left( {\alpha ,\beta } \right)}},\;S\left( {T_i |\alpha ,\beta } \right) = \frac{{{\text{B}}\left( {\alpha ,\beta + T_i } \right)}}{{{\text{B}}\left( {\alpha ,\beta } \right)}} $$
(6.5)

where \(\alpha\) and \(\beta\) are model parameters to be estimated such that the overall likelihood \(P\left( {D| \cdot } \right)\) is maximized.

Beta-Logistic (BL)

Unlike BG that assumes a stationary and common beta distribution of latent churn rate \(\theta\), BL extends BG by leveraging covariate information and results in stationary yet heterogeneous beta distributions. More specifically, BL makes \(\mu_i\) (mean) and \(\sigma_i^2\) (variance) of the latent attrition rate \(\theta\) functions of individual-specific and time-invariant covariates (\({\varvec{X}}_{\varvec{i}} = \left[ {x_{i,1} , \ldots ,x_{i,d} } \right]\)), i.e.,

$$ \begin{aligned} \mu _i & = \log {\text{it}}\left( {\gamma _{\mu ,0} + \gamma _{\mu ,1} x_{i,1} + \ldots + \gamma _{\mu ,d} x_{i,d} } \right), \\ \sigma _i^2 & = \exp \left( {\gamma _{\sigma ,0} + \gamma _{\sigma ,1} x_{i,1} + \ldots + \gamma _{\sigma ,d} x_{i,d} } \right) \\ \end{aligned} $$
(6.6)

Our parameterization \(\left( {\mu ,\sigma^2 } \right)\) differs from \(\left( {{\upalpha },{ }\beta } \right)\) in [17] for ease of interpretation. Technically, the formulas below allow one to transform \(\left( {\mu_i ,\sigma_i^2 } \right)\) back into \(\left( {\alpha_i , \beta_i } \right)\), and compute the likelihood using \(P\left( {T_i {|}\alpha_i , \beta_i } \right)\) and \(S\left( {T_i {|}\alpha_i , \beta_i } \right)\).

$$ \alpha_i = \left( {\frac{1 - \mu_i }{{\sigma_i^2 }} - \frac{1}{\mu_i }} \right)\mu_i^2 ,\;\beta_i = \alpha_i \left( {\frac{1}{\mu_i } - 1} \right) $$
(6.7)

Since the covariates in combination with the coefficients (\(\left[ {\gamma_{\mu ,0} , \ldots ,\gamma_{\mu ,d} } \right]\) and \(\left[ {\gamma_{\sigma ,0} , \ldots ,\gamma_{\sigma ,d} } \right]\)) fully determine the parameters \(\mu_i\) and \(\sigma_i^2\) of latent churn rate distributions, the overall likelihood of BL is conditioned on the coefficients, which leaves them the only parameters to be optimized.

Grassia(II)-Geometric (G2G)

Previous studies have acknowledged that the parametric formulation of BG may fit data poorly sometimes owing to ignoring time-varying factors and duration dependence, e.g., long-living individuals tend to live longer [8, 13]. For such contexts, we need to consider a geometric process with dynamic attrition/churn rate that is generic and flexible. That is,

$$ P\left( {T_i |\theta_{i,1} , \ldots ,\theta_{i,T_i } } \right) = \theta_{i,T_i } \mathop \prod \limits_{t = 1}^{T_i - 1} \left( {1 - \theta_{i,t} } \right),\;S\left( {T_i |\theta_{i,1} , \ldots ,\theta_{i,T_i } } \right) = \mathop \prod \limits_{t = 1}^{T_i } \left( {1 - \theta_{i,t} } \right) $$
(6.8)

However, it is impractical to apply separate prior for each \(\theta_{i,t}\) and integrate over all the priors. Alternatively, Fader and Hardie [9] propose a Grassia(II)-Geometric (G2G) model that replaces the Beta prior in BG with a Grassia(II) distribution. Their formulations based on the clog-log link and gamma heterogeneity areFootnote 1:

$$ \theta _{i,t} = 1 - \exp \left( { - \eta \phi _{i,t} } \right),P\left( \eta \right) \sim Gamma\left( {a,b} \right) $$
(6.9)

where \(\phi_{i,t}\) captures observable heterogeneity and subsumes effects of time-invariant and time-varying covariates over time. Let \({\varvec{X}}_{\varvec{i}}^{\varvec{c}} = \left[ {x_{i,1}^c , \ldots ,x_{i,d}^c } \right]\) be d-dimensional time-invariant covariates (with effect \({\varvec{\gamma}}^{\varvec{c}}\)) and \({\varvec{X}}_{\varvec{i}}^{{\varvec{v}},{\varvec{t}}} = \left[ {x_{i,1}^{v,t} , \ldots ,x_{i,d^{\prime}}^{v,t} } \right]\) be d’-dimensional time-varying covariate at timing \(t\) (with effect \({\varvec{\gamma}}^{\varvec{v}}\)), \(\phi_{i,t}\) brings together the stationary component (\({\varvec{\gamma}}^{\varvec{c}} {\varvec{x}}^{\varvec{c}}\)) and non-stationary component (\({\varvec{\gamma}}^{\varvec{v}} {\varvec{x}}^{{\varvec{v}},{\varvec{t}}}\)):

$$ \phi_{i,t} = \exp \left( {{\varvec{\gamma}}^{\varvec{c}} {\varvec{X}}_{\varvec{i}}^{\varvec{c}} + {\varvec{\gamma}}^{\varvec{v}} {\varvec{X}}_{\varvec{i}}^{{\varvec{v}},{\varvec{t}}} } \right) = \exp \left( {\gamma_1^c x_{i,1}^c + \ldots + \gamma_d^c x_{i,d}^c + \gamma_1^v x_{i,1}^{v,t} + \ldots + \gamma_{d^{\prime}}^v x_{i,d^{\prime}}^{v,t} } \right) $$
(6.10)

The above formulation breaks the limitation of stationarity and exclusively time-invariant covariates in BL. Integrating \(P\left( {T_i {|}\theta_{i,1} , \ldots ,\theta_{i,T_i } } \right)\) and \(S\left( {T_i {|}\theta_{i,1} , \ldots ,\theta_{i,T_i } } \right)\) with respect to the latent \(\eta\), we can derive churn rate (\(P\left( {T| \cdot } \right)\)) and survival rate (\(S\left( {T| \cdot } \right)\)), which constitute the likelihood function and are directly conditioned on Gamma parameters \(\left( {a,b} \right)\) and coefficients (\({\varvec{\gamma}}^{\varvec{c}}\) and \({\varvec{\gamma}}^{\varvec{v}}\)),Footnote 2

$$ \begin{aligned} P\left( {T_i |a,b,\gamma } \right) & = \left\{ {\frac{b}{{b + \sum _{t = 1}^{T_i - 1} \phi _{i,t} }}} \right\}^a - \left\{ {\frac{b}{{b + \sum _{t = 1}^{T_i } \phi _{i,t} }}} \right\}^a , \\ S\left( {T_i |a,b,\gamma } \right) & = \left\{ {\frac{b}{{b + \sum _{t = 1}^{T_i } \phi _{i,t} }}} \right\}^a \\ \end{aligned} $$
(6.11)

Empirical Methods and Data

EMS Workforce Data

Aimed at examining the model’s efficacy and eliciting insights potentially valuable to workforce management practice, we co-work with an anonymous EMS company, Alpha, who provides us with a dataset on 20,000 employees over 145 weeks. In the dataset, in addition to \(T_i\) (duration of survival) and \(C_i\) (censor status), each employee is characterized by 28 time-invariant covariates, including gender, age, hometown, onboard month, category of recruit/contract, mean salary bonus, ratios working on rest day and working on daytime, and mean working hour. In addition, 16 time-varying covariates such as cumulative weeks of survival and season/month indicators for each period are included for capturing time dynamics in distributions of \(\theta_{i,t}\). In Table 6.1, we summarize and explain our operationalization of time-invariant (\(x_1\) to \(x_{28}\)) and time-varying covariates (\(x_{29}\) to \(x_{44}\)).

Table 6.1 Summary of covariates

In Fig. 6.1, we show the distribution of employee’s on-job duration (with density rescaled). Employees onboarding in Year 1 (during weeks 1–52) are on the left panel, whereas employees onboarding in Year 2 (during weeks 53–104) are on the right panel. As can be seen, both distributions are right-skewed and have a majority of employees having duration less than 20 weeks. Moreover, both panels exhibit bimodal curves for duration between 0 and 10 weeks and multi-modals for duration longer than 20. The shape implies mixtures of heterogeneous employee cohorts, echoing the proposal of BL/G2G to bring in covariates to explicitly account for cross-sectional heterogeneity which covariate-free BG cannot capture.

Fig. 6.1
Two line graphs with shaded areas represent censored survival in Years 1 and 2. In years 1 and 2, the curves skewed to the right have a minimum number of crests and troughs of 0 to 10 and multiple crests and troughs after 10.

Density plots of survival (on-job) duration

Bayesian Estimation and Inference

For BL and G2G with covariates, to construct the posterior distributions of covariate effects γ from Bayesian inference, we adopt Gibbs sampling—a Markov Chain Monte Carlo (MCMC) method for model calibration (see [4] for an extensive review). Based on a preliminary test on convergence and distribution of covariate effects, we adopt independent normal priors [15]Footnote 3 with identical parameters. Specifically, we set \(N\left( {0,0.2} \right)\) for G2G and \(N\left( {0,0.5} \right)\) for BL. Based on the simulated posterior distributions, we estimate posterior modes (i.e., maximum a posterior, MAP) and the credible regions of covariate effects. Let \(N\left( \cdot \right)\) be the normal density function, and define collection \(\left\{ \cdot \right\} = \left\{ {\gamma_j , \gamma_{ - j} } \right\}\), where \(\gamma_j\) denotes the jth coefficient being updated and \(\gamma_{ - j}\) be all the other coefficients, the full conditional distribution \(P(\gamma_j |\gamma_{ - j} ,D)\) can be simply written as:

$$ P\left( {\gamma_j |\gamma_{ - j} ,D} \right) \propto P\left( {D| \cdot } \right) \times N\left( {\gamma_j } \right) $$
(6.12)

We calibrate the models on 2000 randomly sampled employees who start their jobs in Year 1, whose survival (employment) durations are censored at week 52. For the models, we run 2000 MCMC iterations and discard the first 1000 draws as burn-in samples. On the other hand, we select employees who join in Year 2 as test samples for the sake of assessing out-of-sample prediction performance. Survival durations for those in test set who survive after week 104 are censored. We apply log-likelihood (LL), Bayesian Information Criterion (BIC), Akaike Information Criterion (AIC) to assess the models’ fitness, and use C-statistics [16] to evaluate the models’ performance on projecting hold-out employees’ job duration:

$$ C = \frac{{\sum_{i \ne j} \left\{ {{I(T_i > T_j ) \cdot I(EL_i > EL_j )}} \right\} \cdot I\left( {C_j = 0} \right)}}{{\sum_{i \ne j} I(T_i > T_j ) \cdot I\left( {C_j = 0} \right)}} $$
(6.13)

of which \(EL\) denotes the expected lifetime assuming no discount (i.e., \(d = 0\)).

Data Analysis

We first assess the computational efficiency that is critical for model adoption and application. In Table 6.2, we report the CPU time it takes to calibrate the foregoing models. We implement the code in R, and execute the programs on a platform with a Win10 operating system, an Intel i7-1165G7 processor, and a 16 GB RAM. Compared with the 2-parameter BG with gradient-based optimization, which takes 35.67 s for calibration, MCMC-based BL (with 30 parameters) and G2G (with 44 parameters) take 701.31 and 10,009.55 s (equivalently, 19.66 and 280.62 times of BG’s execution time). Not surprisingly, the most sophisticated G2G—with time-varying and individual-specific latent churn rate—comes at higher computing cost. Nonetheless, given its unique flexibility in modelling heterogeneous and nonstationary churn distributions, we posit the estimation cost induced by generic formulations is acceptable and affordable in modern computing. We then formally assess the performance gains of G2G.

Table 6.2 Computational cost

Table 6.3 reports the fitness and performance of the foregoing models (i.e., LL, AIC, BIC based on training samples as well as LL and C-index on hold-out samples). Taking BG as our benchmark, we find that modeling time-invariant covariates makes BL rather flexible and able to capture cohort-level stationary heterogeneity, leading to better fitness on in- and hold-out samples. G2G with time-invariant and time-varying covariates enhances performance further and outperforms BL. The improvement of G2G over BL is more salient than that of BL over BG, evidencing the non-trivial value of accounting for time-varying covariates and nonstationary heterogeneity distributions of latent churn.

Table 6.3 Performance evaluation

As for projecting employees’ survival, BL and G2G achieve C-statistics of 0.826 and 0.832, respectively. These are high values considering the high uncertainty of employee churn data and simplicity of the models’ parametric formulations [28]. Because the expected lifetimes that BG produces are not differentiable between individuals, leading to a C-statistic of 0.500, we apply the same protocol to train a survival forest [18]. The survival forest achieves a C-statistic of 0.835. This finding implies that simple parametric yet interpretable formulations do not necessarily undermine prediction accuracy.

Given the significantly better model fitness and prediction performance of G2G, in Table 6.4, we report the MAP effect estimates and 90% credible regions of covariates [2]. For brevity, we report only the top-10 covariates based on their absolute effect size in descending order.

Table 6.4 Effect estimates and credible regions of G2G estimates

As expected, Salary Bonus as a motivating factor outranks other covariates and reduces employee churn. The ratio working on daytime (Shift Ratio) tells if one often works on normal shift hours, thus reducing employee churn. Full-time (R-Category (F)) and dispatch job categories (R-Category (D)), compared to part-time job (the baseline category), are positions seeking regular and long-term workers. Therefore, employees who have full-time and dispatch jobs are less likely to churn. Negative relationship between age and churn rate (Age (≤20), Age (21–30), Age (31–40)) makes sense in that physically fit younger workers are generally better suited for labor-intensive tasks.

Some of the estimated effects are counterintuitive: (a) Why are employees having longer mean working hour (Work Hour) and having higher ratio working on rest day (Rest Day Work) less likely to churn? (b) Additionally, why is the latent churn rate of an employee lower, when the upcoming period is in Summer (Is Summer)? An unpublished interview by the authors’ colleague reveals that the majority of Alpha’s employees are job-hoppers that seek over-time working for extra bonus, thus making (a) reasonable. For (b), it reveals that salary level for the industry is lower during May to June, explaining why employees are unlikely to churn in Summer.

The above discussion focuses specifically on the model fitness and the covariate effects. Below we examine if the estimated effects consistently reflect survival curve at aggregate-level. In Fig. 6.2, we project survival curves using the calibrated BG and G2G models. In addition to the global curve (on the left panel), we take Salary Bonus with the strongest effect as an illustrative example to divide the sample into two sub-groups. We project two local curves, i.e., one with bonus below-median (on the middle) and another with bonus above-median (on the right).

Fig. 6.2
Three line graphs represent B G, G 2 G, and truth line graphs, indicating a full sample in which the 3 curves merge, a lesser value than the median in which the B G curve is on top, and a greater value than or equal to the median in which the truth curve is on top.

The survival curves

For the full sample, except for a few shocks uncaptured, overall both BG and G2G fit the global curve fairly well (with mean discrepancy less than 4%). The result is surprising given the simplicity of BG as we introduced in Sect. 2.2.1. As for the two groups, despite the larger discrepancy between data and model projections, G2G does a decent job at characterizing the two local curves. Arguably, BG will be useful and robust if a decision-maker is only interested in inferring the curve reflecting global survival rate. Nonetheless, the proposed Bayesian G2G will be a more effective and appropriate technique for one to predict the survival rate/remaining job duration of an individual or a specific cohort over time.

To offer managers practical insights in workforce planning, we project survival curves for each individual employee, and by pairwise CORT dissimilarity [3] and hierarchical clustering we cluster the curves. In Fig. 6.3, we show the aggregated survival curves,Footnote 4 and summarize their mean characteristics in Table 6.5. We identify three clusters: High-Risk, Medium-Risk, and Low-Risk. High-Risk has more of 20s and dispatch workers, who seldom work overtime and in irregular hour for extra bonus. Medium-Risk is younger, while having more full-time workers and extra working hours and bonus. The Low-Risk, while being the oldest and having fewer shifts on daytime and rest day, has the highest proportion of full-time workers and the most working hours and bonus. Clearly, the clusters are high in some drivers of survival tendencies but low in the others. For churn anticipation and retention, we suggest one to focus on the factors strictly increasing or decreasing survival risk (e.g., Salary Bonus), or to focus on the combination of drivers (e.g., high Salary Bonus and low Shift Ratio) for the sake of developing countermeasures against employee churn.

Fig. 6.3
The line graph represents three line graphs that relate survival rates to weeks. As the weeks increase, the high-risk graph reaches the lowest value, the medium-risk graph reaches the medium value, and the maximum-value graph reaches the maximum value.

Survival curves clustering

Table 6.5 Averages of features in all clusters

Conclusion

Marketing modelers have invented a lasting stream of contractual churn models and exerted their influences over applications outside the realm of marketing in the past two decades. However, numerous models ignore time-varying covariates by assuming stationary and identical distributions of latent attrition. Such models are thus restrictive and not flexible enough for tenure duration predictions in many business sectors and problem settings. Fader and Hardie [9] propose G2G that offers a generic structure for incorporating time-varying covariates. Instilling Bayesian inference and estimation into G2G, this paper empirically assess the model’s efficacy and the value of time-varying covariates in predicting employee churn, which shares a great similarity with customer churn.

Our analysis of workforce data from manufacturing plants shows that stochastic churn modeling from marketing can be useful for operations management. We find that G2G allowing for nonstationary heterogeneity distributions and time-varying covariates has the best performance in model fitness and survival rate projection. Furthermore, we identify the drivers of employee churn, including factors that are immediately comprehensible (e.g., upcoming seasonality, recruitment category, and age) and counterintuitive ones (mean working hour and working on rest day). When project survival curves globally and locally, the consistently decent performance of G2G suggests that managers can apply G2G to predict churn proportion and remaining lifetime for labor planning and hiring in practice.

Notwithstanding the promising result, our modeling effort is not without limitations and leaves several directions for continual explorations. First, we encourage subsequent studies to triangulate our findings using datasets from different firms and industries. Such efforts will enhance the validity of our Bayesian G2G modeling approach. Second, the fine-tuning of Gamma prior and the MCMC calibration of covariate effects is computationally intensive. Hence, computationally efficient optimization method of G2G deserves more investigations. Third, in this paper, we assume homogeneous employee value and calibrate the models accordingly. Models aimed at delivering maximum workforce performance (e.g., working output and return-on-investment ratio) should take heterogeneous employee value into account. In sum, the probability models provide unique value on prediction and inference front, and G2G having more realistic assumptions is a promising framework in theory and practice of stochastic churn modeling.