Several meta-analytic studies of evaluation research on couple relationship education (CRE) programs have documented their effects on couple outcomes (Arnold & Beelman, 2019; Blanchard et al., 2009; Fawcett et al., 2010; Hawkins et al., 2008; Hawkins & Erickson, 2015; Lucier-Greer & Adler-Baeder, 2012; McAllister et al., 2012; Pinquart & Teubert, 2010). However, to date, none have systematically reviewed whether CRE programs have effects on parenting and child outcomes. This meta-analytic study focuses on a growing body of CRE evaluation studies that have measured effects on a broader set of family outcomes. Our objective is to document the effects of CRE programs on (a) coparenting, (b) parenting, (c) and child well-being and behavior outcomes and what factors moderate any effects. Our focus is on rigorous control-group studies (although we present results for 1-group/pre-post studies—analyzed separately—in an online supplemental appendix).

A relevant backdrop for this review is federal policy supporting CRE programming (Hawkins, 2019; Randles, 2017). Since 2006, the Administration for Children and Families (ACF) has invested more than $1.2 billion in couple relationship education (CRE) and individually-oriented relationship education (RE) for lower income individuals and families to strengthen relationships and increase family stability. About 2.5 million people have participated in these ACF-funded programs at a median cost of about $400 per participant (Hawkins, 2019). The policy—referred to now as the Healthy Marriages and Relationships Education (HMRE) initiative—sparked considerable debate among social scientists and policy scholars (Hawkins, 2019). Early evaluation results of no or small effects generated criticism of the policy (National Academies of Sciences, Engineering, and Medicine, 2019) and calls for abandoning the initiative (Reeves, 2014). Recent results, however, have been more promising (Hawkins, 2019). Many of the studies reviewed here were associated with this federal policy initiative.

From a social policy perspective, the ultimate hope is that stronger couple relationships will improve outcomes for children. That is, an important rationale for this policy initiative is that healthy relationships between parents—regardless of their relationship status—may have spillover effects on parenting and coparenting behavior, which in turn will increase the well-being of disadvantaged children. The logic for this is based in systems theory and a large volume of research that links couple relationship quality to parenting (Carlson & McLanahan, 2006), coparenting (Christopher et al., 2015), and child (Brown, 2010) outcomes. This body of work demonstrates the importance of positive parental relationship quality for effective parenting and child well-being (Cowan & Cowan, 2014; Cummings & Davies, 2002; Knopp et al., 2017). A recent meta-analysis of 230 studies confirms that interparental conflict is concurrently and longitudinally associated with children’s maladjustment (van Eldik et al., 2020). A number of CRE programs now also include some direct instruction on cooperative coparenting and effective parenting. A few CRE evaluation studies have explicitly linked positive intervention changes in couple relationship outcomes to positive changes in parenting and coparenting behaviors, which in turn are associated with positive child outcomes (Fienberg & Jones, 2018; Pruett et al., 2019; Sterrett-Hong et al., 2018; Zemp et al., 2016).

Method

Search Procedure

Meta-analytic studies are exempted from human subjects reviews at the authors’ institution. The researchers did not receive any external funding for the study. The search for relevant studies for this meta-analysis was embedded in a search for a broader set of CRE studies (Hawkins et al., 2021), then modified in the last stages to focus on the specific outcomes of interest. This search process included the following steps. To identify potential studies, we first searched two electronic databases (PsychINFO; Family & Society Studies Worldwide) in May 2020. We experimented with a range of search terms and combinations of terms; we settled on simply “relationship education program,” which proved to be the best starting point because it captured RE evaluation studies while limiting the number of correlational studies. We also supplemented this first-stage electronic search with several other procedures. We looked at several review articles and recently published reports for citations of potential studies. Moreover, we examined references from included studies for potential studies that may have been missed, including so-called “gray literature” (e.g., technical reports, dissertations). Further, we browsed the websites of federally funded CRE programs known to be actively publishing evaluation studies for lists of research reports and published studies. During the process, we also contacted active researchers in this field for clarifications about certain studies (e.g., overlapping samples) and asked them if they had any studies (published or unpublished) not on our accepted-studies list. This yielded several more recent (in press) studies. Unfortunately, research assistants did not account for examined records/studies with each of these steps, a methodological oversight on the first author’s part.

Inclusion/Exclusion Criteria

Outcomes

To be included in the meta-analysis, a study had to be an empirical evaluation of a CRE program and measure at least one of the following constructs: (a) coparenting (e.g., coparenting conflict, competition, positivity, disagreements, hostility); (b) parenting (e.g., harshness, hostility, parental negativity/positivity, responsiveness, supportiveness, father engagement/involvement/care/play); (c) child well-being/behavior outcomes (e.g., externalizing/internalizing behaviors, anxiety/depression, adjustment, aggression, emotion dysregulation, social competence, withdrawal, self-regulation). While it may be appealing to investigate some of these measures separately (e.g., father involvement from parental harshness), doing so risks underpowered analyses.

Relationship intervention

Also, to be included in the study, programs had to give strong attention to the couple relationship. Some programs focus curriculum on more effective parenting and coparenting, with some programs referring to themselves as coparenting education. In these programs, more attention is given to parenting and working together effectively on behalf of their children, but some also include at least an equal weight on couple relationship skills (e.g., Supporting Father Involvement: Cowan et al., 2009; Family Foundations: Feinberg et al., 2010). If so, we included these studies in our meta-analysis.

Study design

We included both control-group studies and 1-group/pre-post studies. In some control-group studies, we could not find evidence that participants were rigorously randomized, but we included them and tested to see if these “quasi-experimental” studies were statistically different from RCT studies. Our primary focus is on rigorous control-group studies that are subject to fewer internal validity problems compared to 1-group/pre-post designs (Reichart, 2019). Still, there have been a significant number of 1-group/pre-post studies in this field over the last five years and they may be able to add a supplemental perspective to the research question. In this context, then, and in field circumstances where program administrators understandably resist excluding interested couples from participating in an intervention to help them strengthen their relationship, examining 1-group/pre-post studies is merited. However, we analyze these studies separately and report the results in online supplemental appendix S2 (along with the study references).

Note that we did not retain four studies that included alternative-treatment comparison groups rather than no-treatment control groups. In each of these excluded studies, we deemed that the alternative treatments were not placebos and could reasonably be expected to produce some similar effects on the outcome variables. Although these were noteworthy studies, group-difference effect sizes would be smaller for these studies and could bias the overall effect size in the meta-analysis.

Couple focus

Some studies focused on individually-oriented relationship literacy RE programs for youth and young adults. These programs do not assume that participants are in couple relationships and do not expect that they attend as couples (although some do). Our focus was on programs that targeted couples, so we excluded individually-oriented RE programs. (And the outcomes they measure are mostly different from couple-oriented programs. See Simpson et al., 2018 for a review of these kinds of programs.) However, in some couple-oriented programs some participants could be attending without their partner.

Qualitative studies

We excluded qualitative evaluations (no effect size to calculate).

Ultimately, we reviewed 662 records and ended up fully coding 40 control-group studies, yielding 406 effect size calculations (and 12 1-group/pre-post studies with 65 effects). This final figure of 40 control-group studies includes adjustments made during the coding process. Some reports included multiple independent sites/samples or reported effects for multiple treatment groups, so they were coded separately. On the other hand, numerous reports from the same research team included samples that overlapped with other reports. Rather than employing the generalized-weights approach described by Bon and Rachinger (2017), we worked to identify overlapping samples and exclude them from our final set of studies. Or, if they added outcomes not reported in previous studies, we combined overlapping-samples studies for coding purposes. This was detail-oriented work that included careful reviews of studies and correspondence with authors. Nevertheless, this overlapping samples work assured us that units of analysis in the study were statistically independent. One study had a longer-term follow-up assessment with the treatment group that was not given to the control group, so we only coded for the immediate assessment. Finally, a few control-group studies had to be excluded because they did not provide sufficient data from which to calculate effect sizes. Fig. 1 graphically summarizes the inclusion/exclusion process and the number of studies excluded at each decision point. (Accepted studies are indicated with * in the reference list). Online supplemental appendix S1 contains a study characteristics table of included studies.

Fig. 1
figure 1

PRISMA flowchart of search process, inclusion/exclusion decisions, and coding adjustments

Variable Coding

We used a 31-item codebook to guide the coding process of methodological, programmatic, and effect size data. Reports were coded independently by four undergraduate research assistants, checked by four trained, advanced MS graduate students. Following independent coding, they met to discuss any coding differences. Although coding discrepancies were not frequent, when they did occur, the original study was examined further and discussed until the coders reached a consensus (sometimes with input from the first author). For instance, a common divergence was with coding sample size. While on the surface coding sample size seems straightforward, recruitment sample sizes often differ from sample sizes at the beginning of the intervention (due to drop-outs) and from the analytic sample (due to measurement non-compliance) for each coded outcome, etc., and coders could code different numbers. A joint inspection of the article focusing on analytic sample size resolves discrepancies.

Computing Effect Sizes

We computed standardized mean differences for control-group studies (and standardized mean gain effect sizes for 1-group/pre-post studies) based on the last outcome assessment available. For some studies this was the immediate post-intervention assessment, but most studies included follow-up assessments of various lengths (86% of effect sizes). We employed Biostat’s Comprehensive Meta-analysis III program to compute effect sizes, which were weighted by the inverse variance, giving greater weight to effects with smaller standard errors (larger samples). We used random effects models to estimate overall effect sizes. Random effects models allow for the possibility that variation in the distribution of effect sizes is a result not only of sampling error but also of differences in programs, research methods, and other factors (Borenstein et al., 2009). Overall effect sizes produced with random effects models are more generalizable to the large variety of CRE programs in the field. We followed up main analyses with Duval and Tweedie’s trim-and-fill analyses to examine potential missing-study bias.

Heterogeneity and Moderator Analyses

We expected significant heterogeneity in the distribution of effect sizes due to the differences in interventions, target populations, methods, and other factors. Accordingly, although we report mean effect sizes, we also give attention to heterogeneity. We report the range of effect sizes to draw attention to potential outlier studies that may indicate especially low or high effects. In addition, we report the prediction intervals (PI) for each of our outcome categories. As a measure of heterogeneity, the prediction interval specifies the two effect size values between which 95% of the true effect sizes would be expected to fall (Borenstein, 2019). When the PI and Q (test of homogeneity of effect sizes) indicated substantial heterogeneity in an outcome, we pursued moderator analyses to explore why some studies had stronger effect sizes. However, we limited moderator analyses to eight planned, substantive variables that likely could explain heterogeneity in the distribution of effect sizes and that had application for practice. (Note that we did not interpret moderator analyses when cells had fewer than five studies contributing to the effect size to diminish the risk of overinterpreting potentially unreliable group differences.) Moderator tests are observational rather than experimental; that is, moderator tests do not reflect an experimental manipulation of the independent variable. Significant moderators should be examined in primary experimental studies before drawing strong causal inferences.

The planned moderator analyses included tests for differences in study methods as well as participant and program characteristics. (a) Relationship status: some studies have found that married participants, who are generally more committed, benefit more from CRE (Hawkins & Erickson, 2015), but other studies have not found relationship status differences (Moore et al., 2018). (b) Relationship distress: many studies have found that couples in more distress at the start of the intervention benefit more from CRE (Amato, 2014; Hawkins et al., 2017), perhaps because they simply have more room to improve. (c) Economic disadvantage: many studies have found that economically disadvantaged couples benefit more from CRE (Hawkins et al., 2017), but one meta-analysis found that studies with samples of mostly near-poor participants did better than studies with samples of mostly poor participants (Hawkins & Erickson, 2015). (d) intervention dosage: some studies have found that programs with low dosage have smaller effects (Hawkins et al., 2008) but there may not be an advantage for the most intensive dosage programs (Stanley et al., 2006). (e) Curriculum content: some CRE programs include explicit coparenting and parenting content (CRE + ) that could boost effects on the targeted outcomes compared to CRE programs with a focus only on the couple romantic relationship. (f) study timing: studies of programs that came later could have learned from earlier programs and strengthened effects. (Among programs funded by ACF, there were explicit efforts to communicate lessons learned and best practices with other program administrators.) (g) Program funding: some evaluation studies were funded directly by ACF-OPRE and conducted by professional policy research organizations with very large budgets (i.e., BSF, SHM, PACT). Also, these studies employed larger samples and multiple (independent) sites. (h) Analyses: some studies employed intent-to-treat (ITT) analyses (outcome analyses on all participants regardless of their level of program participation) while others employed treatment-on-the-treated (TOT) analyses (outcome analyses only on those participants who received a strong dosage of the intervention). Both analytic approaches can yield valuable information; ITT analyses are more conservative and may better estimate the effect of treatment at a population level while ToT analyses may better estimate the effect of the treatment as intended.

Results

Preliminary Analyses

We conducted preliminary analyses to test whether randomized controlled trial (RCT) studies differed significantly from 7 control-group studies for which we could not establish that participants were randomly assigned. These studies can have biased effect sizes but can also be informative and excluding these studies from the meta-analysis could bias results, too (Lipsey & Wilson, 2001). Effect sizes from RCT studies comparing treatment groups to no-treatment control groups were slightly smaller but not statistically different (dexp = 0.052, k = 33; dquasi = 0.152, k = 7; Q = 1.3, p = 0.26). Accordingly, we combined all control-group studies for analyses which provided more statistical power for our main analyses. (Note that 95% of computed effect sizes were from RCT studies).

We conducted power analyses on the set of control-group studies. One study found that five or more studies are needed for meta-analyses to achieve power that is greater than the individual studies that contributed to the overall effect size (Jackson & Turner, 2017). We anticipated small effects on outcomes for this study. For instance, one meta-analysis of couple outcomes with low-income couples assigned to a CRE treatment (compared to no-treatment controls) found a small aggregate effect size for relationship outcomes of d = 0.07 (Hawkins & Erickson, 2015). For control-group studies (k = 40), our power analyses yielded good power of 0.94 to detect a small effect size of 0.07 (with moderate heterogeneity). (For 1-group/pre-post studies (k = 10), however, statistical power was sufficient only when the 3 outcome categories were combined.) We used Tiebel’s meta-analysis power calculator for these calculations: https://www.jtiebel.com/2018/08/26/how-to-calculate-statistical-power-of-a-meta-analysis/.

Analyses for Control-Group Studies

Treatment-versus-no-treatment effect sizes for the three outcome categories and related statistics are presented in Table 1. (Supplemental appendix S3 provides all effect sizes and confidence intervals for all coded studies by outcome.) The average effect sizes for coparenting (d = 0.073, k = 32, p < 0.01) and child well-being/behavior (d = 0.056, k = 28, p < 0.01) were small but statistically non-zero. The average effect size for parenting outcomes (d = 0.023, k = 30, ns) was not significant. We conducted trim-and-fill analyses to examine potential missing-study bias. No bias was detected for coparenting and child outcomes but was detected for parenting, with the adjusted effect size further reduced to virtually zero.

Table 1 Effect sizes for outcome categories

The prediction intervals and Q-statistics of the outcomes revealed substantial heterogeneity only for the coparenting outcome, so we limited the set of planned moderator analyses to this outcome. These analyses are detailed in the upper panel of Table 2. Effect sizes for studies not funded by ACF-OPRE (d = 0.200, k = 14, p < 0.001) were significantly larger than those for the multisite studies funded by ACF-OPRE (d = 0.025, k = 18, p = 0.275; Q = 15.2, p < 0.001). In fact, the effect size for non–ACF-OPRE studies was the largest observed for control-group studies in this meta-analysis. The moderator analyses for study timing showed a similar pattern: program evaluation studies published before 2015, when most of the ACF-OPRE studies took place, had significantly smaller effects (d = 0.033, k = 24, p < 0.152) than those published later (d = 0.197, k = 8, p < 0.001; Q = 11.5, p < 0.001). Although studies that employed TOT analyses (d = 0.155, k = 5, ns) had somewhat larger effect sizes than studies with ITT analyses (d = 0.065, k = 27, p < 0.01), the difference was not statistically significant. Programs that treated married and unmarried couples together (d = 0.147, k = 7, p < 0.01) had larger effects than programs targeting just unmarried couples (d = −0.013, k = 9, p = 0.772; Q = 6.0, p < 0.05), but not statistically higher than programs targeting just married couples (d = 0.091, k = 16, p = 0.003; Q = 0.7, p = 0.415). Effect sizes for moderate-dosage programs (9–19 hours; d = 0.168, k = 9, p < 0.01) were larger than for high-dosage programs (20 + hours; d = 0.054, k = 21, p < 0.05), but not quite statistically different (Q = 3.3, p = 0.069). Finally, CRE + programs (relationship education that included coparenting/parenting education; d = 0.076, k = 24, p < 0.01) did not have significantly larger effect sizes than CRE programs without this added content (d = 0.066, k = 8, ns; Q = 0.4, ns).

Table 2 Results of planned moderator analyses on coparenting outcomes

Results of 1-group/pre-post studies are subject to significant sources of bias (Reichardt, 2019), but we identified 12 of these studies and they do address our research question. However, to avoid distracting readers from the results of the more rigorous control-group studies, we present the results separately in online supplemental appendix S2.

Discussion

An important development in the CRE field over the past 15 years is that researchers have looked systemically beyond how these programs influence couple relationship outcomes to how they subsequently influence coparenting, parenting, and child outcomes, consistent with family systems theory and a large body of research on the effects of parental discord on child development (Coln et al., 2013; Cummings & Davies, 2002; van Eldik et al., 2020). Our meta-analysis reviewed this growing body of evaluation research. Among control-group studies, we found significant but small average effects of CRE programs on coparenting and child well-being/behavior, but nonsignificant effects on parenting.

We acknowledge that the inclusion/exclusion criteria employed for this meta-analysis may bias findings. For instance, studies that include our targeted outcomes may employ different curricula than studies that do not include these outcomes, although most of the included studies used common curricula (e.g., PREP-based programs, Sound Marital House, Together We Can, Family Foundations). Also, our focus was on couple programs rather than the broader set of relationship literacy programs for youth and young adults. Readers should be careful not to overgeneralize the effects seen in this subset of CRE studies to CRE programs in general.

These treatment-to-no-treatment effect sizes for coparenting, parenting, and child well-being/behavior are smaller than the moderate-size effects on couple relationships documented by several meta-analytic studies (Blanchard et al., 2009; Fawcett et al., 2010; Hawkins et al., 2008; Lucier-Greer & Adler-Baeder, 2012; Pinquart & Teubert, 2010). And note that the coparenting effect size in our study is smaller than observed in a meta-analysis focused exclusively on coparenting programs (d = 0.21; Nunes et al., 2021) A number of factors lead us to expect small effect sizes for this set of CRE programs. First, a quarter of the programs analyzed in this meta-analysis did not include specific curriculum on coparenting or parenting. Hence, any effects on these outcomes would be indirect, through improved couple functioning; indirect effects will be weaker than direct effects. Second, most of these studies were conducted in field conditions, not in laboratories with carefully controlled conditions. Field studies of couple interventions usually yield smaller effect sizes than laboratory studies (Bradbury & Bodenmann, 2020). Similarly, program attrition rates are usually higher in field studies than in laboratory studies (e.g., Moore et al., 2012), so many participants do not receive the intended treatment dosage. Third, some couples (especially unmarried couples) who participate in these programs are unsure about the prospects of their current relationship; their participation may be focused on deciding if the relationship has future potential or is too unhealthy to continue. Positive breakups—ending unhealthy relationships with poor prospects—should be seen as a successful outcome of CRE, albeit one that registers statistically as lower scores on outcomes. Finally, most previous meta-analytic studies of CRE programs synthesized studies with middle-class, well-educated samples. In contrast, this meta-analysis was dominated by studies with diverse, lower-income samples. The stressful conditions of these participants’ lives clearly work against maintenance of learned skills (Bradbury & Bodenmann, 2020; Halpern-Meekin, 2019), especially for the most disadvantaged. This is one reason why critics of the policy have predicted that the programs will not be successful (Randles, 2017). Given these realities facing CRE program interventions, we would expect effect sizes to be smaller than in previous studies.

Still, it is helpful to place the effect sizes observed in this study in a broader context by comparing these effects to those of other kinds of family-strengthening programs that receive support from federal policy initiatives. For instance, a rigorous impact evaluation of the well-known Head Start program found that early moderate impacts on cognitive, social, and health outcomes for Head Start children mostly dissipated by the 3rd-grade follow-up (Puma et al., 2012). The few positive effects that were found were generally between d = 0.10 and 0.15, and several negative effects in the same range were observed. Similarly, a rigorous, major impact evaluation of home visiting programs to support low-income new mothers and improve child well-being found few significant impacts at the 18-month follow-up on maternal health, economic self-sufficiency, or child health and development, although there were small, positive impacts on preventing maternal experiences with intimate partner violence and on some parenting outcomes, with effects generally between d = 0.08 and 0.11 (Michalopoulos et al., 2019). Also, a meta-analysis of responsible fatherhood programs for nonresident fathers found an overall effect size of d = 0.10 (Holmes et al., 2020). Effect sizes of CRE programs on coparenting, parenting, and child well-being are similar to those for other important family strengthening programs that target these outcomes more directly.

Note, however, that we observed some important variation in effect sizes in our moderator analyses and variation is often more informative than central tendencies. For instance, there was a significant difference in program effects on coparenting outcomes between the set of ACF-OPRE studies (d = 0.025), which contributed 61% of the total effects in this meta-analysis, and studies conducted outside the purview of ACF-OPRE (d = 0.200). There are likely both methodological and programmatic differences between these two groups of studies that explain the differences but are difficult to tease out. One possibility is that two of three ACF-OPRE studies (BSF and SHM, from which 16 studies came) led out early on in offering CRE to lower income, diverse couples and had a steep learning curve. Most of the non–ACF-OPRE studies in our meta-analysis came later and may have built on the learning curve from earlier programs. (There were explicit ACF efforts to communicate lessons learned to ongoing programs.) Studies conducted earlier rather than later had significantly smaller effects (d = 0.033 vs. d = 0.197) that mirror the difference between OPRE and non-OPRE studies. Optimists could interpret this pattern of results as evidence that programs are getting better at affecting coparenting outcomes. Those earlier studies, however, may have had more disadvantaged participants, on average, and their stressed lives may have restricted their ability to benefit from the programs.

Nevertheless, if the ultimate value of CRE programs is measured by improved lives for children, then these programs will need to strengthen their effects on coparenting, parenting, and child outcomes. Our moderator analyses may provide a few clues on how program administrators might strengthen the effects of CRE programs on coparenting, parenting, and child well-being. Note that studies of programs that included both married and unmarried couples showed significantly stronger effects (d = 0.147) than studies of programs with just unmarried participants (d = −0.013). Some have speculated that mixed-status groups may help unmarried couples gain greater hope and find role models for their relationship aspirations while married couples better sense their relationship success and progress through participation with unmarried couples (Halpern-Meekin, 2019). And it may be especially helpful for unmarried fathers in these groups to see other men who are “digging in” in to strengthen their families.

Note too that higher dosage programs (d = 0.054) provide no significant advantage over moderate dosage programs (d = 0.168) (the small difference is not quite statistically significant). All ACF-OPRE programs and most of the early-stage programs had more intensive curricula, but the evidence from our study does not support the need for higher dosage, a finding consistent with a broader meta-analytic study looking at the effects of CRE on couple outcomes (Hawkins et al., 2012). Perhaps resources that go into more intensive curricula could be retasked for some brief booster sessions, which CRE scholars have encouraged (Halpern-Meekin, 2019).

The fact that the coparenting outcome was significant but the parenting outcome was not might suggest that CRE programs are better equipped to focus on couple outcomes rather than on parenting outcomes (or that participating couples are more attuned to couple issues). Moreover, effects on child well-being/behavior, small as they are, may flow through better couple relationships or improved coparenting practices rather than improved parenting practices. This is speculative, however, and primary studies that explicitly tested possible pathways between intervention effects and the outcomes targeted in this meta-analysis suggest various pathways (Feinberg & Jones, 2018; Pruett et al., 2019; Sterrett-Hong et al., 2018; Zemp et al., 2016). Feinberg and Jones (2018) found that child behavior problems were impacted by a set of changes in both coparenting and parenting outcomes. Pruett et al., 2019 found a significant pathway from reduced couple conflict to reduced harsh parenting then to reduced child behavior problems. Zemp and colleagues (2016) found significant direct paths between improved couple relationships and child behavior problems (without exploring coparenting specifically). Similarly, Sterrett-Hong and colleagues (2018) found significant direct paths between decreased couple conflict and improved child mental health (without exploring coparenting). Our moderator analyses found no difference in effects between programs that included coparenting/parenting content and those that stayed focused on the couple relationship. Again, this might suggest that effects on children can flow simply through improved couple functioning. Our findings, however, should not be taken as challenging other possible pathways through coparenting and parenting practices. Van Elkik and colleagues (2020) have documented many well-tested direct pathways between couple relationships and child well-being, including reductions in parental conflict and improved couple relationships. Going forward, CRE researchers should seek to understand specific (and multiple) pathways for how CRE and CRE + impact children’s well-being. But CRE programs without direct coparenting and parenting curriculum may still have the ability to improve children’s well-being/behavior.

Finally, the findings of our meta-analysis take on greater significance in light of federal policy to provide CRE to disadvantaged couples. While improving the economic circumstances of these couples’ lives is important to reducing daily stress, improving “soft” skills also may help them achieve their aspirations of stable, healthy families and positive child development. Halpern-Meekin (2019), who has studied in depth couples participating in these kinds of programs, found that the primary motivation for their participation was to reduce couple conflict so they can be better parents and provide a more stable and nurturing environment for their children. She points out how small shifts in microlevel interactions, such as learning to take a “time out” to deescalate potential conflict, “can fundamentally change the couple’s day-to-day experience of their relationship—whether it is primarily a source of comfort or conflict” (p. 167). Still, this meta-analytic study documents only small effects of CRE on coparenting, parenting, and child outcomes. And while the trajectory of this work appears to be headed upward, with somewhat stronger program effects apparent in more recent studies, these programs will need to improve their interventions to better achieve the important policy goal of improving children’s well-being.