Abstract
Culturo-behavior science addresses many of the world’s most significant problems and therefore has potential to create world-changing outcomes. Before systems level changes that improve the world can be implemented, however, it is first necessary for the public to know about and take interest in the accomplishments of culturo-behavior science. Measurable evidence that this kind of influence is being achieved is a component of “dissemination impact,” an important but often overlooked form of accountability on sciences that target real-world problems. We describe a method for quantifying some aspects of dissemination impact and present data on how much of this impact has been earned by articles published in Behavior and Social Issues. The results provide considerable food for thought about how culturo-behavior science can proceed toward making the strategic pursuit of dissemination impact a component of its evidence-based practices.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
The culturo-behavior science (CBS) movement in behavior analysis (Cihon & Mattaini, 2020), for which Behavior and Social Issues (BSI) is the primary home, has created a legacy of many valuable contributions. The journal has featured novel conceptual analyses of Big ProblemsFootnote 1; introduced behavior analysts to data and concepts from outside of behavior analysis that are relevant to those problems; and produced some promising empirical analyses related to those problems (for surveys, see Cihon & Mattaini, 2021; Mattaini, 2019). In these ways, the CBS movement has stretched the boundaries of applied behavior analysis, something that is sorely needed in a field that has been criticized for being rigid in its approach. Behavior analysis tends to be methodologically and conceptually conservative, sometimes leaving the impression that topics of study are chosen, not for their relative importance to humankind, but rather at least partly because they fit into a familiar research tradition or a particular style of investigation (Critchfield & Reed, 2017; Friman, 2006; see also Mattaini, 2019). Recently, Heward et al. (2022) catalogued 350 domains of socially important behavior that have been addressed by empirical behavioral analyses. Though the list was long, and included numerous studies aimed at improving individual quality of life, it proved somewhat light on attention to problems at the culturo-behavioral level of analysis. By contrast, CBS has generally been topically and methodologically adventurous, first selecting topics of the highest importance and only then attempting to figure out how to address them (Mattaini, 2019). As many writers in the CBS community have observed, this approach seems absolutely necessary if we are to take seriously Skinner’s (e.g., 1987) famous challenge to “save the world.”Footnote 2 On this basis, it is not unreasonable to describe CBS as potentially the most important thing happening in behavior analysis today.
As integral as an audacious agenda is to “saving the world,” the present article identifies something further that is absolutely needed. This something will, however, be introduced in a roundabout way in order to place CBS into a context that every behavior analyst appreciates: accountability for effectiveness. Skinner (1948) framed this context even before behavior analysis had made the leap from bench to bedside: In Walden Two, he imagined re-engineering society through experimental communities whose practices would be shaped by data – no practice would be sacred unless empirically verified as effective; (e.g., see Glenn, 1986). When applied behavior analysis coalesced as a formal enterprise a generation or two later, it adopted the same perspective, with the only thing stressed as energetically as the need for socially important behavior change being the need to document that change with data (e.g., Baer, et al., 1968). Data-driven practices have also been proposed or established for such diverse pursuits as teaching (Detrich et al., 2013; Spencer et al., 2012), supervising staff (Sellers, et al., 2016; Turner, et al., 2016), managing organizations (Babcock et al., 1998); evaluating graduate programs (Critchfield, 2015), and cross-site scaling up of interventions (Fixsen et al., 2019, Horner et al., 2014). It is no exaggeration to call relying on data to design and refine one’s practices one of the defining features of a behavioral approach.
How does CBS fit into this evidence-based tradition? Empirical accountability requires two things: specifying what outcomes are sought and identifying what evidence describes how well the outcomes have been accomplished (Spencer et al., 2012). Regarding the former, it’s increasingly understood that CBS targets Big Problems such as:
-
Poverty, and growing income and asset disparities
-
Consumer excesses and financial exploitation encouraged by unregulated corporate capitalism
-
Environmental degradation and climate change
-
The escalation of imprisonment, especially among young men of color in the US
-
The widespread failure of public education among disadvantaged populations, despite the availability of effective educational technologies...
-
Collective violence, including political repression, war, genocide, and violence associated with drug, human, and weapons trafficking (Mattaini, 2013, p. 1).
Ultimately, to be called effective, CBS will need to produce implementation evidence (see Malott, 2021; Malott & Glenn, 2006; Spencer et al., 2012) showing that it is making headway toward ameliorating the Big Problems. Skinner, for his part, was impatient about implementation. “It’s not a question of starting,” Skinner (1948) wrote, “The start has been made. It’s a question of what’s to be done from now on” (p. 257). Compared to Skinner, however, the CBS community has been a bit more deliberate and realistic in its approach. A recurring theme in BSI has been that, before “world saving” can commence in a meaningful way, significant preliminary work is needed (e.g., Cihon & Mattaini, 2021; Mattaini, 2010, 2013, 2019; Seniuk et al., 2019). The underlying assumption is that although these activities may not on their own constitute societal change, they plant the seeds of societal change by positioning culturo-behavior scientists to more intelligently steer toward Big Problem solutions.
Preliminary steps are needed, in part, because of a widely recognized constraint on the success of CBS: In the words of Ardila (1990), behavior analysis “has the principles and the laws to change the world, but it doesn’t have the power” (p. 20). Or, as Miller et al. (1992) explained, “The people who implement or decide to implement our procedures are ... caregivers, supervisors, administrators, politicians, parents – ordinary people. We have not been granted dictatorial powers with these people” (Miller et al., 1992, p. 91). Unlike in Walden Two, whose experimental community was imagined into existence, fully formed and operating with the enthusiastic support of its inhabitants, real solutions require societal understanding, acceptance, and interest (see Rogers, 2003). This defines an important type of preliminary milestone for CBS investigations (and one that has generated some traction, e.g., Fawcett et al., 1988; see also recent musings on dissemination efforts by Ardila Sánchez et al., 2020; Mattaini et al., 2020).
For the skeptical reader, we offer the example of the Community Tool Box (https://ctb.ku.edu/en; see Fawcett et al., 2000; Watson-Thompson et al., 2020) as a case for the enduring importance of interest to prospective intervention success. In line with a participatory behavioral community approach, the Tool Box facilitates culturo-behavior science resource accessibility such that communities can take independent and empowered steps toward resolving local instances of Big Problems. No behavior analyst need be present as a catalyst, but logic follows that some member of the local community must be sufficiently convinced of the value of a culturo-behavior science before such practices are adopted. Influence sets the precedent for community interest. We propose that, in a similar fashion, influence is a marker for the ongoing promise for CBS to offer a meaningful solution to any of the world’s Big Problems.
Toward Accountability on Influence
And now, finally, we begin to sidle up to the purpose of the present article. Skinner (e.g., 1978, 1984, 1986, 1993) often observed, in effect, that, “We could do so much good if only the world would let us!” Clearly, a “world-changing” mission like that of CBS requires stimulating interest and investment at a scale needed to set the solutions to Big Problems into motion. In a distinction that is central to the present article’s theme, the quality of a solution and the reach of its influence are more or less independent issues, as Detrich et al. (2013) has observed: “We built a better mousetrap but the world did not beat a path to our door. It is somewhat ironic that what is arguably a science of influence (behavior analysis) has not been more effective at influencing the adoption rate of a science of influence” (p. 541). Of course, behavior analysts have suggested various ways to achieve greater influence (e.g., Cihon et al., 2016; Critchfield, 2011, 2018; Fixsen et al., 2019; Foxx, 1996; Friman, 2006; Hineline, 2018; Lindsley, 1991; Mace & Critchfield, 2010; Poling, 2010; Seniuk et al., 2019; Wolf, 1978). However, for the most part, the effectiveness of these suggestions remains to be systematically evaluated.
In the spirit of the accountability tradition that started with Walden Two, the most important ingredient in meeting any goal, even one as audacious as “saving the world,” is having objective feedback on progress (Spencer et al., 2012). As Mattaini (2009) asserted, CBS “cannot survive without data” (p. 1). Yet, to our knowledge, no systematic data have been presented describing the level of societal influence that CBS has so far attained.Footnote 3 Scholarly citations tell an informative story (e.g., Luke et al., 2017) but may be a bit beside the point of the present discussion because they measure scientists noticing other scientists, and scientists often are a poor proxy for stakeholders (e.g., Detrich, 2018; Mace & Critchfield, 2010; Rogers, 2003). Also available are outcome data showing how well a particular program or intervention worked (e.g., Fixsen & Blase, 2019; Horner et al., 2014) and social validity data showing how much stakeholders liked an intervention (Wolf, 1978). However, these are orthogonal to the problem of up-front influence because they can be obtained only after one has enough influence to implement a solution in the first place. Absent to date has been systematic evidence of how much various kinds of societal stakeholders are noticing advances in CBS.
To quantify this kind of societal attention, one might at least adhere to the general idea behind scholarly citation impact. Citations are how scientists “talk about” their influences, and citation impact reflects the aggregate of these mentions. By extension, to determine how much stakeholders have attended to CBS, it might prove informative to track the extent to which they “talk about” it. For example, how often do professionals writing for official purposes other than publishing scientific reports (e.g., in public policy documents or patent filings) mention CBS? How often do people working in communications mention CBS when writing for the general public (e.g., news articles, blogs, Wikipedia pages)? And how often do everyday people mention CBS when writing for one another (e.g., social media posts)? Early stages of dissemination impact (as contrasted with citation impact; Trueger et al., 2015) would be evident in the aggregate of these kinds of mentions.
Using Altmetric Data to Measure Dissemination Impact
There is, in fact, a convenient way to begin measuring dissemination impact. To promote the concept of accountability on achieving influence, the present article describes it and presents some relevant dissemination-impact data for articles that have appeared in BSI.Footnote 4 The data come from a massive database maintained by a company called Altmetric.com and were obtained using an application called Altmetric ExplorerFootnote 5 on February 12–28, 2022. The database summarizes mentions of journal articles (and some other scholarly outputs) in the types of non-scholarly sources described above (plus a number of others that will not be discussed here). Such data are referred to as altmetricsFootnote 6 (Adie & Roe, 2013).
The Altmetric.com database updates daily to create a “lifetime” record of attention for any article mentioned in its sources, tracing back to the article’s date of publication. At the time this article was written, the database contained about 205 million altmetric mentions of about 20 million scholarly products. The database combines various types of mentions of a given article into an Altmetric Attention ScoreFootnote 7 which, as it gets bigger, indicates more mentions and a higher relative rank of attention among all of the articles in the data base. In the latter case, for instance, among database articles published in July 2021, an Attention Score of 2 defined the median and a score of 10 defined roughly the 75th percentile. To provide some specific examples of Attention Scores, Table 1 lists the BSI articles with the highest ones, as well as their percentile rank for same-aged articles in the database and some of the kinds of mentions they have received. Notice that Attention Scores vary considerably across articles, as do the types of attention comprising those scores.
Perhaps the most important thing to stress about Attention Scores is that they describe a type of impact that is almost completely independent of what scholarly citations describe. To illustrate, Fig. 1 shows the relationship, for all BSI articles in the Altmetric.com data base, between Attention Score and citation count. The weak positive correlation mirrors what has been reported in many previous investigations focusing on a variety of disciplines (e.g., Costas et al., 2015). Put simply, it’s all but impossible to infer an article’s dissemination impact from its citation impact, and vice versa. This makes it vital to directly measure dissemination impact rather than try to infer it from citation counts.
Some types of altmetric mentions (like those in social media, news, and blogs) tend to pile up quickly after an article’s publication, with mention-accumulation half-lives ranging from a few days to a few months. Other types (like those in policy documents, patents, and Wikipedia pages) tend to accumulate more slowly, with half-lives measured in years (Fang & Costas, 2018; Ortega, 2018). In the Altmetric.com database as a whole, different kinds of mentions occur with different relative frequencies that have shifted somewhat over time. Figure 2 shows the prevalence of several types of mentions during the years 2012–2021.Footnote 8 As the figure suggests, mentions overall have been increasing over time, with social media providing the most mentions and Twitter, the largest source of social media mentions, accelerating in prevalence faster than other kinds of mentions. By showing how often different kinds of mentions occur generally, Fig. 2 will provide a good frame of reference for evaluating how BSI’s mentions, specifically, have changed over time, which will be discussed soon.
A final contextual point is that a complete assessment of dissemination impact, as measured via altmetric data, requires more than simply counting mentions. For any given article, it is, of course, important to know what is said. For example, what policy documents mention the article in what capacity? How substantive are mentions of the article in the news, in Wikipedia pages, and in social media posts? Overall, are mentions complimentary or critical? Such details can be mined from the Altmetric.com data base (e.g., see Critchfield & Dixon, 2022), but are beyond the scope of the present discussion. Suffice it to say that before fretting over what has been said about one’s work, it is first necessary to ascertain whether the work was noticed in the first place.
Attention Profile of Behavior and Social Issues
Having now glimpsed the dissemination impact of selected BSI articles (Table 1), let’s elaborate by comparing BSI to some other journals. Deciding what should count as comparison journals is difficult because BSI is the only CBS journal, and also because “saving the world” is complicated: to varying degrees, this is a practitioner mission, a public policy mission, an ABA research mission, and a general behavior analysis mission. Therefore, Table 2 compares BSI mentions, aggregated across all articles published in the journal to date, to those of selected journals relevant to each of those four missions (all N = 5). Most of the comparison journals should be familiar to BSI readers and thereby require no special explanation.
The leftmost data column of Table 2 shows “lifetime” total mentions, those accruing for all of a journal’s published articles. Journals vary considerably in terms of how long they have been publishing and of how many issues and articles they publish per year; obviously, those with more articles have more opportunity to earn mentions. However, the lifetime mention count remains a reasonable way to estimate a journal’s overall influence, as an analogy helps to explain. Polio vaccines have benefitted more people than COVID-19 vaccines. That polio vaccines have existed for longer and been the focus of more public distribution efforts does not diminish this fact. Similarly, for societal mentions of a journal’s contributions, more means more, regardless of other factors.
The second data column of Table 2 normalizes the results a bit by showing the density of attention received by mentioned articles (number of total mentions divided by number of articles that received mentions). We specify this denominator because it isn’t possible to determine from the Altmetric.com database what percentage of a journal’s articles have received mentions, as it lists only articles that have been mentioned at least once.Footnote 9 By focusing only on articles that received mentions, however, the second data column of Table 2 provides a built-in adjustment for the fact that not all of a journal’s articles may have been crafted with dissemination in mind (think of a methodological article like Kaplan et al., 2018; a treatise on technical terms like Baia & Sampaio, 2019; or a state-of-the field discussion like the present article). The interest here, then, is on the level of impact achieved by articles that at least some members of the public have endorsed with their attention.
Among 21 journals, BSI ranks last in terms of lifetime total mentions and mentions per article. To unpack these data a bit, the remaining columns of Table 2 show, for each of the journals, counts of some of the higher-frequency types of mentions in the database (see Fig. 2). This analysis represents the years 2012–2021, and limiting the analysis to the last 10 full calendar years largely avoids the problem of different journals having different lifespans.Footnote 10 Among the journals listed in Table 2, BSI ranks no higher than 18th for any specific form of mention. For interested readers, Table 3 lists all BSI articles that have been mentioned in policy documents, news articles, blogs, and Wikipedia pages.
A different approach to comparing journals is to consider only the most attention-grabbing instances from each, which we define as those earning a journal’s 20 lifetime highest Attention Scores (as per BSI in Table 1). For journals in general, the Attention Score distribution is skewed such that low scores are common and high scores rare (Livas & Delli, 2018), so it’s possible that a journal with limited attention overall could produce viral “greatest hits” that rival those of other journals. However, Fig. 3 provides no support for this hypothesis where BSI is concerned. It compares BSI to some familiar applied and basic/theoretical behavior analysis journals, considering only articles from the interval during which BSI has published under its current name (1991–present; see Luke et al., 2017). Although in some cases ordinate scaling makes details difficult to see, there is no overlap between BSI and any of the other journals at any point in the Top 20 range of Attention Scores (and note that the comparison journals chosen for Fig. 2 are not those from Table 1 with the most dissemination impact). The discrepancy between BSI and higher-attention journals is much greater than what’s shown in Fig. 2 – for instance, compare BSI’s top Attention Score (38) to that of Psychological Science in the Public Interest (2051) or Analyses of Social Issues and Public Policy (3796). BSI’s top Attention Score would be #61 and #24, respectively, for those journals.
It is worth considering that CBS may be a “slow burn” movement that, following a long period of incubation, is only now beginning to find its way to dissemination impact. Figures 4 and 5 test this hypothesis. Figure 4 shows Attention Scores for all BSI articles in the Altmetric.com database with an impact score of 1 or higher, organized by year of publication. Amidst a lot of inter-article variation, which is expected, there is no systematic evidence of increasing dissemination impact across years: The trend line (a least-squares linear fit excluding one outlier data point) is nearly flat.Footnote 11 Figure 5 shows cumulative records of several types of altmetric mentions of BSI (each data point represents cumulative mentions up to and including articles of a given year). A clear positive trend is evident for social media mentions via Twitter, which have increased dramatically in the past 5 or 6 years, and at a faster rate than in the Altmetric.com database overall.Footnote 12 The figure shows no dramatically increasing trend for other types of mentions, despite the increased prevalence of those types in the database overall (Fig. 2). In summary, BSI is gaining Twitter mentions faster than might be expected, but other kinds of mentions more slowly than might be hoped for.
A final point of interest regarding altmetric data concerns who is doing the mentioning. For a given journal, the Altmetric.com database contains a breakdown of this for certain kinds of mentions. For instance, the Table 3 lists the specific mention sources for BSI in the categories of public policy documents, news articles, blogs, and Wikipedia pages. For individual articles, the data base also tallies the types of Twitter and Facebook users that have posted about a given article (self-identified Members of the Public, Scientists, Practitioners, and Science Communicators), and shows the contents of individual tweets. Presumably, the best dissemination impact is shown when favorable mentions come from people other than scientists. Unfortunately, this information is not available in bulk for an overall journal; it must be accessed on an article-by-article basis (see Critchfield & Dixon, 2022).
The database does summarize countries of origin for mentions arising in Twitter, Facebook, news sources, and policy documents. This information is worth noting but must be taken with a grain of salt due to the generic limitation that digital data sources are biased toward countries where computers and internet access are plentiful. From the Altmetric.com data base overall, here is the percentage of origin-labeled mentions arising from four English-favoring, “developed” countries (United Stated, United Kingdom, Canada, and Australia): Twitter (58%), Facebook (68%); News (80%), and Policy Documents (83%). In BSI mentions (Table 1), one sees the same qualitative bias: Twitter (66%), Facebook (90%), News (80%), and Policy Documents (53%). This matters to the extent that a mission of “saving the world” implies all of the world. The four countries just listed comprise less than 6% of the world population, so no matter how much altmetric attention a journal or article can boast of, this represents a tiny slice of the world that needs solutions to Big Problems.
Discussion: Can you “Save the World” Without Engaging it?
No claim is made here that altmetric data tell the whole story of dissemination impact. It is easy to think of many ways, beyond those quantified in altmetric mention counts, that members of the public could demonstrate that they have noticed and understood sociocultural behavior analyses. It is also easy to think of possible validity challenges that might make at least some Attention Scores an imperfect measure of dissemination impact.Footnote 13 And of course there is the demographic problem just noted. In estimating the value of altmetric data as a source of feedback for a developing field that is concerned with dissemination impact, one is reminded of Montrose Wolf’s thoughts about social validity procedures. “Something of social importance would have to be judged by someone as having value to society”, mused Wolf (1978, p. 203, emphasis added), but he could think of no way to measure those judgments except by asking people to make verbal reports. And Wolf knew that verbal-report measurement has many potential problems. In the end, he decided that it was better to know what verbal reports can tell us than not to know anything. Analogously, altmetric data are not perfect, but in the absence of other objective means of measuring dissemination impact, perhaps they will do.
What altmetric data tell us about the dissemination impact of CBS published in BSI provides considerable food for thought. One possible silver lining in the present report is that BSI’s altmetric attention profile is roughly similar to that of several familiar behavior analysis journals. BSI’s mean mentions per article, for instance, is not too different from that of Behavioral Interventions, The Analysis of Verbal Behavior, Journal of the Experimental Analysis of Behavior, and The Psychological Record. When BSI is compared to many other journals, particularly those emphasizing public policy, it seems obvious that there is considerable room for improvement in the level of societal attention that CBS articles attract.
To reiterate a point from our Introduction, societal attention, as defined here, might be considered the low-hanging fruit of dissemination. Even if the entire world were talking constantly about culturo-behavioral behavior analyses, it would not necessarily get “saved” – that would require implementing systems-level interventions. All that is claimed for present purposes is that getting the word out about CBS may promote one day being able to adopt and implement real change, and that a field that is serious about getting the word out should, if it is evidence-based, be systematically monitoring its progress on this front. Consider the present data as providing a baseline from which to move forward.
How might CBS achieve more dissemination impact? Before specific strategies are suggested, a critical step is to embrace the model of accountability already operating in many areas of behavior analysis. In this regard, the most important thing culturo-behavior scientists can do is also the most general: Operate as an “experimental community” whose practices evolve based on objective feedback about their success (e.g., Malott, 2021). This applies both to topics selected for study and means of communicating about them. Regarding topics, it has often been noted that CBS suffers from a shortage of material resources and properly trained professionals (e.g., Mattaini, 2013). This is a strong argument for using the field’s limited person power strategically, for instance, by focusing on a few Big Problems with the greatest chances of remediation, rather than on the whole smorgasbord of them that traditionally have been said to define the field (e.g., see Cihon & Mattaini, 2021; Mattaini, 2013). It’s reasonable to suggest that the Big Problems most likely to be solved are those for which the public shows the most interest in sociocultural behavior analyses. Dissemination impact data may be useful in identifying which Big Problems those are (falling in line with Wolf’s (1978) vision for social validity). We therefore recommend that every CBS scholar dig deep into altmetric data to get an empirical sense of just how much societal interest really exists for their preferred topics of investigation. The results are not always intuitive and, in some cases, may be counter-intuitive.Footnote 14 Tracking these data may better inform effort allocation by guiding both research and dissemination focus.
Regarding communication, we already know that behavior analysts can struggle with making their insights sound palatable to outsiders. Therefore, one critical focus involves how to describe one’s work clearly and persuasively, and behavior analysis is just beginning to grapple with some possible directions for improvement (Becirevic et al., 2016; Chan et al., 2021; Cihon et al., 2016; Critchfield, 2014; Critchfield et al., 2017; Foxx, 1996; Hineline, 2018; Lindsley, 1991; Mace & Critchfield, 2010; Wolf, 1978). Dissemination impact data can be used to glean clues about which communication strategies work best. Such data might also shine light on the value of crossing disciplinary boundaries in collaboration, an arrangement often discussed as having profound benefits for efficiency in communications and overall influence.
Another important focus involves making sure one’s work is displayed where it can be encountered by the right people – and in most cases this does not mean other behavior analysts (e.g., Miller et al., 1992). The present analysis suggests that, compared to BSI, journals with larger and more mainstream circulation bases offer greater potential for attracting attention. From Table 2, for instance, compare BSI’s 382 total altmetric mentions to the 15,709 of Analyses of Social Issues and Public Policy.Footnote 15 Such data suggest that behavior analysis journals may not be the most effective conduit for dissemination.
Here one is reminded of Skinner’s (1993) comment on the tendency for behavior analysts to congregate in comfortable communities of like-minded individuals: “We have been accused of creating our own ghetto.... Rather than break out of the ghetto, I think we should strengthen its walls” (p. 24). We agree with Skinner’s view where the development of solutions is concerned. Getting the analysis of behavior right is a prerequisite to doing anything useful in the world; therefore it makes sense to surround oneself with people who are most likely to help with getting it right. Behavior analysis journals serve the essential function of allowing behavior analytic work to be read by those with the greatest capacity to understand and refine it.
However, we disagree with Skinner where the dissemination of solutions is concerned. Walls do not make good neighborsFootnote 16, at least when the goal is to share scientific insights, and the readership of behavior analysis journals may be limited mainly to behavior analysts (e.g., Critchfield, 2002; Friman, 2014; Krantz, 1971; Poling, 2010), raising the risk that “technologies developed by applied behavior analysts... mostly lie unnoticed in our ever-proliferating professional journals” (Stolz, 1981, pp. 491–492). “Saving the world” therefore requires outreach, and to disseminate their solutions culturo-behavior scientists may need to rely on mainstream journals (for persuasive arguments along these lines, see Friman, 2010, 2014). It is no great secret, of course, that this kind of outreach is difficult. There can be a steep learning curve for writing in a style that will suit mainstream journals (e.g., Friman, 2014). Ditto for designing studies whose methods will receive mainstream approval. However, as with other skill acquisition challenges, the tonic is lots of opportunities to respond under conditions of feedback, and the steeper the expected learning curve, the sooner practice should begin (Ericsson, et al., 2005).
Of course, some culturo-behavior scientists are already publishing in journals outside of behavior analysis, and this creates two opportunities. The first opportunity is to test the speculation that mainstream journals are a superior springboard to dissemination impact. One useful follow-up to the present analysis will be to evaluate the altmetric profiles of CBS articles that already have been published in mainstream journals.Footnote 17 This may provide insights on where CBS articles are achieving the most attention, and on what kinds of CBS work have the best odds of receiving dissemination-impact traction with general audiences. The second opportunity, assuming the first yields anything interesting, is for those who are already achieving dissemination impact via mainstream journals to work with other culturo-behavior scientists to share their outreach expertise, so that more CBS work can be placed in front of more eyes.
Skinner’s (1993) “ghetto” comment, mentioned earlier, raises the possibility that behavioral journals can serve a sort of incubator function, fueling the development of solutions that later are disseminated by other means. This, in turn, raises the question of how one might know when a solution has been sufficiently developed to merit dissemination. Dissemination impact data can provide useful guidance. For instance, preliminary work that begins to attract altmetric mentions may be ripe for taking to the next level. These mentions are objective evidence that a mainstream community took interest in the work and thus may be receptive to adopting and supporting it. By contrast, work that achieves little dissemination impact may require additional development before it is ready to disseminate or, in the spirit of using limited disciplinary resources efficiently (Mattaini, 2013), might be a candidate dropping in order to focus on more promising alternatives.
Where the quest for societal acceptance of behavioral solutions is concerned, it seems a good idea to follow another of Skinner’s admonitions: “Regard no practice as immutable. Change and be ready to change again” (Skinner, 1979, p. 346). Work that meets society’s needs and interests (defined in part by how members of society understand and react to it) has a chance of making a difference. Work that doesn’t is unlikely to better the world and must therefore be either modified or replaced. The key is to know which work is which, and dissemination impact data reveal, in one key way, which efforts do and do not meet society’s needs and interests. In this vein, the present data are a specific instance representing a general ideal. If we have not persuaded that altmetric data provide potentially informative feedback about dissemination impact, let the reader suggest a better source of data. If we have not persuaded that the available data portray CBS as having not yet achieved significant dissemination impact, let the reader make an alternative data-based case. These are matters on which there is room for debate. The take-away should be that, to qualify as an evidence-based behavioral enterprise, and to have the best odds of a successful developmental trajectory, CBS must remain in close contact with, and be responsive to, objective feedback about its progress toward acquiring dissemination impact. On this there should be no debate.
Data Availability
The data analyzed for this investigation are available in the Web of Science and Altmetric.com databases.
Notes
We think “save the world” belongs in air quotes because, in keeping with the present article’s theme of accountability, it’s unclear whether it is possible to ever declare this mission fully accomplished (e.g., Rittel & Webber, 1973). Skinner’s phrase is best regarded as a euphemism for “making some progress on Big Problems.”
True as well for most areas of behavior analysis. For data on the dissemination impact of Behavior Analysis in Practice, see Critchfield et al. (2022).
Our search incorporated BSI Volumes 1 through 30 (1991–2021). The journal’s history actually extends back to 1978 via the predecessor titles Behaviorists for Social Action and Behavior Analysis and Social Action (see Luke et al., 2017; Mattaini & Cihon, 2019). However, the database that we used revealed no evidence of measurable dissemination impact for any article in the predecessor publications, so the present survey is limited to the journal’s current iteration.
The Explorer application is available free to persons with an academic affiliation (https://www.altmetric.com/products/free-tools/). For a way to access altmetric data without the Explorer, see Critchfield & Dixon (submitted).
So, to be clear: The company is Altmetric (capitalized, singular) or Altmetric.com, and the data are altmetrics (lower case, plural).
Different types of mentions are weighted differently in the calculations. For details, see https://help.altmetric.com/support/solutions/articles/6000233311-how-is-the-altmetric-attention-score-calculated-.
This range was chosen because many kinds of electronic sources of mention either did not exist, or were rarely used, before about 2012. Less than 3% of mentions in the Altmetric.com data base date from before then (see https://www.altmetric.com/explorer/timeline).
Examining tables of contents for BSI’s first 30 volumes, we counted 442 published articles of all types. Based on this denominator, about 23% of BSI articles have received altmetric mentions. What’s typical is unclear because we didn’t repeat this tedious evaluation for other journals, but presumably CBS scholars would hope for more than one in four BSI articles get some outside attention.
This is also the timespan during which the most mentions occurred generally. Because the Altmetric.com database is created by monitoring electronic sources, only a small percentage of its mentions date from before 2012.
There are two competing influences in the Altmetric.com database. On the one hand, older articles benefit from having had the most time to accumulate mentions. On the other hand, older articles are disadvantaged because they were published before most of the electronic sources monitored by Altmetric.com existed. We suspect that the latter is the stronger influence given the explosion in recent years of activity in sources like social media and electronic news outlets (see Fig. 2). The expectation, therefore, is that the typical journal would see an increase in mentions per article over the past decade or so.
Whether these mentions are something that CBS scholars should particularly value is something that could be debated, for at least three reasons. First, social media posts can be superficial and unoriginal (e.g., Twitter re-tweets). Second, some posts are “institutional,” that is, originating from professional societies and publishers rather than the stakeholders who are the focus of the present discussion. Third, social media posts don’t necessarily say nice things about a given article. Our view is essentially that, for a fairly young field that is fighting for recognition, attention is better than no attention. If the goal is to be part of a societal conversation, then there may be value in any sort of mention that increases CBS visibility. These points aside, note that social media offer a kind of practical utility that most other kinds of altmetric mentions do not: nearly immediate attentional feedback. Tweets, for instance, can start appearing on the day article is released.
For instance, based on casual observation, it appears that some of the social media attention garnered by some journal articles consists of tweets issued by a publisher or scholarly society, and retweets of those. Since these mentions originate from scholarly sources they clearly violate the spirit of dissemination impact.
For example, we think that one of the most culturally important works in the history of behavior analysis is Hart and Risley’s (1995) Meaningful differences in the everyday experiences of American children. Massive societal ripple effects can be expected from promoting better child language development (e.g., Suskind et al., 2015). Yet, the Altmetric.com database shows exactly zero mentions of this work in public policy documents.
By the way, this phenomenon – of greater attention being available outside of behavior analysis publications than within – is not limited to altmetric attention. Most journals are cited more often than BSI; for instance, compare BSI’s average citations per article (about 2) to that of Behavioural Public Policy (5), Psychology, Public Policy, and Law (27), and Social issues and Policy Review (38). Moreover, most of BSI’s citations (about 97%) come from within behavior analysis (including 31% from BSI itself). We found only a handful of citations of BSI articles in journals devoted to dealing with specific Big Problems (Sustainability, Atmosphere, etc.). Data obtained from Web of Science on February 9, 2022.
Apologies to Robert Frost.
To illustrate the tenor of the proposed analyses, consider as a case study Tony Biglan, a prolific researcher whose work is generally considered to fall under the umbrella of CBS. Biglan has published 2 BSI articles, neither of which has received an altmetric mention. He has published 81 articles in mainstream journals, and at least 11 in behavior analysis journals other than BSI, that have received mentions. Forty of those articles have Attention Scores of at least 10, which places them among the top 25% of all articles in the Altmetric.com data base. Numbers like these, if replicated across other researchers, would support the proposition that mainstream journals may be a preferred way to seek dissemination impact.
References
Adie, E., & Roe, W. (2013). Altmetric: Enriching scholarly content with article-level discussion and metrics. Learned Publishing, 26, 11–17. https://doi.org/10.1087/20130103
Ardila, R. (1990). Walden three: A scientific utopia. Carlton Press.
Ardila Sánchez, J. G., Richling, S. M., Benson, M. L., & Rakos, R. F. (2020). Activism, advocacy, and accompaniment. In T. M. Cihon & M. A. Mattaini (Eds.), Behavior science perspectives on culture and community (pp. 413–436). Springer.
Babcock, R. A., Fleming, R. K., & Oliver, J. R. (1998). OBM and quality improvement systems. Journal of Organizational Behavior Management, 18, 33–59. https://doi.org/10.1300/j075v18n02_03
Baer, D. M., Wolf, M. M., & Risley, T. R. (1968). Some current dimensions of applied behavior analysis. Journal of Applied Behavior Analysis, 1, 91–97. https://doi.org/10.1901/jaba.1968.1-91
Baia, F. H., & Sampaio, A. A. (2019). Distinguishing units of analysis, procedures, and processes in cultural selection: Notes on metacontingency terminology. Behavior and Social Issues, 28, 204–220. https://doi.org/10.1007/s42822-019-00017-8
Becirevic, A., Critchfield, T. S., & Reed, D. D. (2016). On the social acceptability of behavior-analytic terms: Crowdsourced comparisons of lay and technical language. The Behavior Analyst, 39, 305–317. https://doi.org/10.1007/s40614-016-0067-4
Bell, K. M., & Naugle, A. E. (2005). Understanding stay/leave decisions in violent relationships: A behavior analytic approach. Behavior and Social Issues, 14(1), 21–46. https://doi.org/10.5210/bsi.v14i1.119
Chan, S. C., Katz, B. D., Schmidt, D., & Rehfeldt, R. A. (2021). “In” or “out”? An analysis of the use of augmentals in US presidential speeches on the Paris climate agreement. Behavior and Social Issues, 30, 692–711. https://doi.org/10.1007/s42822-021-00070-2
Cihon, T. M., & Mattaini, M. A. (2020). History and progress in cultural and community science. In T. M. Cihon & M. A. Mattaini (Eds.), Behavior science perspectives on culture and community (pp. 1–21). Springer.
Cihon, T. M., & Mattaini, M. A. (2021). The path forward. Behavior and Social Issues, 30, 1–12. https://doi.org/10.1007/s42822-021-00078-8
Cihon, T. M., Cihon, J. H., & Bedient, G. M. (2016). Establishing a common vocabulary of key concepts for the effective implementation of applied behavior analysis. International Electronic Journal of Elementary Education, 9, 337–348.
Costas, R., Zahedi, Z., & Wouters, P. (2015). Do “altmetrics” correlate with citations? Extensive comparison of altmetric indicators with citations from a multidisciplinary perspective. Journal of the Association for Information Science and Technology, 66, 2003–2019. https://doi.org/10.1002/asi.23309
Critchfield, T. S. (2002). Evaluating the function of applied behavior analysis: A bibliometric analysis. Journal of Applied Behavior Analysis, 35(4), 423–426.
Critchfield, T. S. (2011). Translational contributions of the experimental analysis of behavior. The Behavior Analyst, 34, 3–17. https://doi.org/10.1007/bf03392227
Critchfield, T. S. (2014). Ten rules for discussing behavior analysis. Behavior Analysis in Practice, 7, 141–142. https://doi.org/10.1007/s40617-014-0026-z
Critchfield, T. S. (2015). In dreams begin responsibility: Why and how to measure the quality of graduate training in applied behavior analysis. Behavior Analysis in Practice, 8, 123–133. https://doi.org/10.1007/s40617-015-0090-z
Critchfield, T. S. (2018). An emotional appeal for the development of empirical research on narrative. Perspectives on Behavior Science, 41, 575–590. https://doi.org/10.1007/s40614-018-0170-9
Critchfield, T. S., & Dixon, D. R. (2022). A scientist-practitioner's guide to tracking personal dissemination impact using the altmetric attention score and bookmarklet. Behavior Analysis in Practice, 16, 1006–1014. https://doi.org/10.1007/s40617-022-00725-5
Critchfield, T. S., & Reed, D. D. (2017). The fuzzy concept of applied behavior analysis research. The Behavior Analyst, 40, 123–159. https://doi.org/10.1007/s40614-017-0093-x
Critchfield, T. S., Doepke, K. J., Epting, L. K., Becirevic, A., Reed, D. D., Fienup, D. M., Kremsreiter, J. L., & Ecott, C. L. (2017). Normative emotional responses to behavior analysis jargon, or how not to use words to win friends and influence people. Behavior Analysis in Practice, 10, 97–106. https://doi.org/10.1007/s40617-016-0161-9
Critchfield, T. S., Heward, W. L., & Lerman, D. C. (2022). Fifteen years and counting: The dissemination impact of behavior analysis in practice. Behavior Analysis in Practice. Advance online publication. https://doi.org/10.1007/s40617-022-00744-2.
da Rocha, C. A. A., & Hunziker, M. M. H. (2020). A behavior-analytic view on nudges: Individual, technique, and ethics. Behavior and Social Issues, 29(1), 138–161. https://doi.org/10.1007/s42822-020-00037-9
de Fernandes, R. C., & Dittrich, A. (2018). Expanding the behavior-analytic meanings of “freedom”: The contributions of Israel Goldiamond. Behavior and Social Issues, 27(1), 4–19. https://doi.org/10.5210/bsi.v27i0.8248
Detrich, R. (2018). Rethinking dissemination: Storytelling as a part of the repertoire. Perspectives on Behavior Science, 41, 541–549. https://doi.org/10.1007/s40614-018-0160-y
Detrich, R., Slocum, T. A., & Spencer, T. D. (2013). Evidence-based education and best available evidence: Decision-making under conditions of uncertainty. In B. G. Cook, M. Tankersley, & T. J. Landrum (Eds.), Evidence-based practices (pp. 21–24). Emerald Group Publishing Limited.
Ericsson, K. A., Nandagopal, K., & Roring, R. W. (2005). Giftedness viewed from the expert-performance perspective. Journal for the Education of the Gifted, 28, 287–311. https://doi.org/10.4219/jeg-2005-335
Fang, Z., & Costas, R. (2018). Studying the posts accumulation patterns of Altmetric.com data sources. Altmetrics18. Retrieved from http://altmetrics. Org/wp-content/uploads/2018/04/altmetrics18_paper_5_Fang. Pdf.
Fawcett, S. B., Bernstein, G. S., Czyzewski, M. J., Greene, B. F., Hannah, G. T., Iwata, B. A., Jason, L. A., Mathews, R. M., Morris, E. K., Otis-Wilborn, A., Seekins, T., & Winett, R. A. (1988). Behavior analysis and public policy. The Behavior Analyst, 11(1), 11–25. https://doi.org/10.1007/BF03392450
Fawcett, S. B., Francisco, V. T., Schultz, J. A., Berkowitz, B., Wolff, T. J., & Nagy, G. (2000). The community tool box: A web-based resource for building healthier communities. Public Health Reports, 115(2–3), 274–278. https://doi.org/10.1093/phr/115.2.274
Fixsen, D. L., & Blase, K. A. (2019). The teaching-family model: The first 50 years. Perspectives on Behavior Science, 42, 189–211. https://doi.org/10.1007/s40614-018-0168-3
Fixsen, D. L., Blase, K. A., & Van Dyke, M. K. (2019). Implementation Science & Practice. Active Implementation Research Network.
Fournier, A. K., Geller, E. S., & Fortney, E. V. (2007). Human-animal interaction in a prison setting: Impact on criminal behavior, treatment progress, and social skills. Behavior and Social Issues, 16(1), 89–105. https://doi.org/10.5210/bsi.v16i1.385
Foxx, R. M. (1996). Translating the covenant: The behavior analyst as ambassador and translator. The Behavior Analyst, 19, 147–161. https://doi.org/10.1007/BF03393162
Friman, P. C. (2006). The future of applied behavior analysis is under the dome. ABA Newsletter, 29, 4–7.
Friman, P. C. (2010). Come on in, the water is fine: Achieving mainstream relevance through integration with primary medical care. The Behavior Analyst, 33, 19–36. https://doi.org/10.1007/BF03392201
Friman, P. C. (2014). Publishing in journals outside the box: Attaining mainstream prominence requires demonstrations of mainstream relevance. The Behavior Analyst, 37, 73–76. https://doi.org/10.1007/s40614-014-0014-1
Gelino, B. W., Erath, T. G., & Reed, D. D. (2021). Going green: A systematic review of pro environmental empirical research in behavior analysis. Behavior and Social Issues, 30(1), 587–611. https://doi.org/10.1007/s42822-020-00043-x
Glenn, S. S. (1986). Metacontingencies in Walden two. Behavior Analysis and Social Action, 5, 2–8. https://doi.org/10.1007/BF03406059
Greenwood, C. (1997). Classwide peer tutoring. Behavior and Social Issues, 7(1), 53–57. https://doi.org/10.5210/bsi.v7i1.299
Guerin, B., & de Oliveira Ortolan, M. (2017). Analyzing domestic violence behaviors in their contexts: Violence as a continuation of social strategies by other means. Behavior and Social Issues, 26(1), 5–26. https://doi.org/10.5210/bsi.v26i0.6804
Hart, B., & Risley, T. R. (1995). Meaningful differences in the everyday experience of American children. Brookes.
Heward, W. L., Critchfield, T. S., Reed, D. D., & Detrich, R. (2022). ABA from a to Z: Behavior science applied to 350 domains of socially-significant behavior. Perspectives on Behavior Science, 45, 327–351.
Hineline, P. N. (2018). Narrative: Why it’s important, and how it works. Perspectives on Behavior Science, 41, 471–501. https://doi.org/10.1007/s40614-018-0137-x
Horner, R. H., Kincaid, D., Sugai, G., Lewis, T., Eber, L., Barrett, S., Dickey, C. R., Richter, M., Sullivan, E., Boezio, C., Algozzine, B., Reynolds, H., & Johnson, N. (2014). Scaling up school-wide positive behavioral interventions and supports: Experiences of seven states with documented success. Journal of Positive Behavior Interventions, 16, 197–208. https://doi.org/10.1177/1098300713503685
Kaplan, B. A., Gelino, B. W., & Reed, D. D. (2018). A behavioral economic approach to green consumerism: Demand for reusable shopping bags. Behavior and Social Issues, 27, 20–30. https://doi.org/10.5210/bsi.v27i0.8003
Knapp, T. J. (1997). Behaviorism and public policy: B.F. Skinner’s views on gambling. Behavior and Social. Issues, 7(2), 129–139. https://doi.org/10.5210/bsi.v7i2.311
Krantz, D. L. (1971). The separate worlds of operant and non-operant psychology. Journal of Applied Behavior Analysis, 4(1), 61–70.
Lamal, P. (2009). From rats to pigeons to cultural practices: A review of Beyond the box: B.F. Skinner's technology of behavior from laboratory to life, 1950s to 1970s, by Alexandra Rutherford. Behavior and Social Issues, 18(1), 1–3.
Levin, K., Cashore, B., Bernstein, B., & Auld, S. (2012). Overcoming the tragedy of super wicked problems: Constraining our future selves to ameliorate global climate change. Policy Sciences, 45, 123–152. https://doi.org/10.1007/s11077-012-9151-0
Levy, S., Vaidya, J., Dettmering, J., Siebold, A. N., Mittelman, C., & Garner, J. (2019). Addressing social and global issues: Viewing the Syrian refugee crisis through a behavior-analytic lens. Behavior and Social Issues, 28(1), 77–98. https://doi.org/10.1007/s42822-019-00018-7
Lindsley, O. R. (1991). From technical jargon to plain English for application. Journal of Applied Behavior Analysis, 24, 449–458. https://doi.org/10.1901/jaba.1991.24-449
Livas, C., & Delli, K. (2018). Looking beyond traditional metrics in orthodontics: An altmetric study on the most discussed articles on the web. European Journal of Orthodontics, 40, 193–199. https://doi.org/10.1093/ejo/cjx050
Luke, M. M., Roose, K. M., Rakos, R. F., & Mattaini, M. A. (2017). The history and current status of behavior and social issues: 1978–2016. Behavior and Social Issues, 26, 111–127. https://doi.org/10.5210/bsi.v26i0.7728
Mace, F. C., & Critchfield, T. S. (2010). Translational research in behavior analysis: Historical traditions and imperative for the future. Journal of the Experimental Analysis of Behavior, 93, 293–312. https://doi.org/10.1901/jeab.2010.93-293
Malott, M. E. (2021). The nature of culturo-behavioral science interventions. Behavior and Social Issues, 30, 83–93. https://doi.org/10.1007/s42822-021-00081-z
Malott, M. E., & Glenn, S. S. (2006). Targets of intervention in cultural and behavioral change. Behavior and Social Issues, 15, 31–57. https://doi.org/10.5210/bsi.v15i1.344
Mattaini, M. A. (2003). Constructing nonviolent alternatives to collective violence: A scientific strategy. Behavior and Social Issues, 12(2), 148–163. https://doi.org/10.5210/bsi.v12i2.41
Mattaini, M. A. (2009). Editorial: The data are coming! Behavior and Social Issues, 18, 1–3. https://doi.org/10.5210/bsi.v18i1.2828
Mattaini, M. A. (2010). Editorial: Cultural analysis and social change in Medellín. Behavior and Social Issues, 19, 1–5. https://doi.org/10.5210/bsi.v19i0.3346
Mattaini, M. A. (2013). Editorial: Constructing justice. Behavior and Social Issues, 22, 1–4. https://doi.org/10.5210/bsi.v.22i0.5028
Mattaini, M. A. (2019). Out of the lab: Shaping an ecological and constructional cultural systems science. Perspectives on Behavior Science, 42, 713–731. https://doi.org/10.1007/s40614-019-00208-z
Mattaini, M. A., & Cihon, T. M. (2019). A new beginning-and a new commitment. Behavior and Social Issues, 28(1), 1–7.
Mattaini, M. A., & Rehfeldt, R. A. (2020). Rendezvous with truth and discovery. Behavior and Social Issues, 29(1), 1–14. https://doi.org/10.1007/s42822-020-00034-y
Mattaini, M. A., Esquiero-Leal, J. L., Ardila Sánchez, J. G., Richling, S. M., & Ethridge, A. N. (2020). Public policy advocacy in culturo-behavior science. In T. M. Cihon & M. A. Mattaini (Eds.), Behavior science perspectives on culture and community (pp. 385–412). Springer.
Miller, L. K., Altus, D. E., & Welsh, T. M. (1992). Must people be forced to use behavior analysis? A review of Walden three by Ruben Ardila. Behavior and Social Issues, 2, 89–93. https://doi.org/10.5210/bsi.v2i1.177
Molina, B., Deochand, N., & McGee, H. (2019). Learning from experimental communities using behavioral systems analysis. Behavior and Social Issues, 28(1), 174–188. https://doi.org/10.1007/s42822-019-00005-y
Niland, H., & Ortu, D. (2020). Confessions selected by consequences: An operant analysis of false confessions and interrogation techniques. Behavior and Social Issues, 29(1), 162–194. https://doi.org/10.1007/s42822-019-00025-8
Ortega, J. L. (2018). The life cycle of altmetric impact: A longitudinal study of six metrics from PlumX. Journal of Informetrics, 12, 579–589. https://doi.org/10.1016/j.joi.2018.06.001
Poling, A. (2010). Looking to the future: Will behavior analysis survive and prosper? The Behavior Analyst, 33(1), 7–17. https://doi.org/10.1007/BF03392200
Rittel, H. W., & Webber, M. M. (1973). Dilemmas in a general theory of planning. Policy Sciences, 4, 155–169. https://doi.org/10.1007/BF01405730
Rogers, E. M. (2003). Diffusion of innovations ((5th ed.). ed.). Free Press.
Sasson, S., & Paul, L. A. (2014). Labeling acts of sexual violence: What roles do assault characteristics, attitudes, and life experiences play? Behavior and Social Issues, 23(1), 35–49. https://doi.org/10.5210/bsi.v23i0.5215
Schlinger, H. D. (2002). Not so fast, Mr. Pinker: A behaviorist looks at the Blank Slate. A review of Steven Pinker's The blank slate: The modern denial of human nature. Behavior and Social Issues, 12(1), 75–79.
Sellers, T. P., Valentino, A. L., & LeBlanc, L. A. (2016). Recommended practices for individual supervision of aspiring behavior analysts. Behavior Analysis in Practice, 9, 274–286. https://doi.org/10.1007/s40617-016-0110-7
Seniuk, H. A., Cihon, T. M., Benson, M., & Luke, M. M. (2019). Making a footprint in environmental sustainability: A behavioral systems approach to engaging the behavioral community. Perspectives on Behavior Science, 42, 11–926. https://doi.org/10.1007/s40614-019-00233-y
Sidman, M. (1993). Reflections on behavior analysis and coercion. Behavior and Social Issues, 3(1), 75–85. https://doi.org/10.5210/bsi.v3i1.199
Signal, T., & Taylor, N. (2008). Propensity to report intimate partner violence in Australia: Community demographics. Behavior and Social Issues, 17(1), 8–19. https://doi.org/10.5210/bsi.v17i1.1759
Simon, C. (2017). Why Norwegians don’t have their pigs in the forest: Illuminating Nordic “co-operation.” Behavior and Social Issues, 26(1), 172–186. https://doi.org/10.5210/bsi.v26i0.7317
Skinner, B. F. (1948). Walden two. Macmillan.
Skinner, B. F. (1978). Why don’t we use the behavioral sciences? Human Nature, 1, 86–92.
Skinner, B. F. (1979). The shaping of a behaviorist. Knopf.
Skinner, B. F. (1984). The shame of American education. American Psychologist, 39, 947–954. https://doi.org/10.1037/0003-066X.39.9.947
Skinner, B. F. (1986). What is wrong with daily life in the western world? American Psychologist, 41, 568–574. https://doi.org/10.1037/0003-066X.41.5.568
Skinner, B. F. (1987). Why we are not acting to save the world. In B.F. Skinner, upon further reflection (pp. 1–14). Prentice Hall.
Skinner, B. F. (1993). A world of our own. Behaviorology, 1, 1–5.
Spencer, T. D., Detrich, R., & Slocum, T. A. (2012). Evidence-based practice: A framework for making effective decisions. Education and Treatment of Children, 35, 127–151. https://doi.org/10.1353/etc.2012.0013
Stolz, S. B. (1981). Adoption of innovations from applied behavioral research: “Does anybody care?”. Journal of Applied Behavior Analysis, 14(4), 491–505. https://doi.org/10.1901/jaba.1981.14-491
Suskind, D., Suskind, B., & Lewinter-Suskind, L. (2015). Thirty million words: Building a child's brain. Dutton.
Trueger, N. S., Thoma, B., Hsu, C. H., Sullivan, D., Peters, L., & Lin, M. (2015). The Altmetric score: A new measure for article-level dissemination and impact. Annals of Emergency Medicine, 66, 549–553. https://doi.org/10.1016/j.annemergmed.2015.04.022
Tsipursky, G., Votta, F., & Roose, K. M. (2018). Fighting fake news and post-truth politics with behavioral science: The pro-truth pledge. Behavior and Social Issues, 27(1), 47–70. https://doi.org/10.5210/bsi.v27i0.9127
Turner, L. B., Fischer, A. J., & Luiselli, J. K. (2016). Towards a competency-based, ethical, and socially valid approach to the supervision of applied behavior analytic trainees. Behavior Analysis in Practice, 9, 287–298. https://doi.org/10.1007/s40617-016-0121-4
Watson-Thompson, J., Francisco, V. T., & Anderson-Carpenter, K. D. (2020). A behavioral community approach to community health and development: Tools for collaborative action. In T. M. Cihon & M. A. Mattaini (Eds.), Behavior science perspectives on culture and community (pp. 333–356). Springer.
Watson-Thompson, J., Rakos, R. F., & Anderson-Carpenter, K. (2021). An editorial on revitalizing behavioral community psychology: Where do we go from here? Behavior and Social Issues, 30(1), 514–524. https://doi.org/10.1007/s42822-021-00080-0
Wolf, M. M. (1978). Social validity: The case for subjective measurement or how applied behavior analysis is finding its heart. Journal of Applied Behavior Analysis, 11, 203–214. https://doi.org/10.1901/jaba.1978.11-203
Wyatt, W. J. (2002). What was under the McMartin preschool? A review and behavioral analysis of the “tunnels” find. Behavior and Social Issues, 12(1), 29–39. https://doi.org/10.5210/bsi.v12i1.77
Ward, T. A., Eastman, R. L., & Ninness, C. (2009). An experimental analysis of cultural materialism: The effects of various modes of production on resource sharing. Behavior and Social Issues, 18(1), 58–80. https://doi.org/10.5210/bsi.v18i1.195
Funding
No external funding was associated with this work.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Ethical Review and Human Subjects Protections
Informed Consent and Ethics Board review are not relevant to this work because the data come from public, archival sources rather than original investigations with human or nonhuman participants.
Competing Interests
The authors declare no competing interest, financial or otherwise, related to this article.
Supplementary Information
ESM 1
(PDF 93 kb)
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Gelino, B.W., Critchfield, T.S. & Reed, D.D. Measuring the Dissemination Impact of Culturo-Behavioral Science. Behav. Soc. Iss. 32, 88–114 (2023). https://doi.org/10.1007/s42822-022-00120-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42822-022-00120-3