Introduction

Reading comprehension is a complex process that involves various competencies (Perfetti & Stafura, 2014). Research on the contributions of these competencies has primarily involved students in the K-12 grades and postsecondary programs (e.g., Aaron, Joshi, Gooden, & Bentum, 2008; Hannon, 2012) and has largely ignored the one in five adults in the United States who read at or below elementary levels (National Center for Education Statistics, 2019). One major challenge facing researchers and practitioners is the heterogeneity of the struggling adult reader population, especially in terms of reading comprehension (Comings & Soricone, 2007; Lesgold & Welch-Ross, 2012). Moreover, complicating the heterogeneity challenge, some norm-referenced assessments do not seem to function as expected for adults who struggle with reading (Nanda, Greenberg, & Morris, 2014; Pae, Greenberg, & Williams, 2012). The goal of this study was to address these issues by examining the effects of reading-related competencies on the reading comprehension performance of struggling adult readers across (a) different reading comprehension tests and (b) different levels of performance.

A theoretical framework for reading comprehension

Since there is no reading comprehension framework specific to struggling adult readers, we turn to theories that have been validated with children and adolescents who read below the high school level. The influential Simple View of Reading (SVR) emphasizes the contributions of decoding and linguistic comprehension to reading comprehension performance (Gough & Tunmer, 1986). Decoding refers to the basic process of translating print words to spoken language. Linguistic comprehension aids the reader in understanding oral language and can be subdivided into meaning-making at the word level (i.e., oral vocabulary) as well as the discourse level (i.e., listening comprehension) (Sabatini, Sawaki, Shore, & Scarborough, 2010; Tunmer & Chapman, 2012). Additionally, reading fluency may have a place in the SVR, because being able to read words efficiently frees up attentional resources and allows the reader to focus on processing the meaning of the text (Tilstra, McMaster, Van den Broek, Kendeou, & Rapp, 2009).

Moving beyond the SVR, the Direct and Inferential Mediation (DIME) framework highlights the importance of background knowledge and inference generation to reading comprehension (Cromley & Azevedo, 2007). Background knowledge, or prior knowledge, refers to the generic and specific knowledge encoded in an individual’s long-term memory (Graesser, Singer, & Trabasso, 1994). Inferencing refers to the skill of understanding implicit associations within a discourse (Kintsch, 1988). These competencies facilitate deep comprehension of the text content and help the reader make connections between what is known and what is learned. When the DIME framework was tested with children and adolescents, both background knowledge and inference explained variance in reading comprehension after controlling for decoding and oral vocabulary (Ahmed et al., 2016; Cromley & Azevedo, 2007).

Together, the SVR and DIME frameworks point to different competencies that may influence reading comprehension: decoding, oral vocabulary, listening comprehension, fluency, background knowledge, and inference. Across multiple investigations involving struggling adult readers, reading comprehension exhibited moderate to strong correlations with decoding, oral vocabulary, listening comprehension, and fluency (e.g., Braze et al., 2016; Mellard, Fall, & Woods, 2010; Talwar, Tighe, & Greenberg, 2018). Moreover, each of these four skills uniquely accounted for variance in reading comprehension scores for struggling adult readers (Braze et al., 2016; Greenberg, Levy, Rasher, Kim, Carter, & Berbaum, 2010; Mellard et al., 2010; Sabatini et al., 2010; Tighe, Little, Arrastia-Chisholm, Schatschneider, Diehm, Quinn et al. 2019). Although less work with this population has focused on background knowledge and inference, the limited findings suggest that both competencies are uniquely important to reading comprehension over and above the skills associated with the SVR (Talwar et al., 2018; Tighe, Johnson, & McNamara, 2017).

Reading comprehension tests

The measurement of the reading comprehension construct is an important issue to consider. Past research suggests that different reading comprehension assessments are not equivalent to one another. Multiple studies have shown that examinees can successfully answer some multiple-choice questions even without reading the source passage (Coleman, Lindstrom, Nelson, Lindstrom, & Gregg, 2010; Katz, Lautenschlager, Blackburn, & Harris, 1990; Keenan & Betjemann, 2006). Additionally, reading comprehension tests may be differentially impacted by word reading and oral language skills (Cutting & Scarborough, 2006; Keenan, Betjemann, & Olson, 2008; Nation & Snowling, 1997).

Specific to struggling adult readers, Mellard, Woods, Desa, and Vuyk (2015) reported that vocabulary made the largest unique contribution to reading comprehension performance on the Test of Adult Basic Education (TABE; CTB/McGraw-Hill, 1996), whereas working memory made the largest unique contribution to reading comprehension performance on the Comprehensive Adult Student Assessment System (CASAS, 2004). Similarly, Tighe et al. (2017) found that vocabulary was a significant predictor of performance on the Woodcock Reading Mastery Test Passage Comprehension subtest (Woodcock, 2011) but not on the Gates–MacGinitie Reading Test (MacGinitie, MacGinitie, Maria, & Dreyer, 2000), whereas the opposite pattern was observed for inferencing. These findings support the notion that literacy research and assessment should include multiple measures of reading comprehension and examine the constructs tapped by each measure (Keenan, 2016).

As computerized testing has become more common, it is valuable to explore reading performance across modes of administration. Some researchers have observed with children and proficient adult readers that examinees score similarly on paper-based and computer-based tests targeting the same domain (Achtyes et al., 2015; Bodmann & Robinson, 2004; Srivastava & Gray, 2012). However, other investigations indicate that adolescents are less likely to identify important information within a passage if it is presented on a screen versus paper (Kobrin & Young, 2003) and adolescents with higher test anxiety generally perform worse on computerized tests (Lu, Hu, Gao, & Kinshuk, 2016). In fact, according to a recent meta-analysis that largely included studies with college students, reading from screens is associated with lower comprehension performance compared to reading from paper, especially for expository texts (Clinton, 2019). Such mode comparison studies have not been conducted specifically with struggling adult readers. Individuals in this population can lack basic computer skills such as identifying specific keys on the keyboard and right-clicking the mouse (Olney, Bakhtiari, Greenberg, & Graesser, 2017), and it is important to investigate whether these deficits are related to performance on computerized tests.

The utility of quantile regression for reading comprehension research

As mentioned previously, struggling adult readers comprise a heterogenous population (Comings & Soricone, 2007; Lesgold & Welch-Ross, 2012). Quantile regression presents a unique way to address the variability in reading levels among these adults. Unlike ordinary least squares (OLS) regression, which estimates the effect of a predictor at the average level of the criterion, quantile regression can estimate this effect at different quantiles (or locations) in the criterion distribution (Davino, Furno, & Vistocco, 2014; Koenker & Basset, 1978; Koenker & Hallock, 2001). With school-age readers, quantile regression analyses have shown that the contributions of vocabulary, oral fluency, and motivation to reading comprehension can differ for readers at different proficiency levels (e.g., Cho, Capin, Roberts, & Vaughn, 2018; Frijters et al., 2018; Lonigan, Burgess, & Schatschneider, 2018). Most pertinently, Hua and Keenan (2017) examined the relations of word recognition and listening comprehension to reading comprehension across five reading comprehension measures. From the 0.1 quantile to the 0.9 quantile, the unique effect of listening comprehension increased on two tests, decreased on two other tests, and stayed the same on one test. Mixed trends were also observed across tests for the effect of word recognition.

Similar quantile regression work with struggling adult readers is strikingly sparse. As a proof of concept, Tighe and Schatschneider (2016) demonstrated with a sample of struggling adult readers that two component skills exhibit differential effects on reading comprehension at different quantiles of performance. Given that the adult literacy field serves a heterogeneous group of learners (Comings & Soricone, 2007; Lesgold & Welch-Ross, 2012), a high research priority should be the identification of competencies that are important at different reading proficiency levels.

The current study

Our main goal in this exploratory study was to examine the importance of different reading-related competencies for struggling adult readers, a historically underserved and understudied population. Informed by the SVR and DIME frameworks, we explored the contributions of decoding, oral vocabulary, reading fluency, listening comprehension, background knowledge, and inference to reading comprehension. We explored these contributions for different reading comprehension measures and different performance quantiles within each measure, based on prior research reporting differential findings across reading tests and proficiency levels (Hua & Keenan, 2017; Mellard et al., 2015; Tighe & Schatschneider, 2016). Our approach addressed gaps in the literature involving competencies that are understudied among struggling adult readers (e.g., background knowledge, inference) and the relative importance of different competencies across reading comprehension performance levels within this population. Two research questions guided our investigation:

  1. 1.

    For struggling adult readers, what are the joint and unique contributions of decoding, oral vocabulary, fluency, listening comprehension, background knowledge and inference to performance on different reading comprehension tests?

  2. 2.

    To what extent do the effects of these predictors vary across levels of performance on different reading comprehension tests?

Methods

Participants

As part of a larger project (Institute of Education Sciences, U.S. Department of Education, Grant R305C120001), data were collected from individuals who attended adult literacy classes that targeted the third through seventh grade reading levels. For the purposes of the current study, we limited the sample to 168 adults who were native English speakers and had completed three reading comprehension measures. The sample included individuals who were 16 or 17 years old, since anyone who is at least 16 years of age and not attending high school can enroll in an adult literacy program. The mean age was 42.19 years (SD = 14.39). Participants attended adult literacy programs in two major metropolitan areas in the Southeast of the United States (57.1%) and in Ontario, Canada (42.9%). The major demographic groups were women (71.4%) and individuals who identified as Black or African American (76.8%). About a quarter (24.4%) of the sample reported having a paying job. The majority (73.8%) reported receiving at least one form of government support (e.g., food assistance, disability assistance). More detailed demographic information is reported in Table 1.

Table 1 Demographic information

Procedure

Trained graduate research assistants administered measures of reading comprehension, decoding, oral vocabulary, fluency, listening comprehension, background knowledge, and inferencing to participants at their adult literacy program sites. Test administration was started at easier items to reduce testing anxiety. Basal and ceiling rules were followed. Each item was scored as correct or incorrect according to the scoring rules in the publishers’ examiner manuals. All scored measures underwent two additional rounds of checking. In addition to completing literacy measures, participants also answered questions about their demographic backgrounds and computer experience.

Measures

Reading comprehension

All participants completed three measures of reading comprehension.

WJ Passage Comprehension

One of the reading comprehension measures was the Passage Comprehension subtest of the Woodcock–Johnson III Normative Update (WJ; Woodcock, McGrew, & Mather, 2007). The items in the WJ Passage Comprehension (WJ-PC) subtest were connected texts comprised of one or two sentences with missing words indicated by blanks. The participant silently read each item and filled in the blank by speaking the missing word out loud. Administration started at Item 14. This measure was standardized on individuals 2 years old to over 80 years old and the internal reliability estimates ranged from 0.73 to 0.96 (McGrew, Schrank, & Woodcock, 2007).

RISE Reading Comprehension

Another reading comprehension measure was the Reading Comprehension subtest of the Reading Inventory and Scholastic Assessment (RISE), a Web-based assessment battery developed by the Educational Testing Service (Sabatini, Bruce, Steinberg, & Weeks, 2015). On the RISE Reading Comprehension (RISE-RC), the participant was instructed to read passages and answer multiple-choice questions about each passage with the passage still in view. All questions included three answer choices and the participant selected the answer by pressing the 1, 2, or 3 keys on the keyboard.

We used a version of the RISE battery that had been previously used extensively with adolescent readers. Internal reliability estimates on RISE-RC ranged from 0.72 to 0.82 for students in the sixth through eighth grades (Sabatini et al., 2015). The current study was the first to administer this version of the RISE to struggling adult readers. Each adult was administered one of two test forms. Therefore, their RISE-RC performance was analyzed using Scale Scores instead of raw scores. The IRT-based Scale Scores provide a way to link the two forms and ensure that all examinees’ performance is on a common scale (Sabatini et al., 2015).

RAPID Reading Comprehension

The third reading comprehension measure was the Reading Comprehension subtest of the Reading Assessment for Prescriptive Instructional Data (RAPID), a Web-based assessment battery developed by Lexia Learning (Foorman, Petscher, & Schatschneider, 2017). On the RAPID Reading Comprehension (RAPID-RC), the participant was instructed to read passages and answer multiple-choice questions about each passage with the passage still in view. All questions included four answer choices and the participant selected the answer using mouse clicks. Because the RAPID-RC is an adaptive test, not all participants were administered the same passages. Each participant’s starting passage was determined by performance on the other RAPID subtests, and administration continued until a precise estimate of performance was reached or the participant completed three passages.

To our knowledge, the current study was the first to administer the RAPID battery to struggling adult readers. Internal reliability estimates on RAPID-RC ranged from 0.83 to 0.93 for students in the third through ninth grades (Foorman et al., 2017). As recommended in the technical manual, the RAPID-RC Performance Scores were used for analyses (Foorman et al., 2017).

Decoding

WJ Word Attack

In the WJ Word Attack subtest, the participant read pseudowords out loud. Administration started at Item 4. This measure was standardized on individuals 4 years old to over 80 years old and the internal reliability estimates ranged from 0.78 to 0.94 (McGrew et al., 2007).

WJ Letter-Word Identification

In the WJ Letter-Word Identification subtest, the participant read real words out loud. Administration started at Item 33. This measure was standardized on individuals 4 years old to over 80 years old and the internal reliability estimates ranged from 0.88 to 0.99 (McGrew et al., 2007).

Challenge Word Test

In the Challenge Word Test (Lovett et al., 1994, 2000), the participant read aloud real words that contain multiple syllables. Administration started at Item 1. Because this is an experimental measure, no standardization information is available.

Oral vocabulary

WJ Picture Vocabulary

In the WJ Picture Vocabulary subtest, the participant looked at pictures and named the depicted objects or actions. Administration started at Item 15. This measure was standardized on individuals 2 years old to over 80 years old and the internal reliability estimates ranged from 0.70 to 0.93 (McGrew et al., 2007).

Fluency

WJ Reading Fluency

In the WJ Reading Fluency subtest, the participant was given a list of statements printed on paper and given 3 min to silently read as many statements as possible, decide if each statement is true or false, and circle “Y” (for “yes”) or “N” (for “No”) next to each statement. These statements were short, isolated sentences (3–11 words long) and consisted of high-frequency words. The technical manual describes WJ Reading Fluency as a measure of reading speed and semantic processing speed (McGrew et al., 2007). Informed by the manual, the item stimuli, and the timed nature of the task, we treated this measure as an assessment of reading efficiency. Administration started at Item 1. This measure was standardized on individuals 6 years old to over 80 years old and the internal reliability estimates ranged from 0.72 to 0.96 (McGrew et al., 2007).

Listening comprehension

CELF Understanding Spoken Paragraphs

In the Understanding Spoken Paragraphs subtest of the Clinical Evaluation of Language Fundamentals IV (CELF; Semel, Wiig, & Secord, 2003a), the participant listened to very short stories and then answered questions about them. Administration started at Item 1. This measure was standardized on individuals 5–21 years old and the internal reliability estimates ranged from 0.54 to 0.81 (Semel et al., 2003b).

Background knowledge

WJ General Information

The WJ General Information subtest had two subscales: Where and What. The examiner asked questions about where one would usually find certain objects and what one would usually do with certain objects, and the participant provides answered verbally. For both subscales, administration started at Item 1. This measure was standardized on individuals 2 years old to over 80 years old and the internal reliability estimates ranged from 0.82 to 0.96 (McGrew et al., 2007).

Inference

CASL Inference

In the Inference subtest of the Comprehensive Assessment of Spoken Language (CASL; Carrow-Woolfolk, 1999), the examiner read aloud short passages that have missing information and asked a question about the missing information in each item. The participant answered the question using world knowledge or clues in the passage. Administration started at Item 1. This measure was standardized on individuals 7–18 years old and the internal reliability estimates ranged from 0.86 to 0.90 (Carrow-Woolfolk, 2008).

Data analysis strategy

To answer the first research question, OLS regression models were estimated separately for each reading comprehension test. The dependent variable in this model was reading comprehension and the independent variables were decoding, oral vocabulary, fluency, listening comprehension, background knowledge and inference. All independent variables were entered in one step into the model, based on the absence of control variables and the heterogeneous reader profiles that have been observed in the struggling adult reader population (MacArthur, Konold, Glutting, & Alamprese, 2012; Mellard, Woods, & Lee, 2016; Talwar, Greenberg, & Li, 2020). For the significance testing of each independent variable, an alpha level of 0.05 was used in the analysis.

Prior to estimating OLS regression models, the data were examined to determine whether the assumptions of linear regression were tenable. For each reading comprehension test, scatter plots indicated approximately linear relations with all independent variables. Scatter plots of residuals and fitted values did not exhibit a distinct pattern, which supported the assumption of homoscedasticity. Additionally, residuals appeared to be normally distributed as indicated by Q–Q plots and residual means of approximately zero.

To answer the second research question, quantile regression models were estimated separately for each reading comprehension test using the quantreg package (Koenker, 2020) in the R statistical environment (R Core Team, 2020). As in the OLS regression model, the dependent variable was reading comprehension and the independent variables were decoding, oral vocabulary, fluency, listening comprehension, background knowledge and inference. Model parameters were estimated at the 0.1, 0.5, and 0.9 quantiles, which correspond to low, average, and high levels of reading comprehension within the sample (Hua & Keenan, 2017). Additionally, between-quantile slope comparisons were conducted across the 0.1, 0.5, and 0.9 quantiles for each model.

Results

The reliability and mean performance of the current sample are reported in Table 2. Reliability for WJ-PC was estimated at 0.81. Reliability for RISE-RC was estimated at 0.63, which is similar to the estimate of 0.60 reported for fifth grade students but lower than the estimates for middle school students (Sabatini et al., 2015). Reliability could not be computed for RAPID-RC because we did not have access to item-level data from this adaptive assessment.

Table 2 Performance on measures

Correlations across measures are reported in Table 3. The correlation coefficients among WJ Letter-Word Identification, WJ Word Attack, and the Challenge Word Test ranged from 0.79 to 0.90. Due to these strong relationships, a decoding composite was computed for subsequent analyses by taking the mean of z-scores on these measures.

Table 3 Correlations across measures

Computer experience

Since RISE-RC and RAPID-RC are administered at computers, it was important to examine whether participants’ performance on these assessments were related to their computer experience. All but two participants indicated that they had used a computer before. Of those who had used a computer, approximately 41% said that they use a computer every day, 31% said that they use a computer a few times a week, 13% said that they use a computer once a week, and 15% said that they use a computer less than once a week. One-way ANOVAs indicated that there were no significant differences among these respondent groups on RISE-RC (F(3,162) = 2.59, p > 0.05) and RAPID-RC (F(3,162) = 2.01, p > 0.05). Additionally, participants reported the number of hours they usually use a computer per day, which ranged from zero to 18 h, with a mean of 2.93 h (SD = 3.19). Number of hours of computer use was not significantly correlated with scores on either test (ps > 0.05). Since computer experience did not appear to be related to performance on RISE-RC and RAPID-RC, it was not included as a covariate in any subsequent analyses.

Research Question 1: What are the joint and unique contributions of reading-related skills to performance on different reading comprehension tests?

The parameter estimates of the OLS regression model for each reading comprehension test are reported in Table 4.

Table 4 OLS regression estimates

OLS Regression model for WJ Passage Comprehension

The model explained 66% of variance in WJ-PC performance (F(6,153) = 48.99, p < 0.001). Decoding, oral vocabulary, fluency, background knowledge, and inference had significant unique effects on reading comprehension scores. Approximately 10% of the reading comprehension variance was uniquely contributed by decoding, 2% by fluency, 2% by inference, 1% by background knowledge, and 1% by oral vocabulary.

OLS Regression model for RISE Reading Comprehension

The same model was estimated for RISE-RC and explained 31% of variance in reading comprehension performance (F(6,153) = 11.65, p < 0.001). Only decoding had a significant unique effect on reading comprehension scores and uniquely contributed 12% of the reading comprehension variance.

OLS Regression model for RAPID Reading Comprehension

Finally, the OLS regression model was estimated for RAPID-RC and explained 43% of variance in reading comprehension performance (F(6,153) = 19.09, p < 0.001). Decoding and listening comprehension had significant unique effects on reading comprehension score. Approximately 14% of the reading comprehension variance was uniquely contributed by decoding and 4% was uniquely contributed by listening comprehension.

Research Question 2: Do the effects of predictors vary across levels of performance on each reading comprehension tests?

The quantile regression estimates are reported in Table 5.

Table 5 Quantile regression estimates

Quantile regression model for WJ passage comprehension

For WJ-PC scores, the quantile regression model explained approximately 43% of the variance at the 0.1 quantile, 42% at the 0.5 quantile, and 49% at the 0.9 quantile. At the 0.1 and 0.5 quantiles, significant predictors were limited to decoding and oral vocabulary. At the 0.9 quantile, the significant predictors were decoding and background knowledge. Between-quantile slope comparisons indicated that the unique effect of decoding was stable across the 0.1, 0.5, and 0.9 quantiles of WJ-PC performance (ps > 0.05). The effect of oral vocabulary on reading comprehension was greater at the 0.5 quantile than at the 0.9 quantile (p = 0.022). Additionally, the effect of background knowledge on reading comprehension was greater at the 0.9 quantile than at the 0.1 quantile (p < 0.001).

Quantile regression model for RISE reading comprehension

For RISE-RC scores, the quantile regression model explained approximately 15% of the variance at the 0.1 quantile, 17% at the 0.5 quantile, and 23% at the 0.9 quantile. At the 0.1 and 0.5 quantiles, decoding emerged as the only significant predictor of reading comprehension. At the 0.9 quantile, the significant predictors were decoding and background knowledge. Between-quantile slope comparisons indicated that the effect of decoding on reading comprehension was stable across quantiles (ps > 0.05), while the effect of background knowledge increased in magnitude from the 0.5 quantile to the 0.9 quantile (p = 0.039).

Quantile regression model for RAPID reading comprehension

For RAPID-RC scores, the quantile regression model explained approximately 14% of the variance at the 0.1 quantile, 28% at the 0.5 quantile, and 35% at the 0.9 quantile. At the 0.1 and 0.5 quantiles, decoding was the only significant predictor of reading comprehension. At the 0.9 quantile, decoding and listening comprehension emerged as significant predictors. Between-quantile slope comparisons indicated that the effect of decoding on reading comprehension increased in magnitude from the 0.1 quantile to the 0.5 quantile (p = 0.006), from the 0.1 quantile to the 0.9 quantile (p < 0.001), and from the 0.5 quantile to the 0.9 quantile (p = 0.003). Additionally, the effect of listening comprehension was greater at the 0.9 quantile than at the 0.1 quantile (p = 0.040) of reading comprehension.

Discussion

The aim of this study was to examine the shared and unique explanatory effects of reading-related competencies on reading comprehension performance across different tests and proficiency levels for a sample of adults who struggle with reading. Based on the SVR and DIME frameworks, we focused on the constructs of decoding, oral vocabulary, listening comprehension, fluency, background knowledge, and inference. The results suggest that the effects of certain competencies on reading comprehension were influenced by how comprehension was assessed. Furthermore, the magnitude of some effects changed across levels of performance. The only common finding across tests was that decoding made the largest unique contribution to reading comprehension performance.

Differential effects across tests

Similar to research with children (Cutting & Scarborough, 2006; Keenan et al., 2008), the current study found differences in predictor-comprehension relations across reading comprehension tests. With the exception of listening comprehension, all the other predictors made significant unique contributions to WJ-PC performance. Significant contributors to reading comprehension performance were limited to decoding and listening comprehension on RAPID-RC and only decoding on RISE-RC. Although decoding contributed significantly to performance on all three measures, the proportion of variance explained by decoding varied across measures. Similarly, the total variance in reading comprehension performance also differed across measures.

There are a few possible explanations for the differential findings, which are rooted in method differences between WJ-PC and the computerized tests. Given the multiple-choice format of the RISE-RC and RAPID-RC, it is conceivable that an examinee could correctly answer some questions by relying on informed guesses and without engaging in subprocesses of deep comprehension (Coleman et al., 2010; Katz et al., 1990; Keenan & Betjemann, 2006). Another factor to consider is the modality of the examinee’s response. To respond to an item on RISE-RC and RAPID-RC, the examinee has to click a mouse or press a key on a keyboard. In contrast, the items on WJ-PC require an oral response, which is similar to the response requirements of seven of the eight tests that were used to measure the predictor competencies, including the decoding tests. It is possible that examinees may behave differently and may be more or less likely to use certain competencies on a test that involves interaction with a human examiner as opposed to a test taken silently at a computer. The reading comprehension measures also differed in the length of text stimuli; the WJ-PC featured short passages consisting of one or two sentences, whereas the RISE-RC and the RAPID-RC included longer texts. Word-level processing, as indexed by the decoding measures, may be more important for shorter passages, whereas longer text may require competencies that were not measured in this study, such as working memory.

Despite these differential findings, the models for all three reading comprehension tests emphasize the importance of decoding to struggling adult readers. Decoding was the strongest predictor of success across tests regardless of administration mode and question format. Research with children in elementary grades indicates that much of the variance in reading comprehension at that developmental stage is carried by decoding skills (e.g., Hoover & Gough, 1990; Lonigan et al., 2018). Similarly, decoding influenced reading comprehension performance more than higher-level competencies for our sample of adults with elementary-level skills. Overall, the current adult sample demonstrated skill associations that are more reflective of developing young readers than proficient adult readers.

Differential effects across proficiency levels

A novel feature of our research design was the evaluation of predictors at low, average, and high levels of the reading comprehension performance of struggling adult readers. It should be noted that these proficiency labels are relative to this particular sample; overall, all of these adults would be classified as having reading difficulties. Although quantile regression has been utilized in service of this broad question in a handful of prior studies with children (e.g., Cho et al., 2018; Hua & Keenan, 2017; Language and Reading Research Consortium [LARRC] & Logan, 2017), similar work in the adult literacy context is very limited (Tighe & Schatschneider, 2016).

There are multiple interesting findings regarding the stability of predictors across different levels of reading comprehension proficiency. First, the effect of oral vocabulary decreased between average and high levels of WJ-PC performance. This declining importance of oral vocabulary echoes a trend observed by Ahmed et al. (2016) in their investigation of the DIME framework with adolescent readers in seventh through twelfth grades. As grade level increased in their sample, oral vocabulary exhibited a gradually smaller effect on reading comprehension. This is not surprising, because more proficient readers are adept at activating and integrating word meanings in text processing (Perfetti & Stafura, 2014), which would be expected to reduce the influence of word-level semantic representations on comprehension.

Second, the effect of background knowledge increased between low and high levels of WJ-PC performance as well as RISE-RC performance. Poor readers tend to have gaps in the knowledge domains that are generally covered in formal education (Strucker, 2013) and may be less likely to connect their prior knowledge to what they are reading. It is not surprising that the relationship between background knowledge and reading comprehension was observed only at the 0.9 quantile of reading comprehension of both tests. These results add valuable nuance to past work with struggling adult readers reporting that background knowledge makes a unique contribution to reading comprehension scores (Talwar et al., 2018); this effect appears to exist only for relatively stronger comprehenders.

Third, the effect of decoding consistently increased across low, average, and high levels of RAPID-RC performance. This trend was not observed for WJ-PC and RISE-RC, on which decoding emerged as a stable predictor of reading comprehension across proficiency levels. It can be argued that the adaptive algorithm of the RAPID amplifies the influence of word reading skills on comprehension performance, since the starting passage administered to each examinee on this assessment is determined by the examinee’s performance on the other subtests of the RAPID, one of which measures word recognition.

Finally, the effect of listening comprehension increased between low and high levels of RAPID-RC performance. This finding reflects a trend uncovered in cross-sectional research with children and adolescents. As readers’ proficiency level increases, listening comprehension makes larger contributions to reading comprehension variance (Tilstra et al., 2009; Vellutino, Tunmer, Jaccard, & Chen, 2007). Since listening comprehension was a significant predictor of reading comprehension only at the 0.9 quantile, it can be concluded that the adaptive algorithm did not unduly inflate this relationship.

Overall, the quantile regression results imply that oral language skills and higher-order competencies are more important at higher levels of reading comprehension performance. These adults with greater comprehension proficiency are also likely to have advanced decoding skills. For these more proficient readers, it is plausible that decoding is a reliable competency that does not draw greatly on available mental resources, thereby allowing the reader to focus on higher-level processing of text (Kendeou, van den Broek, Helder, & Karlsson 2014; Perfetti & Hart, 2002).

Implications for assessment and instruction

Our findings indicate that different reading comprehension assessments do not appear to target the same underlying construct for struggling adult readers. This pattern has also been observed with other tests, including those that are commonly administered in Adult Basic Education programs (Mellard et al., 2015; Tighe et al., 2017). Although such work is recent in the adult literacy context, these findings have been reported with child samples for at least two decades (Cutting & Scarborough, 2006; Keenan et al., 2008; Nation & Snowling, 1997). The format, modality, and intended purpose of reading comprehension tests should therefore be considered when evaluating adults for educational progress or research purposes.

The findings also provide support for delivering more targeted instruction in adult literacy programs. Instruction in both vocabulary and background knowledge has yielded gains in reading skills for children and adolescents (e.g., Dole, Valencia, Greer, & Wardrop, 1991; Scammacca, Roberts, Vaughn, & Stuebing, 2015), yet the quantile regression results suggest that focusing on both areas may not be appropriate for all adult learners. Vocabulary instruction may be most beneficial for lower-level readers and can be interwoven with lessons on parsing new words and using contextual clues to guess word meanings (Bromley, 2007). Higher-skilled readers may find it more useful to receive instruction in general and academic knowledge, which can improve their comprehension of academic texts (Ozuru, Dempsey, & McNamara, 2009) and better equip them for high school equivalency tests (Strucker, 2013).

Limitations and future research

A caveat commonly noted in research with struggling adult readers is the heterogeneity of this population. It is very likely that our results are only applicable to adults who are native speakers of English and attend classes targeting the third through seventh grade reading levels. Future research should explore the predictors of reading comprehension across different assessments and proficiency levels with adults who are more skilled and who do not speak English as a native language. Some limited evidence suggests that the explanatory effects of certain competencies will be different in such cases (Herman, Cote, Reilly, & Binder, 2013; To, Tighe, & Binder, 2016).

Due to the ad hoc nature of this study, the design was shaped by the measures available from the larger project. The data analyzed here were collected at a single time point. Thus, any inferences about important predictors of reading comprehension are based on correlations among observed variables. Numerous measures were from the WJ III Normative Update battery of subtests, which may have resulted in inflated associations with WJ-PC scores. Additionally, WJ-PC differed from RISE-RC and RAPID-RC in terms of modality (human-administered vs. computer-administered), question and response format (short oral response vs. multiple-choice), and text stimuli length (long vs. short passages). These differences make it difficult to attribute differential results to a specific test characteristic.

Some of the measures used in the study were not normed on adults over the age of 21 years, which can be problematic because some child-normed tests may not function as expected for adult examinees (Nanda et al., 2014). In addition, the sample exhibited low reliability on RISE-RC (0.63), and reliability could not be estimated for RAPID-RC. These reliability issues raise the possibility that one or both of these measures do not present an accurate account of reading comprehension for the sample. Unfortunately, there is a lack of assessments that have been specifically normed on struggling adult readers (Nanda et al., 2014; Tighe, 2019). Until this problem is remedied, research with this population should strive to include multiple measures of key constructs and model latent factors, which would address measurement error to a greater degree.

In the analyses for the current study, we estimated separate OLS regression models for the three reading comprehension assessments. It should be noted that a multivariate regression approach would have been appropriate and statistically sound. We opted for the separate regression approach to allow for comparable quantile regression analyses.

A consideration for future research is the assessment of different types of background knowledge in relation to reading comprehension. The current study followed the measurement approach used in prior research with struggling adult readers, which is to administer assessments of general background knowledge involving broad domains (Strucker & Davidson, 2003; Talwar et al., 2018). In contrast, researchers studying school-age and college readers typically measure specific knowledge that pertains to the content of the texts used to assess reading comprehension (e.g., Ahmed et al., 2016; Ozuru et al., 2009). Recent work with high school students indicates that general and topic-specific background knowledge exhibit differential explanatory effects on reading comprehension performance (McCarthy et al., 2018). Thus, it is important to explore different ways of assessing background knowledge with struggling adult readers and evaluate these knowledge types in terms of their predictive utility to reading comprehension and their suitability as potential targets of instructional interventions.

An important future direction that arises from the current findings involves the computerized testing of reading comprehension with struggling adult readers. Although our results showed that responses to questions about computer experience were not related to performance on RAPID-RC and RISE-RC, it is possible that participants’ self-reports of computer experience were not accurate. This potential relationship between computer familiarity and test performance should be probed with behavioral measures of computer skills, such as the Northstar Digital Literacy Assessments (Minnesota Literacy Council, 2018). Additionally, the field will benefit from a test modality comparison study with struggling adult readers in which the same reading comprehension passages are administered on paper and on a screen (e.g., Mangen, Walgermo, & Brønnick, 2013). Furthermore, it would be valuable to investigate whether adaptive tests like the RAPID function appropriately for this population in terms of item difficulty and progression rules (Greenberg, Pae, Morris, Calhoon, & Nanda, 2009).