Introduction

Writing skill depends on an intricate interweaving of cognitive, linguistic and affective processes. The writer must rapidly retrieve knowledge from long term memory, manage working memory and attention, express ideas cohesively in verbal form, and think clearly about topics (Kellogg & Whiteford, 2009). It has been proposed that proficient writers utilize a “task environment” made up of social and physical factors including awareness of the audience for the writing, and the text written so far. Further, writers draw on processes at the level of “the individual,” including planning, drafting, and revision, and motivation and affect. Affective factors in writing include one’s goals, predispositions and beliefs about writing (Hayes, 1996).

At the postsecondary level, proficient writing relies on strategies intended to ensure an appropriate response to an assigned prompt (Fallahi, 2012; MacArthur, Philippakos, & Ianetta, 2015). The writer is now expected to understand the informational needs of a reader, generate text appropriate to the purpose and convey information accurately and coherently, while at the same time obeying conventions of grammar and spelling. At this point, there is a focus on discipline-specific writing that summarizes, synthesizes, analyzes and responds to information, and offers evidence to support a position (O’Neill, Adler-Kassner, Fleischer, & Hall, 2012).

For many years, two types of writing that have featured prominently in postsecondary education are persuasive writing and written summarization (Bridgeman & Carlson, 1984; Hale et al., 1996; Wolfe, 2011). These two forms of writing present difficulty for low-skilled adults in this setting (Johns, 1985; MacArthur et al., 2015), and understanding predictors of performance on the two tasks would be an important contribution to intervention planning. The purpose of this study is to determine the extent to which general reading and writing ability, self-efficacy and teacher judgments served as predictors of persuasive writing compared to summarization among low-skilled adult students enrolled in postsecondary developmental education (also known as college remedial programs). In particular, predictors were examined for persuasive writing and summaries written based on source text.

Persuasive writing

The goal of persuasive writing is to convince readers of a position. The writer should state and defend an opinion on an issue, and, at an advanced level, acknowledge and rebut an opposing position, and draw a conclusion (De La Paz, Ferretti, Wissinger, Yee, & MacArthur, 2012; Ferretti, Lewis, & Andrews-Weckerly, 2009; Wolfe, 2011). Writers of well-written persuasive essays are aware of the level of knowledge of their audience and can anticipate readers’ views (Ferretti et al., 2009). Discourse knowledge important in persuasive writing includes an awareness of customary norms of thought of members of a disciplinary community (Goldman & Bisanz, 2002; Shanahan & Shanahan, 2008; Wineburg, 1991; Yore, Hand, & Prain, 2002). Little information is available in previous literature on the persuasive writing abilities of postsecondary remedial students. However, it has been reported that the persuasive essays of students in adult literacy programs may be of low overall quality and display problems with all elements of the persuasive genre, although these weaknesses may decrease after a focused intervention (MacArthur & Lembo, 2009).

Written summarization

Summarization involves arriving at the gist of a message by condensing information to its most important ideas (Perin, Keselman, & Monopoli, 2003; Brown & Day, 1983; Kintsch & van Dijk, 1978; Westby, Culatta, Lawrence, & Hall-Kenyon, 2010). Key processes in summarization are knowing the purpose of a summary, abstracting only the most important information, integrating ideas, and creating a brief but cohesive message expressing condensed information in one’s own words (Westby et al., 2010). Several studies have reported difficulties in text-based summarization among academically underprepared students in postsecondary education, and specifically, with including important ideas from the source text and representing information accurately (Perin et al., 2003; Perin, Bork, Peverly & Mason, 2013; Johns, 1985; Selinger, 1995). For example, Johns (1985) reported that only six percent of a sample of developmental education students included two-thirds of the main ideas in a reading passage when writing a summary and that over half of the information in the summaries was inaccurate. Perin et al. (2003), in a study of written summarization in a similar population, found that approximately 15% of sentences written were extraneous to information in source text, and about one-third of sentences portrayed information inaccurately.

Text-based writing

When the stimulus for writing is written text, reading comprehension skills are required (Fitzgerald & Shanahan, 2000; Mateos, Martin, Villalon, & Luna, 2008). As students enter postsecondary education, they should be able to analyze text to identify the most important information, utilize background knowledge from specific content areas, interpret language and vocabulary appropriately for the context, use strategies to understand new concepts, and draw analogies between different pieces of information (Holschuh & Aultman, 2009; Macaruso & Shankweiler, 2010; Paulson, 2014; Wang, 2009). The ability to comprehend complex expository text is particularly important (Armstrong, Stahl, & Kantner, 2015). Competent reading at the postsecondary level depends on self-regulatory and metacognitive mechanisms involved in setting goals for reading a particular text, applying knowledge of text structure to comprehension, assessing one’s understanding while reading, and evaluating the trustworthiness of text (Bohn-Gettler & Kendeou, 2014; Bråten, Strømsø, & Britt, 2009).

From a cognitive and linguistic perspective, reading and writing are closely related (Fitzgerald & Shanahan, 2000). In higher education, students need to integrate these skills since, at this level, writing assignments tend to be text-based (Carson, Chase, Gibson, & Hargrove, 1992; Jackson, 2009; McAlexander, 2003) and require critical reading of source text (O’Neill et al., 2012; Yancey, 2009). Many students entering postsecondary education have low literacy skills (Sparks & Malkus, 2013) and do not meet college readiness standards on reading and writing (National Governors’ Association and Council of Chief State School Officers, 2010).

Self-efficacy for reading and writing

Self-efficacy, defined as the level of confidence a person has in his/her own ability to perform a challenging task (Bruning, Dempsey, Kauffman, McKim, & Zumbrunn, 2013), is an important construct in postsecondary education. This variable is associated with academic achievement and perseverance, including a tendency to increase effort or attempt new strategies in the face of academic difficulty (Conley & French, 2014).

Self-efficacy appears to mediate learning among college students (Kitsantas & Zimmerman, 2009), and systematic relationships between this construct and both reading and writing achievement have been reported (MacArthur, Philippakos, & Graham, 2016; Pajares & Valiante, 2006). College developmental (remedial) reading students report lower levels of self-efficacy than students not placed in developmental reading courses (Cantrell et al., 2013). Further, low-skilled postsecondary students may increase their self-efficacy for reading and writing tasks as their skills improve (Caverly, Nicholson, & Radcliffe, 2004; MacArthur et al., 2015).

Teacher judgments of students’ reading and writing skills

Teachers have been found to be reliable judges of their students’ reading ability (Ritchey, Silverman, Schatschneider, & Speece, 2015). An early review of studies of the relation between teacher ratings and student test scores reported correlations from r = .28 to r = .86, with a median correlation of r = .62 (Hoge & Coladarci, 1989). A more recent review corroborated this finding, with a mean effect size of .63 (Südkamp, Kaiser, & Möller, 2012). Teacher judgments of students’ general writing ability have been found to be moderate predictors of students’ motivational beliefs about writing, including self-efficacy (Troia, Harbaugh, Shankland, Wolbers, & Lawrence, 2013). However, it has also been reported that teacher ratings of reading skill are more reliable predictors for higher than lower achieving students achieving students (Begeny, Krouse, Brown, & Mann, 2011; Feinberg & Shapiro, 2009), and that teachers may overestimate their students’ ability, especially for students of average reading ability (Martin & Shapiro, 2011). One question that has rarely been addressed in the literature is the relation between teacher judgments and students’ self-efficacy. Especially where low-skilled adults are concerned, it is not known whether students and/or their instructors are able to predict performance levels accurately, and the extent to which instructors’ judgments match their students’ estimates of their own performance.

Current study

Previous studies of persuasive writing and summarization have focused on typically-performing postsecondary students, and students of a range of ability in elementary and secondary education. To our knowledge, no study has reported on predictors of the two types of writing among low-skilled adults in higher education. Seeking to address this gap in the literature, the current study examined predictors of persuasive writing and summarization performance by developmental education students in two colleges in the U.S. The study asked, what are the contributions of general reading and writing ability, self-efficacy ratings, and teacher judgments to measures of text-based summarization and persuasive writing, controlling for college attended? College was entered as a control variable because initial analyses showed differences in reading scores as a function of the participants’ college.

Method

Participants and setting

The participants were N = 211 students attending community college developmental education courses in which reading and writing instruction were integrated. The purpose of developmental education is to provide instruction in basic skills to prepare low-skilled students for the academic demands of the postsecondary curriculum (Boylan, 2002). Entrants to community colleges (also known as two-year and junior colleges) are assessed for reading, writing and mathematics skills and referred to developmental education if test scores are below a predetermined cut point (Bailey, Jeong, & Cho, 2010). The courses attended by participants were one level or two levels below freshman English composition courses in two community colleges in a southern US state (N = 123 in College 1, N = 88 in College 2). The initial sample was N = 212 but one participant was dropped from the analysis because the written summary task was not completed. Fifty-four percent of the participants were Black or African-American, 64% were female, and 75% spoke English as a native language. All were fluent in oral English. Mean age was 24.55 years (SD = 10.66), with 71% aged 18–24 years. All participants had completed secondary education; 83% had a regular high school diploma and the others had a General Education Development (GED) equivalency diploma. Thirty-one percent of participants were responsible for children in the home, and 59% were employed.

The students’ developmental courses centered on three state-mandated goals: demonstrate the use of reading and writing processes; apply critical thinking strategies in reading and writing; and recognize and compose well-developed, coherent, and unified texts. Competencies prescribed by the state relevant to the current study were to compose a text with a consistent point of view, and summarize complex texts. These goals and competencies were developed in a statewide educational reform effort that had occurred prior to the current study and pertained only to postsecondary developmental education.

Classroom observations and interviews with instructors indicated that students did in fact write summaries and persuasive essays during the course. The courses lasted eight weeks and at the time of data collection, the participants were nearing the end of the course. Instructors reported anecdotally that almost all of the study participants were expected to pass the course, meaning that they were considered to be almost ready for postsecondary-level reading and writing.

Measures

Text-based writing tasks necessarily involve both reading comprehension and writing skills (Fitzgerald & Shanahan, 2000). Further, text-based persuasive writing and summarization may trigger different reading processes (Gil, Bråten, Vidal-Abarca, & Strømsø, 2010a, b). In addition, self-efficacy has been found to be related to writing skills among low-skilled adults in postsecondary education (MacArthur et al., 2015). For these reasons, general reading and writing ability, and self-efficacy were selected as basic measures of skill and affect in the current study, which we see as an initial foray into this area. A fourth predictor, teacher judgments, was selected for the purpose of eventual intervention planning, as previous work with children has found this variable to be systematically related to classroom literacy performance (e.g., Südkamp et al., 2012). As a practical matter, we were also interested in the extent to which self-efficacy ratings matched teacher judgments in the target population.

General reading and writing skill

The Nelson-Denny Reading Test, Comprehension subtest (Brown, Fishco, & Hanna, 1993) and the Woodcock–Johnson III (WJ-III, Woodcock, McGrew, & Mather et al., 2001) Writing Fluency subtest, were used as measures of general reading and writing skills. Raw scores, obtained according to instructions in the test manuals, were used in the analysis.

The Nelson-Denny Comprehension subtest consists of 38 multiple-choice factual and inferential questions based on seven unrelated reading passages. The test is normed for Grades 9 through 16. The publisher reports Kuder–Richardson Formula 20 reliability coefficients of .85–.91 for the Comprehension subtest but does not provide information on validity. However, the measure is thought to have reasonable face validity for screening general reading skills (Smith, 1998).

The WJ-III Writing Fluency subtest assesses the ability to write grammatically correct sentences using target words, in speeded conditions. The WJ-III battery is normed for ages 2 through 90. The median score reliability, using a Rasch procedure appropriate for speeded measures, is .88 (reliabilities of .80 and above interpreted as desirable, Schrank, McGrew, & Woodcock, 2001). Further, previous research with low skilled adults found a reliable relationship between the measure and the quality of persuasive essays (MacArthur et al., 2015).

Text-based writing

Two researcher-designed, 30-min tasks were administered to measure text-based writing. Intact text from newspaper articles was used in the interest of ecological validity (Neisser, 1976). To select source text, articles from several newspapers were scanned for readability in combination with appropriateness of topic, and length. We used the readability level of typical introductory college textbooks as the standard for our search for text for the study. These textbooks tend to be written at the 12th grade readability level, although developmental reading texts, as used in the courses in which the current participants were enrolled, are generally written at a lower level (Armstrong et al., 2015). The introductory college level was considered optimal as participants were considered by their instructors to be almost ready for college reading.

Topics were selected to correspond to those taught in high-enrollment, introductory college-level general education courses in which reading and writing assignments were of importance in both learning and assessment. The highest enrollments in the participating colleges in courses with such literacy demands, which students had not yet taken, were in psychology and sociology. Interviews with content-area faculty indicated that newspaper articles were assigned regularly to supplement readings in course textbooks. Further, it was important that the articles were readable by participants within the time allowed for the task. Instructors consulted during the design of the study provided guidance on text length.

Articles meeting the readability and topic criteria were found in the newspaper USA Today, and selected as source text. The articles selected were on the psychology topic of stress experienced by teenagers and the sociology topic of intergenerational tensions in the workplace. The psychology topic was used for the persuasive essay, and the sociology topic was used for the written summarization task. Because instructors requested that assessment time be kept short, it was necessary for the research staff to reduce word length by eliminating several paragraphs in each article. The deleted paragraphs presented examples to illustrate main points in the articles and did not add new meaning. Two teachers with experience in reading instruction read the reduced-length articles and verified that neither cohesion nor basic meaning had been lost as a result of the reduction.

Word count was 650 for the psychology text and 676 for the sociology text. Flesch-Kincaid grade levels were 10–11, and Lexile levels were 1250–1340, corresponding to an approximately 12th grade level. High school students who can comprehend text at a 1300 Lexile level are considered ready for college reading (National Governors’ Association and Council of Chief State School Officers, 2010, Appendix A), and the syllabus for the top level developmental course indicated that texts were within a Lexile range of 1185–1385. Retrospective reports on the summary task (Perin, Raufman & Kalamkarian 2015), which included comprehension questions, indicated that students had no difficulty understanding the article used for that task. Comprehension of the other article, used for the persuasive essay, was not assessed. However, developmental course instructors stated that both articles were within students’ reading ability.

Persuasive essay

Students were asked to read the article on stress experienced by teenagers and write an essay expressing their opinion on a controversy discussed in the article. The prompt instructed students to pretend that they were trying to persuade a friend to agree with them, and use their own words, with no quotations. It was also stated that students could mark the article if they wished, and that a dictionary could be used during the task.

Summary

The text-based summarization task required students to read a second newspaper article, on intergenerational conflict, and summarize it in one or two paragraphs using their own words, again with no quotations. As with the essay, students were permitted to use a dictionary and mark the source text as they wished. The task had another part, in which students stated how information in the article could be applied in a job setting. However, only the summary portion of the response was analyzed for the current study.

Self-efficacy

On a scale designed for the study based on the work of MacArthur, Philippakos, and Ianetta (2015), participants were asked to respond to 16 questions that asked them to rate the level of confidence they felt in their ability on the text-based tasks and related skills (items listed in “Appendix” section). The ratings were obtained prior to the administration of the summarization and persuasive writing tasks so that they could be used predictively in the analysis. The instructions informed respondents that they would shortly be asked to read two newspaper articles and write a summary and persuasive essay based on them. They were directed to rate their confidence on these tasks by selecting, for each of the 16 questions, one point on a 100-point scale reflecting their level of confidence. Participants had to circle one of 11 points on the scale (0, 10, 20, 30, 40, 50, 60, 70, 80, 90 and 100). Examples of points on the scale were provided in the instructions: 0 (“you are sure you cannot do it”), 50 (“there is an equal chance that you can do it or not do it”), and 100 (“you are sure you can do it”). The same 16 questions were used for teacher judgments, as described below, although the wording of the student items was simplified to ensure comprehension.

Teacher judgments

A teacher judgment scale was developed for the study based on work suggesting that teacher judgments are predictive of student performance (Hoge & Coladarci, 1989; Speece et al., 2010; Südkamp et al., 2012; Troia et al., 2013). Using the online Qualtrics platform, instructors were asked to provide ratings on the same items for which students had rated their self-efficacy, using the same 11 points on a 100-point scale. Thus, each teacher judgment was keyed to a student self-efficacy item. As with the self-efficacy scale, the teachers were asked to rate their level of confidence that the student was able to perform the task indicated. A rating of 100% indicated that the teacher was sure that the student had the skill, 50% meant that the teacher thought that there was an equal chance that the student did or did not have the skill, and 0% meant that the teacher thought that the student did not have the skill.

Scoring of text-based writing

The persuasive essays and summaries, which had been handwritten by participants, were word processed, correcting for spelling, capitalization, and punctuation (Graham, 1999; MacArthur & Philippakos, 2010; Olinghouse, 2008). Two genre-specific scores were obtained for each type of writing: holistic persuasive quality and number of persuasive essay parts for the essays, and proportion of main ideas from the source text and analytic quality for the summaries. In addition, the use of academic words, and length of the writing sample, in number of words, was obtained for both types of writing. Thus, there were four scores for each type of writing.

Holistic quality of the persuasive essay was scored using a 7-point rubric (based on MacArthur et al., 2015). Raters were asked to bear in mind the clarity of expression of ideas, the organization of the essay, the choice of words, the flow of language and variety of sentences written, and the use of grammar. For the essay parts measure (adapted from Ferretti, MacArthur, & Dowdy, 2000), each essay was first parsed to idea units (Gil et al., 2010b; Magliano, Trabasso, & Graesser, 1999) and then labeled as a functional or non-functional element. Following Ferretti et al. (2000), functional elements consisted of a proposition, or statement of belief or opinion; a reason for the position stated; an elaboration of proposition or reason; an alternative proposition, or counterargument; a reason for the alternative proposition; a rebuttal of the counterargument; or a concluding statement. Nonfunctional persuasive elements were defined as repetitions or information not relevant to the prompt. Almost all of the functional persuasive units in the current corpus were propositions, reasons, elaborations of propositions or reasons, or conclusions, with very few counterarguments or rebuttals. The functional and non-functional elements were summed and functional elements were expressed as a proportion of total elements for a proportion elements score.

On the summaries, the proportion of main ideas from the source text was calculated (Perin et al., 2003; Perin et al., 2013). To identify the main ideas, the first author and two research assistants who were experienced English language arts teachers each independently read the article and listed the main ideas. Then, through discussion consensus was reached on which ideas to retain, eliminate, or add to the list. Through this process, nine main ideas were identified in the source text. For the analysis, the number of main ideas found in each summary was expressed as a proportion of the nine main ideas in the source text.

The quality of the summary was assessed using an analytic summarization quality rubric with four 4-point ratings, for a maximum score of 16 (based on Westby et al., 2010). Ratings were made for four aspects of summary quality: (1) topic/key sentence, main idea; (b) text structure; (c) gist, content (quantity, accuracy, and relevance); and (d) sentence structure, grammar, and vocabulary.

Both the essays and summaries were scored for the number of academic words used (Lesaux, Kieffer, Kelley, & Harris, 2014; Olinghouse & Wilson, 2013). Academic words are defined as those that appear frequently in academic texts but are not specific to any specific subject area (Lesaux et al., 2014) (Examples: circumstances, category, debate, demonstrate, estimate, interpret, guarantee). The percentage of academic words was obtained from the automated vocabulary profiler software program VocabProfile (Cobb, n.d.), which computes the percentage of words that occur on the Academic Word List (Coxhead, 2000) and contains words not among the 2000 most frequent words in the English language. The academic word measure reflects the assumption that less mature writing contains a predominance of highly frequent words and fewer low-frequency words (McNamara, Crossley, & McCarthy, 2010; Uccelli, Dobbs, & Scott, 2013).

Finally, the number of words written on the persuasive essays and summaries was obtained using the word count feature in the Word program. This measure was taken because low-achieving students often produce inappropriately short compositions (Doolan, 2014; Nelson & Van Meter, 2007; Puranik, Lombardino, & Altmann, 2008).

All of the measures were scored by three trained research assistants who were unfamiliar with the goals of the project. Each assistant scored one of the three full sets of writing samples (WJ-III Writing Fluency, essay, summary). A random sample of 30% of each set was scored by another assistant to determine the interrater reliabilities of scores on the WJ-III Writing Fluency, essay functional elements and quality, and the summary proportion of main ideas and analytic quality. Anchor papers for the essays and summaries obtained in piloting were used, and pairs of assistants were trained by scoring sets of 10 protocols until they reached a criterion of 80% exact agreement. It is noted that training was a lengthy process because the writing was poor and sometimes difficult to understand.

Interrater reliabilities (Pearson’s r) for the measures were adequate: WJ-III Writing Fluency: .98; essay functional elements: .92; essay quality: .85; summary proportion of main ideas: .72; summary quality: .85. Exact agreement for persuasive essay quality was adequate at 77%. However, exact agreement on the other variables was moderate or low: 40% for writing fluency, 43% for essay functional elements, 14% for summary main ideas, and 23% for summary quality. Agreement within one point improved to 79% for writing fluency, 68% for essay functional elements, 100% for essay quality; 63% for summary main ideas and 44% for summary quality. It can be seen that the inter-scorer agreement of summary quality measure was problematic; however, agreement within two points on this 16-point measure improved to 71%.

Procedure

Participants were recruited by college liaisons through classroom presentations and flyers. The measures were administered by the third and fourth authors, following a script, in 2.5 h group sessions, outside of class time, with several rest breaks. Order of administration was fixed, as follows: self-efficacy, persuasive essay, WJ-III Writing Fluency, summary, Nelson-Denny Comprehension, and a questionnaire on demographic and academic information. The teacher judgments were submitted on the Qualtrics platform. Along with the link to the scale, instructors received copies of the reading and writing tasks on which their judgments focused, and were given monetary compensation for completion of the task. Participants received gift cards. All data in the study were collected in accordance with the standards and guidelines of the Institutional Review Board of Teachers College, Columbia University.

Results

Descriptive statistics

Raw and standard scores on the general reading and writing measures (Nelson-Denny Comprehension and WJ-III Writing Fluency tests) are shown in Table 1 (Only raw scores were used in the correlations and regression analyses reported below). The scores verify that participants were low-skilled, and translate to the 22nd and 27th percentiles for reading and writing, respectively, using as a reference group 12th graders at the end of the school year. There were no statistically significant differences between developmental course levels on either test at either college. However, there was a statistically significant between-college difference on the comprehension scores, with higher scores at College 1 (t = 2.5, df = 210, p < .01) although there was no systematic difference in writing fluency (t = .33, df = 209, p = .744).

Table 1 General reading and writing measures

Mean (unadjusted) writing scores, and the self-efficacy ratings and teacher judgments are shown in Table 2. Intercorrelations between general reading and writing, and text-based writing scores are shown in Table 3, and between self-efficacy ratings, teacher judgments and text-based writing scores in Table 4. Pearson product-moment correlations (Pearson’s r) were performed. Data are provided for all N = 211 participants on the reading and writing variables. Instructors provided teacher judgments for N = 162 of the participants, i.e., data were missing on this variable for N = 49 participants, and only these individuals were included in the analysis of the self-efficacy and teacher judgement data. As mentioned earlier, the teacher judgments were provided by instructors on an online platform; follow-up with instructors did not result in successful provision of the missing information.

Table 2 Writing scores*, Self-efficacy ratings** and Teacher judgments**
Table 3 Intercorrelations: general reading and writing, and text-based writing scores
Table 4 Intercorrelations: self-efficacy ratings, teacher judgments, and text-based writing scores

Table 3 shows that the correlation between reading comprehension and writing ability was statistically significant. Both of these variables correlated significantly with proportion of functional elements, holistic persuasive quality, length of the persuasive essay, and summary analytic quality. Only reading comprehension correlated significantly with proportion of main ideas in the summary and percentage of academic words in the summary. Neither variable correlated significantly with percent academic words in the persuasive essay or length of the summary.

In Table 4 it can be seen that the correlation between the self-efficacy ratings and teacher judgments was statistically significant. Both of these measures were significantly correlated with the proportion of functional persuasive elements, and summary analytic quality. Only the teacher judgments ratings correlated significantly with holistic persuasive quality, proportion of main ideas in the summary, length of the summary, and proportion of academic words in the summary. Thus, overall, the teacher judgments appeared more predictive of student performance than the self-efficacy ratings.

Contribution of general reading and writing ability, self-efficacy, and teacher judgments to text-based writing

A series of hierarchical regressions was completed in order to determine the contribution of general reading and writing ability (as measured by scores on the Nelson-Denny and WJ-III Writing Fluency tests), self-efficacy ratings, and teacher judgments to scores on the persuasive essays and written summaries. We modeled six variables: the proportion of functional persuasive elements in the essay, essay quality, the percentage of academic words in the essay, the proportion of main ideas in the summary, summary quality, and the percentage of academic words in the summary.

College was entered in the first block as a control variable, since there were statistically significant differences in general reading and writing ability (Nelson-Denny Comprehension and WJ-III Writing Fluency, which we refer to as standardized tests below) between the two sites. The scores on the standardized tests were entered in the second block, and the third block consisted of the self-efficacy ratings and teacher judgments. The results of the third block indicate whether the self-efficacy ratings and teacher judgments were statistically significant predictors of text-based writing skills over and above the contributions of the scores on the standardized tests. The results of the analyses are shown in Tables 5, 6, 7, 8, 9 and 10. Total R 2, or the total amount of variance accounted for, ranged from 9% (percentage of academic words in the summary) to 18% (percentage of academic words in the essay and summary quality).

Table 5 Proportion of functional elements in the essay: summary of hierarchical regression analysis
Table 6 Essay quality: summary of hierarchical regression analysis
Table 7 Percentage of academic words in the essay: summary of hierarchical regression analysis
Table 8 Proportion of main ideas in the summary: summary of hierarchical regression analysis
Table 9 Summary quality: summary of hierarchical regression analysis
Table 10 Percentage of academic words in the summary: summary of hierarchical regression analysis

Proportion of functional persuasive elements in the essay

As seen in Table 5, in the full model for the proportion of functional elements in the essay, Model 3, the change in R 2 was statistically significant (F = 4.580, p = .012), which can be explained by the contributions of the WJ-III Writing Fluency scores and the self-efficacy ratings. The change in R 2 in Model 2 (scores on standardized tests only), was also significant (F = 7.106, p = .001), which can be attributed to the contribution of the scores on the standardized writing test.

Essay quality

The change in R 2 in the full model for essay quality was not statistically significant, as seen in Table 6, although, again, the WJ-III Writing Fluency scores made a significant contribution. However, the change in R 2 in Model 2 (scores on the standardized tests only) was significant (F = 4.960, p = .008), which was a result of the contribution of the standardized writing scores.

Percentage of academic words in the essay

Neither Model 2 nor Model 3 was statistically significant in explaining the percentage of academic words used in the essay, as shown in Table 7, suggesting that performance could not be predicted by general reading or writing ability, self-efficacy, or teacher judgments.

Proportion of main ideas from the source text in the summary

As seen in Table 8, the change in R 2 for the proportion of main ideas in the summary was statistically significant in the full model (F = 7.659, p = .001). This result was explained by the significant contributions of the Nelson-Denny Comprehension scores and the teacher judgments. Model 2 was also significant (F = 5.190, p = .007) because of a significant contribution of the scores on the standardized reading tests.

Summary quality

The change in R 2 was statistically significant in the full model for summary quality (F = 8.725, p = .000) as well as Model 2 (F = 5.178, p = .007). In the full model, as seen in Table 9, only the teacher judgments made a significant contribution to summary quality, and in Model 2, only the scores on the standardized reading tests made a significant contribution.

Percentage of academic words in the summary

As shown in Table 10, the full model of the percentage of academic words in the summary was statistically significant (F = 3.901, p = .022) because of the significant contribution of teacher judgments.

Discussion

Description of skills

The present study set out to contribute information about the writing of an often-overlooked population in the research literature, low-skilled adults who have completed secondary education and are in higher education. The participants tested at the lower end of the average range for end-of-year 12th graders on the general reading and writing measures. The scores can be compared with those found in previous research on developmental education students (MacArthur et al., 2016; Perin et al., 2003; Perin et al., 2013). In the (Perin et al., 2015) study, top level developmental education students obtained a mean raw score of 30.95 (SD = 15.22), or 41% correct, on the Nelson-Denny Comprehension subtest, which is similar to the score of 29.32 (SD = 12.29), or 38% correct, found in the current sample of intermediate and top level developmental students. MacArthur et al. (2015) reported a mean raw score of 19.9 (SD = 4.5), or 50% correct, for upper level developmental students on WJ-III Writing Fluency, which is somewhat similar to the mean raw score of 22.50 (SD = 5.25), or 56% correct, in the present sample.

Performance on the text-based writing tasks provides insight into participants’ readiness for college-level literacy demands because they required both reading comprehension and writing skill and were typical of classroom assignments. When summarizing a newspaper article, the participants included 24% of the main ideas from the source text. Although it is not expected that all of the main ideas would necessarily be included in a summary written even by the most proficient writer, 24% of the main ideas is low if the summary is to capture the gist of the source text. Further, although norms are not available for this task, performance for this sample fell below that of two other developmental education samples, 28% in Perin et al. (2003) and 42% in Perin et al. (2013). The quality of the written summaries, measured by an analytic rubric that focused on four components of summarization, also tended to be somewhat low, with a mean score of 8.09 (SD = 2.64) on a 16-point scale.

Participants also demonstrated weakness in text-based persuasive essay writing, with a mean quality score of 2.58 (SD = .80) on a 7-point holistic scale. This is similar to the scores of 2.44-2.96 reported by MacArthur et al. (2015) on experiential (i.e., not text based) persuasive writing of top-level developmental students. Almost one half of the content written in the persuasive essays in the current study, measured in terms of functional persuasive elements, was not pertinent to the development of a persuasive argument. Comparable data on the use of persuasive elements are not available in the literature, although Ferretti et al. (2009; Table 2) reported that 50.8% of learning disabled and 25% of typically developing sixth graders included non-functional elements in their persuasive essays. Further, although there are no prior studies of the use of academic words in the writing of adults (or adolescents), a study of children found that one percent of the words used in the writing of typically-developing fifth graders were academic words (Olinghouse & Wilson, 2013). In this light, the three percent seen in the current postsecondary sample seems low.

Relationships between measures

Scores on the two text-based writing tasks had a statistically significant but relatively weak relationship to each other (r = .16, p < .05 for summary and essay quality), suggesting that the two tasks called for different skills. This possibility is supported by the different relationships of the scores on the standardized reading and writing tests to the two tasks. The reading scores had a higher correlation with summary quality (r = .31, p < .01) than with essay quality (r = .16, p < .05), and the writing scores had a higher correlation with essay quality (r = .28, p < .01) than with summary quality (r = .14, p < .05). Although both tasks required both reading and writing, it appears that reading skills were more important for the written summarization task and writing skills were more important for the text-based persuasive essay.

Another notable finding was that, although the scores on the standardized reading and writing tests were reliably related to performance on the text-based writing measures, even the highest correlations were moderate, suggesting that they were tapping different skill sets. Of the 45 correlations we ran between the scores on reading and writing skills, only five exceeded r = .40: essay quality and essay word count (r = .51, p < .01), essay quality and the percentage of academic words in the essay (r = .57, p < .01), proportion of main ideas in the summary and summary quality (r = .61, p < .01), proportion of main ideas in the summary and summary word count (r = .47, p < .01), and summary quality and summary word count (r = .50, p < .01). Of these five relatively strong correlations, three concerned word count. Although the direction of the relationship between each pair of variables is unknown, a hypothesis can be proposed that working with students to lengthen their writing samples may help them improve their writing.

The self-efficacy ratings were significantly correlated with teacher judgments, but the correlation was only moderate (r = .30, p < .01), suggesting that for individual students, there were discrepancies between students’ and teachers’ level of confidence in the student’s proficiency.

Self-efficacy ratings correlated significantly with only two of the eight text-based writing variables, whereas teacher judgments correlated significantly with six of these variables. Therefore, the current data suggest that the teachers may be better than their students at predicting the students’ reading and writing skills. Although previous research has found self-efficacy to be a significant predictor of literacy skills (MacArthur et al., 2016; Pajares & Valiante, 2006; Proctor, Daley, Louick, Leidera, & Gardner, 2014), the current study has the added advantage of being able to compare self-efficacy ratings with teacher judgments for the same sample.

Contribution of general reading and writing, self-efficacy ratings and teacher judgments to variance in text-based writing

The results of hierarchical regression analyses indicated the importance of the general writing scores in predicting the proportion of functional elements in the persuasive essay and the quality of the essay, while the general reading scores were important in predicting the proportion of main ideas in the summary and the quality of the summary. Although directionality is not established in the analysis, these findings suggest a hypothesis that improvement in text-based summarization may require particular attention to reading comprehension skills, while improvement in text-based persuasive essay writing may depend more on developing general writing skills.

Self-efficacy was only important in predicting the proportion of functional elements in the essay, and teacher judgments were only important in predicting the percentage of academic words in the summary. Thus, self-efficacy and teacher judgments had only a small explanatory role in text-based writing once the specific college attended and the standardized test scores were taken into account. Previous research suggests that self-efficacy is a reliable predictor of literacy performance (e.g., MacArthur et al., 2016; Martinez, Kock, & Cass, 2011) and that teacher judgments correlate with literacy skills, although relationships are stronger at higher skill levels (Begeny et al., 2011; Feinberg & Shapiro, 2009). Perhaps the lack of high achievers in the current sample, in conjunction with the ceiling effect on self-efficacy (with many scores at the highest part of the scale), at least partly explains the relatively small amount of variance of self-efficacy and teacher judgments, compared with standardized test scores, in accounting for the text-based writing scores.

Limitations and research directions

We see two limitations in this study. First, interscorer agreement on some of the measures was low despite extensive training of raters, which we attribute to the low-quality and high variability of the writing samples analyzed. Additional research is needed to determine whether more refined scoring techniques can be developed for this population to ensure higher reliability. Second, interpretation of the current data is limited by the lack of prior literature that could be used for comparison. Future research is needed not only on text-based writing itself, but this particular skill among students with low writing skills.

Implications

The present findings help illuminate the challenges to supporting struggling writers at the postsecondary level and suggest some possible pathways to explore in order to improve literacy outcomes in this population. As discussed in the introduction, postsecondary writing places a multiple demands on the student’s cognitive and self-regulatory skills. However, it appears from the r = .16 correlation between the written summarization and the persuasive essay quality scores that these tasks are very different and that the writers themselves have uneven writing skills. In examining the low correlations between some of the tasks and the low levels of predictiveness, it seems that there are different and unique ways to be poor writer.

We would hypothesize that the low interrater reliability on the scoring of the writing could be related to the complexity of poor writing, making it difficult to evaluate an evidently wide range of approaches to writing among this population. While there clearly needs to be more research with this population, our data suggest there are numerous of “moving parts” when it comes to understanding what makes poor writers poor writers.

However, our research can begin to suggest some of the ways poor postsecondary writers can be supported. First, it is clear that for text-based writing (a common postsecondary task), reading comprehension needs to be supported. Second, the relationship between length of writing and the quality of the writing sample warrants further exploration in order to determine whether supporting writing stamina leads to better writing or if better writing skills themselves serve to increase stamina. Additionally, the complexity of the pattern of skill displayed by the low-achieving writers in the current study constitutes a challenge to educators to develop customized interventions that take into account both the individual writer and the writing context.

Conclusions

The study revealed low levels of college-readiness on an important postsecondary literacy skill for a sample of students many of who were about to exit developmental (remedial) coursework and proceed to college-level courses. Further, the hierarchical regressions suggest that improvement in general reading skills may promote better text-based summarization while improvement in general writing skills may be key to better persuasive essay-writing skills.

The text-based writing measures, as examples of authentic college level assignments, may have good potential for adding to currently-used measures of college readiness such as college placement tests. Problems have been identified in the use of single measures for placing students into developmental education (Hughes & Scott-Clayton, 2011), and it is possible that the use of multiple measures might lead to more accurate course placement (Scott-Clayton, 2012). Finally, given the lack of prior data against which to compare current findings, this study may be best seen as an exploration of the writing of a large, high-need population.