Introduction

Fairness is an essential quality of the language assessment process (Wallace, 2018; Kunnan, 2000, 2004; 2018; Rasooli et al., 2018). It is defined as treating test takers equally (Kunnan, 2000), regardless of individual characteristics (i.e., age, ethnicity, race, religion, and first language; Joint Committee of Testing Practices, 2005). Typically, this is addressed by ensuring that the procedures through which assessments are administered are the same for every test taker, that the content on the test is neutral and does not bias one group over another, and that there is sufficient information about the test given to test takers before they take it (Kunnan, 2018). To determine if an assessment is fair, researchers traditionally examine its psychometric properties and if there is an absence of bias, then it is deemed a fair assessment. The current study follows a more recent trend in the fairness literature and takes a subjective view of fairness. A test is fair if the stakeholders (e.g., the test takers and test administrators) perceive it to be fair (Wallace, 2018; Wallace & Qin, 2021). This perspective is based on the large body of literature showing that learners who perceive assessment to be unfair will experience negative emotions (e.g., lower motivation) and behaviors (e.g., acting out in class) in the classroom (Chory-Assad, 2002).

Another important aspect of fair assessment practice, and one that has been overlooked, is the assessment approach used to evaluate performance. Assessments can compare performance against a set of pre-determined criteria, against other test takers, or against an earlier performance (Dalbert et al., 2007). Applying an inappropriate form of assessment can be perceived by the test takers to be an unfair evaluation of their performance because there is a mismatch between the intended purpose of the assessment and the approach taken. For example, if scores were awarded for a language performance by comparing the performances among all the students in an English as a foreign language (EFL) course, this could be perceived as an unfair assessment by the test takers because the purpose of a classroom test is to evaluate how well learners meet the standards of a course, and not how well they performed on the test compared with their peers (Brown & Hudson, 2003).

Most research attention into fairness has focused on high-stakes tests (i.e., standardized language tests such as TOFEL and IELTS), but it is also important to take the fairness of low-stakes tests (i.e., performance on classroom assessments) into account (Wallace, 2018). Assessment accounts for nearly half of the language teachers’ professional time (Lan & Fan, 2019), thus highlighting its importance in language teaching and learning. The fairness of assessment is especially important in language classrooms, because not doing so can lead to several undesirable consequences (e.g., complaining to the teacher or acting out in class; Chory-Assad, 2002). To avoid these consequences, it is important to understand what students and teachers perceive to be fair assessment approaches of learning.

Approaches to Assessment

There are three approaches to assessment: criterion-referenced assessment (CRA), norm-referenced assessment (NRA), and individual-referenced assessment (IRA) (Dalbert et al., 2007; Geisinger et al., 1980). CRA refers to evaluating an individual learner’s performance when compared with a set of descriptive standards (Dalbert et al., 2007). These standards are typically pre-determined by the school administration (Neil et al., 1999) and often are based on pre-established criteria determined by an external organization (Nightingale, et al., 1996). For example, the writing performance criteria may be taken directly from the Common European Framework of References for Languages descriptors (Council of Europe, 2001), and writing performance is evaluated based on these descriptors. The goal of CRA in the classroom is to determine how well students meet the learning objectives of the course or program. For example, if a student performs well on a test, then the interpretation of that performance may be that the student has acquired a sufficient amount of the knowledge, skills, or competencies required for success in that program.

In NRA, teachers compare a student’s performance against other students’ performances and then assign a score for that performance in relation to the others (Neil et al., 1999). The goal of NRA is to achieve a normal curve, or the “bell curve” distribution of scores, where a small number of scores are high and low, respectively, and a majority of them are in the middle. The outcome of the scoring procedure is a ranked list of performances, where students with the best X% of performances get higher scores and those with the worst X% get lower scores (Nightingale et al., 1996). Scoring to achieve a normal curve can result in students getting higher scores for weak performance when the overall performance of the class is poor. Similarly, the students may get lower scores if their performance is lower than their classmates, regardless of how well the students performed on the task.

Finally, for IRA, also referred to as ipsative assessment (Hughes et al., 2014), students’ performance is compared to their previous performance (Dalbert et al., 2007). For example, students’ scores of writing assignment may be compared with their previous pieces of writing. This may result in students receiving higher scores for improvement on their own performance even if they are not among the best-performing students in the class.

Assessment Types in the Classroom

The assessment type taken in the classroom can affect the design of the test content (e.g., the skills, abilities, or knowledge measured on the test), the analysis of the tests, as well as the validity of the tests (e.g., whether they measure what they intend to measure; Brown, 1996). The assessment approach also influences the interpretation of the assessment results, and perceptions of performance may differ depending on the approach taken. For example, a student who scored 85% on a test can be interpreted in multiple ways. It could represent that the student knew 85% of the test content (i.e., using CRA), or it could mean that the student’s score was better than 85% of the students and worse than 15% of them (i.e., using NRA). Because interpretations of an assessment score may vary according to the assessment approach, it is important to understand the advantages and disadvantages of the approach that is utilized.

CRA and NRA are two commonly used assessment types that have been utilized in the classroom to evaluate performance. It has been suggested that of the two, CRA is more suitable for the classroom. The purpose of classroom assessment is for teachers to obtain information where students learnt best and where they need improvement (Airasian, 2005), and for teachers to check students’ understanding of the knowledge acquired and assist their learning with information generated in the assessment (Mckay, 2006). The benefit of CRA is that it evaluates whether students have acquired a certain ability, knowledge, or proficiency (Lok et al., 2016; Notar et al., 2008) and provides learners with feedback on their assessment so as to achieve said criteria. When students receive a score on their performance, then they know what that score means in relation to the quality of their performance against a standard. Students are not just given a score and told that they were better than some students and worse than others as they would be with NRA. Therefore, a benefit of CRA is that it is aligned with the purpose of classroom objectives and is recommended in classroom assessment (Brown & Hudson, 2003).

On the other hand, it has been claimed in the literature that using NRA in the classroom would be inappropriate because the purpose of classroom assessment is not to distinguish the better from worse, but to evaluate whether students meet the certain criteria to acquire the knowledge delivered and strive for improvement (Airasian, 2005). Only finding out where students performed in relation to one another is not very informative for learning (Brown & Hudson, 1998). The origins of NRA may help explain its appeal for the classroom. It was originally used to make diagnoses in psycholinguistic research (Kirk et al., 2014), particularly for identifying language impairment (Ebert & Scott, 2014; Hendricks & Adlof, 2017; Norbury & Bishop, 2003; Paul & Norbury, 2012) and distinguishing among the degrees in which a participant’s language is impaired (Spaulding et al., 2012). In order to determine how impaired one’s language was, researchers compared performance on language-oriented tasks against others. Unfortunately, using assessment for this purpose may not be useful or appropriate for the classroom. However, NRA is still frequently used in the classroom today (Dalbert et al., 2007), despite recommendations to the contrary.

So why does NRA still persist in classrooms despite recommendations that it be avoided there? One reason may be the learning context. When learners study in a test-oriented environment, like most EFL contexts in Asia, then classroom teachers may evaluate their students’ learning in a similar manner to prepare them for their high-stakes exams. When Cheng et al. (2004) compared the assessment purposes, methods, and procedures for ESL/EFL teachers in mainland China, Canada, and Hong Kong, they reported that language teachers in mainland China heavily relied on using assessment items from standardized tests, or those designed to discriminate among a population of test takers, for their classroom assessments. The authors reasoned that this was likely due to the learning context involving students needing to successfully perform on a high-stakes language assessment (e.g., TOEFL), so teachers intentionally selected these kinds of assessments to evaluate their students’ learning. However, this tendency may be influenced by the level of assessment literacy of the teachers. The Hong Kong teachers in Cheng et al. (2004) seldom used these types of items and relied on their own test item development despite also being in a test-oriented environment. This result was encouraging, but assessment literacy appears to be limited in many learning contexts. Sevimel-Sahin (2020) observed that studies examining EFL teacher perceptions of language assessment consistently reported that in-service teachers felt unprepared to deliver effective assessment of learning (e.g., Vogt & Tsagari, 2014) and had limited knowledge of assessment (e.g., Semiz & Odabas, 2016). This likely explains why teachers rely on using NRA in the classroom—they lack the knowledge and competence to be able to develop their own tests to measure their learning objectives and/or they do not understand that classroom-based assessments should not be used to discriminate among the students in class. A final reason why NRA is still used today may be attributed to cultural tradition. Sasaki (2008) acknowledged that Japan has traditionally used NRA in the classroom. However, a change towards using CRA methods in the educational curriculum in the 1990s was met with several challenges—a lack of transparency for the criteria and scores not being useful for feedback—resulting in it being less favored than NRA still today.

The final assessment approach, IRA, has received little attention in the literature compared with CRA and NRA. This is unfortunate because it has many benefits to learning. Certainly one benefit is that it shows students’ own individual learning progression. Similarly, it has been found to have the strongest relationship with students’ intrinsic motivation (Gipps, 2011). For example, students will be more willing to help each other with their learning and less likely to compete with each other. Likewise, IRA has the potential in motivating learning since it is not related to external criteria and standards, but associated with internal and personal performance (Hughes, 2011). The other two assessment types (NRA and CRA) do not provide this kind of individualized assessment. However, one drawback is that it may not be possible to ensure equal measurement across all learners, potentially causing it to be perceived as unfair.

The current study’s context is set in a Macau, where an exam-oriented educational system similar to mainland China and Hong Kong is in place. It is likely that a range of assessment approaches are utilized in Macau schools, but it is unclear how fairly students or teachers perceive the classroom assessment approaches to be.

Fairness in Assessment

Fairness is a test quality that refers to perceptions of stakeholders towards the assessment (Wallace, 2018; Wallace & Qin, 2021). Language tests should be fair, especially for the classroom-based tests since most language assessment is done in the classroom (Wallace, 2018). If the type of assessment is used inappropriately in the classroom, it can be perceived as unfair. Bempechat et al. (2013) discovered this when discussing classroom assessment experiences with Russian grade nine students, who reported being more wary of assessment when the scoring approach was unclear and the evaluators applied an idiosyncratic scoring system to their work. This suggests that learners may be more sensitive to assessment practices when they perceive such practices as questionably fair. This also shows that it is important to make students feel that they are treated fairly because once students perceive they have received unfair treatment, this perception is likely to spread. Students perceiving tests as fair would have positive influences, like being more engaged in learning (Wentzel, 2009) and having higher-level of satisfaction towards instructors, assessment procedures and their scores (Chory-Assad, 2002; Wendorf & Alexander, 2005). In contrast, they act out against their teacher when students perceive tests as unfair (Chory et al., 2017; Chory-Assad & Paulsel, 2004; Lemons et al., 2011). One source of unfair perceptions may be the type of assessment used in the class. If it is inappropriately used, learners may consider it to be unfair (Wallace, 2018) and cause unwanted reactions, both emotionally and behaviorally (Chory et al., 2017; Chory-Assad & Paulsel, 2004; Wendorf & Alexander, 2005), such as giving out verbal aggression, arousing hostility and expressing resistance (Chory et al., 2017; Chory-Assad & Paulsel, 2004), cheating (Lemons et al., 2011), dissenting to others (e.g., complaining to friends), and not engaging the class (Chory et al., 2017). As students’ perceptions of fairness are associated with learning outcomes and behaviors (Wentzel, 2009), it is important for teachers and students to perceive their classroom assessment to be fair.

Little research has examined the fairness perceptions of each assessment type for classroom-based tests across curriculum. One study that has done so, Dalbert et al. (2007), reported that German adolescents considered CRA to be the fairest assessment approach for English, German, and mathematics classroom evaluation. They also reported that IRA was fair, but that NRA was unfair. These findings indicate that the learners felt that being evaluated against a standard was fair assessment practice, as was evaluation against a prior performance. However, comparing performance against others was not viewed as being fair. It is unclear if these perceptions of fairness would be similar in the Asian EFL classroom, especially when the learners are immersed in a test-oriented culture. Therefore, this study aims to investigate local Macau secondary school students’ perceptions of fairness on three assessment approaches in the language classroom.

While the empirical research has examined the perceptions of assessment fairness on behalf of the students, less has investigated the teachers’ perspectives. It would be expected for teachers’ views on assessment fairness to mirror what is recommended in the literature for classroom-based assessments—that CRA and IRA are fair, but NRA is unfair–but this is yet unknown. To address this need, the following research questions were posed.

  1. 1.

    How fairly do Macau secondary school students and teachers perceive CRA, NRA, and IRA approaches to be?

  2. 2.

    Are there significant differences among the perceptions of the three assessment approaches for the students and teachers?

Methodology

Participants

In total, 344 students and 14 teachers (3 male, 11 female) from a Chinese-medium school (the language of instruction is all in Chinese) in Macau were invited to participate in the study. The students, aged 12 to 16 (M = 13.9, SD = 1.3), were from 13 classes. Ultimately, 276 students (115 female; 161 male) and all 14 teachers elected to participate. Because the students were learning in a Chinese-medium school, their exposure to English was limited to their English language classes. Their English courses focused on increasing vocabulary and grammatical knowledge and improving their reading, writing, and speaking skills. The proficiency level of the students was anticipated to range between Common European Framework of References for Languages A1-A2.

Out of the 14 teacher respondents, 8 taught at the junior high school level (students from age 12–14) and 6 were from the senior high school level (students from age 14–16). The number of years’ experience in teaching English ranged from 1.5 years to 26 years, with the average being 9.6 years of experience and the median 8.5 years of experience. All of the teachers delivered a blended syllabus, focusing on improving reading, writing, and speaking skills and increasing vocabulary and grammatical knowledge.

Instrument

The questionnaire used in this study (see Appendix) was adapted from Dalbert et al. (2007) for the EFL context. The original questionnaire elicited fairness perceptions of three assessment types (CRA, NRA, IRA) across three school subjects—German, English, and math. The survey used in the current study focused on one subject, English, but elicited fairness perceptions of the assessment approaches across the language skills and content taught in Macau secondary schools—reading, writing, speaking, vocabulary, and grammar. The questionnaire contained five scenarios for each skill. For each scenario, participants were presented with three scoring options to choose from, each representing one of the assessment types: CRA, NRA, and IRA. Using a six-point Likert scale, participants indicated their perception of fairness of each scoring option from 1 (“totally unfair”) to 6 (“totally fair”). The questionnaire was presented in Chinese to ensure participants clearly understood the scenarios and optional scoring procedures.

An example scenario and scoring options from the questionnaire are provided below. The scenario sets the assessment context and describes the performance of a hypothetical student. The scenarios are contexts that the participants may experience at school. Accompanying them are three scoring options (a–c) that are intended to represent the three scoring approaches. In the example below, a., b., and c. elicit fairness perceptions of NRA, IRA, and CRA, respectively.

figure a

Procedures

After receiving ethical clearance from the university and informed consent from the participating school, the participants, and their parents, the questionnaires were distributed in their electronic form during class. Students who elected not to participate in the study were given an alternative assignment. Students who completed the questionnaire were given the option of completing the alternative assignment at a later date. The teachers completed the questionnaire at the same time as the students and followed the same procedures.

Data Analysis

Data was first screened at the item level to ensure the quality of the data was appropriate for subsequent analysis. First, the researcher inspected the data for straight-line responses (every response is the same), resulting in four participants being removed from the dataset. This was done to ensure that the data we used for analysis came from genuine responses from the participants. Straight-line responses indicate that the participants might have completed the survey as quickly as possible without giving adequate thought to the items. Reliability estimates were then calculated to examine the consistency of scores on the instrument. Cronbach’s alpha values above 0.70 would be considered an acceptable level of consistency for the current study (Dornyei, 2007). The unidimensionality of the constructs measured by the questionnaire (CRA, IRA, and NRA) were then examined using Winsteps (Linacre, 2021), a Rasch Analysis software program. Five items intending to measure each assessment approach were subjected to the analysis. Winsteps produces Eigenvalues and percentage of variance explained by the Rasch model and those left unexplained by the model (called residuals). Items are considered to be unidimensional if the amount of variance explained by the residuals is below 2.0 Eigenvalues or the percentage of variance explained by the Rasch model is more than four times that explained by the residuals (Linacre, 2016). After confirming that the constructs were unidimensional, the individual item scores were averaged together to form aggregate variables representing each assessment type (NRA, CRA, and IRA). Descriptive statistics were then calculated on the aggregate scores to verify that the data met the assumptions of normality. Items with skewness and kurtosis values within the absolute value of 2.0 were univariate normal (Field, 2009).

To answer research question one, the descriptive statistics for the students and teachers were consulted. To answer research question two, a within-groups Analysis of Variance was conducted to determine if there was a statistical difference among the means reported for the three assessment types. The effect size (η2) for the within-subjects variable (i.e., perceptions of CRA, NRA, and IRA) was also calculated. Results showing effect sizes up to 0.06 would be considered small, up to 0.16 considered medium, and 0.36 and higher indicate a large effect (Amoroso, 2018). Post hoc tests using the Bonferroni correction were conducted to determine statistical differences among the three assessment type.

Results

Results from Winsteps indicated that the constructs were unidimensional. The items on the questionnaire were internally consistent (ɑ = 0.73), as were the items representing each construct (CRA: ɑ = 0.72; NRA: ɑ = 0.72; IRA: ɑ = 0.77). The results from Winsteps and reliability estimates are presented in Table 1.

Table 1 Unidimensionality and reliability estimates of the study variables

The first research question asked how fairly Macau secondary school students and teachers perceived CRA, NRA, and IRA approaches to be. The descriptive statistics show that the students perceived CRA and IRA to be a somewhat fair assessment methods and NRA to be somewhat unfair. The teachers also reported similar results. CRA was the perceived to be a fair assessment method, IRA was viewed as quite fair, and NRA was viewed as quite unfair. The descriptive statistics for the three assessment approaches are presented in Table 2.

Table 2 Descriptive statistics for the study variables

The second research question asked if there are significant differences among the three assessment approaches for the students and teachers. Mauchly’s test of sphericity indicated the assumption of sphericity had not been met for the student data. Therefore, a Greenhouse–Geisser correction was performed. Results showed statistically significant differences among assessment types for students: F [1.79, 492.12] = 154.61, p < 0.001; η2 = 0.360. The effect size equaled the value for a large effect, suggesting that the within-subjects variable accounted for 36% of the total variance in the data for the students. Post hoc tests with Bonferroni correction revealed that students viewed CRA to be significantly fairer than IRA (p < 0.001) and NRA (p < 0.001). IRA was also significantly fairer than NRA (p < 0.001). For teachers, the assumption of sphericity had not been met, so a Greenhouse–Geisser correction was performed. The results indicate that there were differences among the three assessment types with a large effect: F [1.16, 15.13] = 28.15, p < 0.001, η2 = 0.684. The effect size exceeded the value for a large effect, suggesting that the within-subjects variable accounted for 68% of the total variance in the data for the teachers. Post hoc tests using Bonferroni showed that CRA was viewed as significantly fairer than IRA (p < 0.001) and NRA (p < 0.001). IRA was significantly fairer than NRA (p < 0.001). Altogether, for both students and teachers, CRA was the fairest assessment method, IRA was fairer than NRA, and NRA was the least fair.

Discussion

Fairness Perceptions from Students

The students perceived CRA to be quite fair as an assessment approach. This means that when their performance was measured against stated criteria, they viewed that assessment to be a somewhat fair evaluation of their ability. The CRA score in our study was lower than that reported by Dalbert et al. (2007) for German adolescent students, who perceived CRA to be fair, rather than quite fair. One reason for this difference may be the age differences between the participants in the studies. Our students were younger (M = 13.9, SD = 1.3) than the learners in Dalbert et al.’s study (M = 16.9, SD = 1.2), and the greater learning experiences of their students may have contributed to them feeling more strongly about CRA than our students. Another reason may be the cultural differences between the schools in Germany and the schools in Macau. German schools tend to use CRA in classes and Macau schools often use NRA to evaluate performance. The familiarity with CRA by the German students, or the lack of familiarity of the approach by Macau learners, could explain why the German students viewed CRA more fairly.

Students perceived NRA to be quite unfair. This means that when their performance was measured against their peers, our participants viewed the assessment approach to be somewhat unfair. This is somewhat surprising because our students studied in a NRA-oriented context, and despite being accustomed to and familiar with it, they perceived it to be unfair. In contrast, Dalbert et al.’s students were consistently evaluated in a CRA environment and also perceived NRA as quite unfair. One reason for this perceived unfairness in both contexts may be the subjective nature in which scores can be assigned in NRA. For productive language tasks (e.g., speaking and writing),Footnote 1 teachers may apply their own assessment criteria in addition to or in place of stated assessment criteria to distinguish among performances, especially if several performances were similar and the teacher needed to discriminate among them. These hidden criteria may not be available to students, which makes it challenging to determine if it is applied fairly and consistently to all performances. This could lead to challenges in the classroom because students have shown to be quite sensitive to how evaluators score their performance. For example, Bempechat et al. (2013) reported that Russian grade nine students were quite sensitive to evaluators’ lack of clarity and inconsistency in applying scoring criteria to their performance. Our participants may have felt similarly, highlighting an important concern that students may have with NRA: that the evaluation criteria and methods can be left vague and unsystematic, and the performance scores can be awarded based solely on evaluators’ subjective opinions without reference to a common metric. This lack of transparency among the scoring procedures would undoubtedly leave students feeling uncomfortable about their scores.

These results support notions that not only is it inappropriate to use NRA in classroom assessment (Brown & Hudson, 2003), but that it is unfair to students to do so. As highlighted earlier, using unfair evaluation methods can be problematic because it has been reported that when students feel they are assessed unfairly they may have a negative emotional response (e.g., stress), and then act on that emotional response through negative behaviors (e.g., complaining about the teacher; Chory et al., 2017; Chory-Assad & Paulsel, 2004). To avoid this, it is important to ensure that students feel they are being assessed fairly, or at the very least, that they are not assessed unfairly. One way to do this is by avoiding NRA in the classroom. However, it is acknowledged that teachers may not have a choice in the assessment methods they use, as administrations may require a normal curve for each assessment. Schools in some educational contexts require scores to match a normal distribution, with a certain percentage of scores reserved for high (e.g., maximum 10% A or A −), average (e.g., 40% Bs, 40% Cs), and low marks (Tan et al., 2020). The justification for this is to prevent score inflation, where too many high scores are awarded for performance, thereby assuring the integrity of instruction quality (Lok et al., 2016). However, schools appear to be taking a risk in doing this when their students perceive such an assessment approach to be unfair.

The students perceived IRA to be quite fair, meaning that when performance was measured against a prior performance, they considered it to be a fair evaluation method. Again, these results are consistent with those reported in Dalbert et al. Though IRA is considered somewhat fair by these two groups of learners, one reason why it may not be considered to be a fairer method is that it is infrequently utilized as a means to evaluate performance in the classroom and is typically considered a process-oriented and informal means to evaluate learning improvement over time (Brown & Lee, 2015). The benefit of using IRA is that it tailors assessment of learning to individual students and may be more effective than the other methods to address their unique needs by assessing learner improvement. In contrast, CRA and NRA apply a general scoring method to everyone and the underlying assumption of these two approaches is that every learner is at the same point of knowledge or skill level at the outset of learning. Students with less knowledge and/or experience with English writing, for example, would be disadvantaged in meeting learning outcomes compared with students with greater knowledge and experience, who may not need to work as hard to achieve higher scores. Because IRA evaluates growth, the implicit bias in CRA and NRA would be avoided. Certainly one challenge for IRA is to ensure equal measurement across all learners. Determining how much improvement would be needed to achieve a certain score and applying that equally across all learners would be difficult to accomplish. This may explain why IRA has not been utilized as much as the other assessment approaches in the classroom, despite it being viewed as a fair approach.

Of the three assessment approaches, the students perceived CRA to be fairest classroom assessment method. The reasons for this may stem from CRA having greater transparency in scoring compared with the other assessment types. In CRA, students are provided with the standards they have to meet in order to be given a score for a performance. Evaluators apply these criteria equally to all performances, making it clear to students how they were scored. This transparency can address the noted concern for learners that evaluators may unsystematically score performance (Bempechat et al., 2013). In contrast, scoring performance using NRA can be less transparent because the subjective opinion of the evaluator without reference to pre-established criteria determines whether one performance is better than another. It is understood that this can be done by evaluators relying on their gut feeling to assign scores or by applying a hidden scoring criteria in addition to listed standards so they can discriminate among the test takers. This latter approach may be taken when it is challenging for evaluators to distinguish among several similar performances, and adding unseen criteria (perhaps unintentionally) would help rank the test takers; though it may not be a fair approach to classroom assessment.

Students thought IRA was a fairer assessment method than NRA was well. IRA is a less interpersonally competitive evaluation method than NRA because learners need to outperform themselves to score well, regardless of what their peers do. It also eliminates the potential bias in scoring by learners who have more or less knowledge of the topic at the outset of learning. Whether or not these conceptions were considered by the learners in our study is speculative, however.

Fairness Perceptions from Teachers

The teacher perceptions largely mirrored those of the students, though they held stronger opinions about the fairness of CRA and IRA and the unfairness of NRA. These beliefs align with recommendations in the theoretical literature that CRA and IRA be used in classroom-based assessment and that NRA be avoided (Brown & Hudson, 2003). Our results contribute another layer of support for these recommendations, showing that teachers believe that, in addition to efficiency of measurement, the assessment approach is also a matter of fairness. If the approach taken is not appropriate for the classroom, then it can be seen as an unfair evaluation method. Our results further indicate that the Macau EFL teachers may have a high degree of assessment literacy. The teacher beliefs in our study are more similar to the Hong Kong teachers’ beliefs than the mainland Chinese teachers’ beliefs reported by Cheng et al. (2004). All three contexts involve Chinese EFL learning and are heavily test-oriented in their educational systems, but the Macau and Hong Kong teachers appear to be more literate in assessment practices by supporting the use of CRA and avoiding NRA in the classroom. An explanation for this is that these teachers may feel less constrained by the historical practice of using NRA in classroom assessments (Sasaki, 2008) and adopt alternative methods to evaluate whether learners are able to meet course objectives. However, this is only speculative because we were unable to confirm these opinions from the participants. Altogether, these findings support the use of CRA and IRA in the classroom instead of NRA because they are more appropriate for evaluating course objectives and fairer to the students.

Conclusion

Both students and teachers perceived CRA and IRA to be fair classroom assessment approaches and NRA to be unfair. Of the three, CRA was the fairest approach and NRA was the least fair. Theoretically, our findings add further support for the use of CRA and IRA in classroom evaluation because, in addition to them being more appropriate methods for evaluating student performance, they are considered to be fairer assessment methods. NRA remains a popular classroom assessment method today, but we hope that our findings have demonstrated that continuing to use it may be viewed as unfair by learners and teachers alike. This is especially important in the learning process because students can behave negatively when they feel their assessment is unfair.

We also acknowledge a few limitations of the study. Firstly, data was collected from students and teachers in one Chinese-medium of instruction school. To allow for our results to be more generalizable, more schools, including English-medium of instruction schools, would need to be included in the sample. Another limitation is that we primarily collected quantitative survey data. Future studies may consider using qualitative methods to elicit why teachers and students perceived the assessment methods as they did. Finally, the number of teachers participating in the study was limited. Including more teachers in future research may yield a more variety of opinions on the fairness of assessment. Overall, the current study makes an important contribution to the literature, showing that the classroom assessment method taken by teachers can have a strong impact on how fairly students feel they are evaluated.