A well-functioning education system requires adequate educational expenditure and basic infrastructure. Yet, research shows these alone will not ensure higher learning levels (Muralidharan 2013). Schools in India have received increased financial allocations (PAISA 2012), leading to some improvements in woefully inadequate school infrastructure, but student learning has not improved commensurately (ASER 2013). This raises an important question: In addition to basic infrastructure, what other school resources are important to improve learning?

One key challenge in addressing this question is identifying and quantifying relevant school resources. Studies typically rely on limited measures of school resources that capture just the basic infrastructure. When these infrastructure items appear insignificant, researchers may prematurely conclude that schools resources in general cannot make a significant difference to learning levels (see Chudgar and Luschei 2009 for a related discussion); however, they may not be measuring important aspects of school resources.

A related challenge is isolating the importance of school factors on learning. Students attending well-resourced schools likely belong to more affluent families; thus, it is not immediately evident whether higher performance at such schools reflects these schools’ resources or their students’ overall privileged circumstances. In such a case, knowing the students’ prior test scores may be particularly useful, allowing researchers to account for the influence of students’ circumstances on their learning thus far. However, prior test score data are also difficult to obtain from most standard data-collection exercises.

We overcome these data challenges via a collaborative partnership between researchers at universities in the United States and a large, well-established Indian educational foundation. The foundation has an intimate knowledge of school environments and school operations in the study location. Its ongoing and extensive school-based data-collection effort is a primary strength of our study. In consultation with the US-based scholars, the foundation team added a pre-test to its existing protocol of testing students once yearly and added modules to collect background data on students and teachers. These pre-test scores and rich background data are our study’s second core strengths. This partnership resulted in a rich and unique test-score dataset of 2,072 fifth-graders in 88 government schools located in 1 district in a southern Indian state. These data helped us explore this study’s guiding question: What school resources may be associated with improved student achievement?

The challenge of isolating the importance of school resources

The research literature most relevant to our work follows the “production function” framework of economics. Using this framework, researchers attempt to identify the role of such “inputs” as a child’s own characteristics, home background, classroom and school background, and teachers in “producing” a given educational “output”, usually measured as test scores.

Researchers generally agree that a child’s home background is crucial in explaining his/her educational outcomes: children in more educated and better-resourced families tend to have better educational outcomes. We focus on the role of inputs that policymakers can influence, such as teacher, classroom, and school attributes. Research on these factors is not consistent (e.g., Chudgar and Luschei 2009), as discussed below.

Production-function research in the United States: Identifying and measuring school resources

Extensive research from the United States shows that traditional school resources are not systematically associated with improved learning levels (Hanushek 1996, 1997). This research, however, is not without its critiques (e.g., Greenwald, Hedges, and Laine 1996; Krueger 2003); thus, debate continues about the role of resources and money in improving learning in the US context (e.g., Burtless 1996).

Uneven or inadequate quantification of key educational inputs such as school resources may explain, at least in part, why these debates have remained unsettled. But developing a comprehensive understanding of what is happening inside a school, and the related taxonomy, is a complex task. Cohen, Raudenbush, and Ball (2003) summarize the US-based literature that attempts to address these challenges. They identify three strands of research that aim to understand the role of school resources: one that investigates the “instructional processes and the resources used therein”, a second that focuses on the “schools’ collective characteristics”, and a third that examines “teachers’ and students’ interactions over specific content” (p. 121). The authors note the importance of what they term “personal resources” as a mediator between “conventional resources” and learning. They urge future researchers to pay close attention to “coherent systems of instruction” in order to understand how they influence learning. In other words, they argue in favor of developing a more nuanced understanding of what occurs inside a school.

Norton Grubb’s (2008) empirical research is an example of such a move beyond conventional resources to a more coherent and systemic understanding of school resources. Using a slightly different nomenclature, he identifies conventional school resources as “simple resources” and introduces the notion of “compound resources, complex resources, and abstract resources”. Compound resources become meaningful “in conjunction” with another resource, such as teacher experience and teacher training to teach the particular subject. Complex resources, such as innovative teaching approaches, are important but cannot be easily bought. Finally, abstract resources are “often embedded within a web of relationships and practices in a school” (p. 108); these include factors like positive school climate. Grubb’s (2008) research indicates that using these concepts greatly improves the explanatory power of models elucidating student performance. Thus, the resources he identifies as important for improved learning are not those that traditional measures identify.

Our brief review of the US-based literature underscores the potential limits of what researchers have termed “basic”, “traditional”, “conventional”, or “simple resources” for identifying factors significant for improved learning. We also highlight the value of adequately measuring and accounting for the complexities of school resources to understand what matters for improved learning. The presence of these challenges is equally unmistakable in the literature from developing countries.

Production-function research in developing countries

Scholars have argued that school resources may be especially important for improving learning outcomes in poor countries (Heyneman and Loxley 1983). Multi-country research following Heyneman and Loxley has questioned these findings (Baker, Goesling, and LeTendre 2002; Hanushek and Luque 2003). Most recently, Chudgar and Luschei (2009) revisited this question and found that schools are, indeed, relatively more important for learning in poor and unequal countries. However, that paper did not identify specific school resources that matter for improving learning, leaving unanswered the question we are interested in.

In terms of “simple” or “conventional resources”, it appears that the most recent and exhaustive review of the literature, from 1990 to 2010 (Glewwe, Hanushek, Humpage, and Ravina 2011), identifies few school resources that matter for learning in developing countries. From a database of 9,000 studies, the authors focus on 43 “high quality” studies. Simple or conventional resource measures in their synthesis include the availability of textbooks, furniture, technology, electricity, blackboards, a library, and other basic infrastructure. Some studies also include measures of the length of a school day, provision of meals, and parent and community participation, which Glewwe et al. group together as measures of “school organization”. The authors find that the impact of most school and teacher characteristics on student learning is statistically insignificant. Their review reveals the potentially limited importance of simple resources in developing countries and also underscores the fact that most studies from developing countries rely on limited school resource measures.

Production-function literature, or literature that aims to identify the role of educational inputs, in the Indian context is limited but growing. A recent exhaustive literature review from India (Muralidharan 2013) notes that this literature emphasizes teacher attributes over school attributes. An important feature of several of these studies (e.g., Atherton and Kingdon 2010; Kingdon and Banerji 2008; Kingdon and Muzammil 2013; Kingdon and Sipahimalani-Rao 2010) is their taking account of extensive student background variables, including prior-test scores.

This approach, also referred to value-added modeling (VAM), uses students’ prior performance as a control in production-function models (Koedel and Betts 2011). This allows researchers to better isolate inputs introduced after prior test–score data. However, researchers debate what are appropriate VAM model specifications; they have examined assumptions underlying various value-added specifications (Guarino, Reckase, and Wooldridge 2012; Harris and Sass 2006) and often critiqued the standard VAM approaches for producing implausible results (Baker et al. 2010; Corcoran 2010; Rothstein 2010). In general, though, agreement is emerging that including a lagged or prior achievement score, as several Indian studies do and as our study proposes to do, may be an adequate value-added model specification (Harris and Sass 2006; Guarino et al. 2012).

Thus, our work is similar to some of the recent production-function studies from India regarding its use of prior-test scores as a control. However, our study differs from this existing work from India in its exploration of school resources as important for achievement. Existing research as noted above primarily investigates the impact of various teacher attributes on learning.

To summarize: while school resources may play a role in learning (especially in poor and unequal contexts), researchers have not ascertained which are important. Researchers are often limited by their ability to identify and measure relevant and complex aspects of school resources. We have overcome this challenge through a close collaboration between experts in the field and researchers in the US, resulting in the collection of a unique dataset.

The school resources instrument: Development and existing use

Brief note on the administrative hierarchy in the Indian education system

India has a federal system, with 29 states (divided into 600 districts) and 7 union territories. Districts are divided into blocks, and blocks into clusters. The educational infrastructure exists at the national, state, district, block, and cluster levels.

While the national and state levels set agenda, the block- and cluster-level machinery, particularly the Cluster Resource Coordinators (CRC), support the day-to-day functioning of schools within their purview. The CRCs are responsible, on average, for 10–15 schools and are the first point of academic contact for teachers. Their primary responsibility is to provide academic support to teachers by visiting their assigned schools regularly and by conducting regular cluster-level meetings to enable exchange of pedagogic ideas between the teachers.

Foundation’s work and development of the school resource instrument

This study is located in a relatively underdeveloped, northeastern region of a southern Indian state, where the foundation has worked in education for over a decade. Its efforts support the state government’s work to improve education in this region. The foundation initially recruited 25 school “guides” (SGs) to work closely with schools (SG to school ratio is currently 1:14). The work of SGs is similar to that of CRCs. SGs visit their assigned schools regularly and maintain close contact; they support the teachers, the principals, school organizers, and the community in various educational activities. They also facilitate various foundation interventions. Their long and extensive association with these schools gives SGs a unique understanding of the school context. A key aspect of their work, relevant to our study, is to gather data on their schools.

The foundation has adopted an “integrated and holistic” approach toward education (APF 2013, p. 9). Initially, the foundation identified five areas for work and intervention: school environment, class environment, teaching learning process, teacher development, and community participation. It has launched interventions in each of these areas (APF 2013, p. 29) and regularly evaluates this work through systematic data collection activities.

To gather these data, the foundation team originally developed a 214-question school indicator survey (SIS) that encompassed the 5 key areas of intervention discussed above. It used these data to inform a school-improvement plan that local stakeholders (including parents and community members) developed jointly. SGs participated in these meetings and helped stakeholders interpret data and generate school-improvement plans.

The team recognized, over time, that the 214-item SIS did not track schools effectively. In their own words, “the indicators captured infrastructure and other tangible items rather than the progress in the teaching-learning process” (APF 2013, p. 26). This group of practitioners thus arrived at a conclusion similar to the literature about the relative importance of simple versus complex and compound resources in understanding what might matter in improving educational outcomes.

They produced a 60-item SIS and began collecting the variables at different time intervals. They collected some information fortnightly and some monthly, quarterly, and even annually. This time frame of data collection became a primary mode of grouping these variables. The 20 fortnightly indicators included variables related to cleanliness, hygiene, food and water provision, and teacher attendance; and other variables where changes occur even on a daily basis. The 6-month indicators included such variables as School Development and Monitoring Committee (SDMC—a parent-teacher body required in all public schools) meetings or teachers’ lesson plans. Quarterly indicators included 8 measures that varied every few months, such as community contributions and student dropout rates. Finally, the SG monitored 26 variables on an annual basis: these included information on whether an SDMC existed, whether teachers had their mandatory annual trainings, whether the required infrastructure was in place, and whether annual events (e.g., sports meet, cultural day) had occurred. This 60-item SIS provides a unique and comprehensive picture of the schools we analyze here.

Data

The study uses data from one rural block in the region where the foundation has been active. The block consists of 190 villages, with over 300,000 people, including the largest proportion of rural population in the district (nearly 87%) and a sizeable minority population. It has high socioeconomic disparity and low female literacy (just under 40%). The foundation, active in over 340 government schools in this block, works with 1,600 teachers and more than 60,000 children. For our study, we excluded lower primary schools (LPS), or schools that teach only up to grade 5. We excluded 7 Urdu medium schools due to lack of Urdu language assessment tools at the time of the survey. Our study uses data from 88 higher primary government schools (HPS) that teach up to grade 7, and from 2,072 fifth-grade students at these schools.

The foundation team entirely managed and carried out data collection for this study; the researchers obtained the data from the foundation in a de-identified format. The team collected data in August–September 2011, toward the beginning of the school year, and again in March 2012, toward the end of the academic year. The SIS formed one of the key questionnaires for data collection; the foundation team and the researchers jointly developed additional questionnaires as needed. SGs administered the questionnaires in their respective schools.

The resultant data had information: (a) from the 60-item SIS; (b) on teachers’ gender, experience, training, and education; (c) on principal-reported school enrolment and principals’ perceptions of their schools’ percentage of children from poor families; (d) on student-reported demographics and home-background; and (e) on student baseline and end-line test-scores in math and environment science (EVS). SG gathered these test data by administering each child with a subject-specific test instrument.

EVS’s inclusion in school curricula in India reflects global educational discourses around environment education (EE) since UNESCO’s Tbilisi Declaration in 1977. In 2003, an order from the Supreme Court of India made the “green curricula” mandatory for all states (see Iyengar and Bajaj 2011). Per the National Curricular Framework 2005, EVS’s aim is to make children aware of environmental issues by helping them to make a connection with their natural and human-made surroundings (Directorate of State Education Research and Training 2012). The EVS curriculum is action-oriented and project-centered, with participatory approaches to teaching. It envisions multidisciplinary and integrated courses, drawing from the sciences (physical, chemical, and biological), from social studies (history, geography, civics, etc.) and from environmental education (protection and conservation). Thus, the two subjects we have test-score data for—math and EVS—might be particularly divergent: math relies on a formal, codified body of knowledge, while EVS attempts to create a more holistic, experiential learning experience.

Preparing variables for analysis

To prepare the variables for analysis, we collapsed the teacher information at the school level, such that for each school we generated values representing average teacher experience, proportion of female teachers, proportion of teachers with a bachelor’s degree or more, and proportion of teachers with a diploma in education. Teacher attributes in our analysis, therefore, do not account for the students’ specific teachers but, instead, provide an overall sense of the teachers at a given school.

We next had to group the SIS’s 60 items, as it was untenable to use these items separately in a regression framework. We decided to use 57 items, excluding data on total numbers of students and teachers, and information that no students had dropped out that quarter—these variables departed from the more standard “resources” we focused on.

Based on concerns about sample size, we rejected using factor analysis in constructing our composite measures of school resources. The literature’s recommendations for item-to-observation ratios for factor analysis were 1:5, 1:10, and even 1:20 (Kline 2013). Researchers have noted that a small sample size where the item-to-observation ratio is as low as 1.5:1 (as in this study) will produce incorrect solutions and misclassify items (Costello and Osborne 2005). Small sample sizes are associated with the data’s failure to converge to a solution and the possibility of producing outcomes that do not make practical sense. After extensive discussion and, exploratory analysis we adopted a more organic iterative approach to data reduction, which helped us identify and group variables into eight indices, each based on variables reflecting a similar underlying concept. After grouping these variables, we generated a simple average measure for each of the eight indices, ranging from 0 to 1—where a larger number meant a greater presence of the indicated items.

These data allowed us to produce an extensive set of simple and more complex resources. In addition to measures of basic facilities, we have information on tangible (e.g., learning facility, school organization, mid-day meals) and intangible (teacher preparedness, learning environment) school resources available to children, as well as their parents and community’s participation in their schools’ day-to-day functioning. We use these school indicators to investigate if, and to what extent, any of these resources matter in understanding variation in student performance.

Method

Given the continuous nature of the dependent variables (math and EVS performance) and the clustering of children within schools, we used an ordinary least squares (OLS) regression analysis with cluster correction at the school level. For both math and EVS, we first conducted regression analyses including student, teacher, and school background as controls, and excluding any of the eight school resource indices. We then introduced these eight indices one at a time.

$$ {\text{A}}_{\text{ijt}} =\upgamma1 +\upgamma2 \,{\text{A}}_{{{\text{ij}} \left( {{\text{t}} - 1} \right)}} + {\mathbf{X}}_{\text{ij}}\, \upgamma3 \, + {\mathbf{ST}}_{\text{j}}\, \upgamma4 +\upvarepsilon{\text{ij}} $$
(1)
$$ {\text{A}}_{\text{ijt}} =\uppi1 +\uppi2 \,{\text{A}}_{{{\text{ij}} \left( {{\text{t}} - 1} \right)}} + {\mathbf{X}}_{\text{ij}}\, \uppi3 \, + {\mathbf{ST}}_{\text{j}}\, \uppi4 +\uppi5{\text{SC}}_{\text{j}} { + \varepsilon }{\text{ij}} $$
(2)

where,

\( {\text{A}}_{\text{ijt}} = \) EVS or math scores for the individual i in school j at time t;

\( {\text{A}}_{{{\text{ij}}\left( {{\text{t}} - 1} \right)}} = \) EVS or math scores for individual i in school j at time t−1;

\( {\mathbf{X}}_{\text{ij}} = \) vector of background characteristics of individual i in school j;

\( {\mathbf{ST}}_{\text{j}} = \) vector of teacher and school-related characteristics for school j;

\( {\text{SC}}_{\text{j}} = \) school resource index, introduced one at a time.

Results

Descriptive analysis

Table 1 provides a detailed list of all the child, teacher, and school variables, with their description. The EVS and math baseline and end-line sample sizes reveal a slight discrepancy, since some children might have been absent during tests.

Table 1 Student, teacher, and school background characteristics, means, standard deviations, and frequencies

The data indicate an improvement in mean scores between the baseline and end-line testing, though the change is not directly comparable since these two sets of tests had slightly different ranges for both EVS and math. In our sample, 47% of the respondents are girls, their average age is nearly 11 years, and 37% of them reported belonging to the scheduled castes (SC) or scheduled tribes (ST). The average household size is around 5 persons. A majority (68%) of these children report that their mother is illiterate; 46% report an illiterate father; and in 43% of the families no one speaks English at home, where “English at home” is treated as an additional proxy of the family’s ability to access and navigate complex educational choices. Reflecting the overall low economic development of the region, 54% report having a vehicle at home and 66% report regular electricity in the house. Most children’s families have at least one “possession”—mobile phone, land line, or television. In terms of the learning environment: 36% of children receive no homework help at home, 25% report not discussing their schools at home, 80% do not receive private coaching or tuition, 58% report not reading at all at home, and over 60% report not using the library. Overall, thus, these children belong to poor families with limited education. Although it is unremarkable, given the extent of illiteracy among their parents, that most of these children do not read at home or use the library, these numbers are large and striking.

Table 1 also describes the average teachers in this dataset. Only 31% of teachers in our sample are female—not surprising, given that our sample excludes lower primary schools and that we collected these data from a rural block. The literature has associated both these factors with a lower presence of female teachers. In our sample, 28% of teachers have a BA degree or higher, and 70% have a diploma in education. Average teacher experience is approximately 11 years. These schools enroll approximately 300 students on average. According to the principals, on average, 68% of these schoolchildren belong to “economically disadvantaged (economically weaker) sections of society”. Regarding the eight school resource indicators: on average, schools tend to have high levels of basic facility (0.73) and school organization (0.79). Similarly, mid-day meal indicators are also present more widely, on average (0.69). School learning environment and teacher preparedness indicators are, however, less evident, with an average of 0.60 and 0.57, respectively. Finally, the presence of school learning facilities, co-curricular activities, and community participation seems relatively less consistent. On a scale of 0 to 1 (where 1 implies presence and 0, absence) schools, on average, score 0.50 or lower.

Correlation analysis (not shown) indicated that these eight school resource indicators are positively and significantly correlated with one another. Yet, the correlation between most of these variables is not so large as to obviate the need for these separate variables. The correlation also signifies that the presence of one school-based advantage makes it more likely that the child will experience other school-related advantages as well.

Regression analysis

Regression results for equations (1) and (2) are presented in Table 2a, b. In both of these analyses controlling for students’ prior performance was crucial as indicated by the consistent, positive, and significant coefficients associated with baseline scores. In analysis not shown here we also noted an appreciable increase in the overall model R-square after inclusion of prior student performance.

Table 2 OLS regression, with baseline test-score

Focusing first on EVS results, girls seem to perform systematically poorer than boys. Fathers’ illiteracy is also statistically significant and negatively associated with EVS performance. A child who receives no help for homework, or has no discussions about school at home, performs poorly, but these relationships are not statistically significant. The negative and significant coefficients associated with a child who does not read and does not use the library are especially striking given that we are also controlling for the child’s prior performance. These two variables may be a proxy not just for the child’s reading behavior but also for his or her overall learning environment outside of school. Of the school variables, average teacher gender, experience, education, or training do not appear to be significantly associated with EVS performance. However, we do note that children attending large schools tend to perform poorly in EVS.

For math, many of the patterns are similar; we find that girls perform significantly more poorly than boys. Also, receiving no homework help, and not engaging in any out-of-school reading are all associated with negative math performance. Fathers’ illiteracy or library use are, however, not associated with math performance. Once again, large school size and also greater presence of children from low-income families is associated with lower math performance.

The remaining columns in each table (Table 2a, b) add different school indices one at a time. The column title in each case indicates the school resource index that was introduced in the model. Before we discuss the role of these specific school resource indices, a few additional observations emerge from this analysis. We note that almost all the child-level variables and even the few school-level variables that were significant in the initial analysis remain significant throughout. Inclusion of these exhaustive school resource indices does not seem to diminish the importance of, say, fathers’ illiteracy or household size. However, in case of math, English speaking ability at home actually becomes significant in some of these later cases. Perhaps most striking is the consistent significance of the child’s reading habits. A child who does not read at home performs poorly regardless of what other explanatory variables we add to the model, and regardless of the subject. Similarly, in their respective models, library use and homework support both remain important.

In terms of school indices, we observe more significant relationships for math compared to EVS—not altogether surprising, given the distinct nature of these two subjects and the potentially greater role for school-based work in math. For math, we find that basic facilities, teacher preparedness, and learning environment indices remain insignificant. School organization and community participation coefficients are also insignificant. The coefficient associated with provision of the school mid-day meal is insignificant but with a p-value of 0.108. However, for math, after an extensive set of student- and school-background controls, including prior test scores, we find that having learning facilities in school (e.g., laboratory, library, resource center) and having co-curricular activities are both significantly and positively related to math performance (at p≤0.05 and p≤0.10), respectively). The analysis thus provides some support for the argument that some specific, nuanced school resource measures may be important in understanding variation in student performance in math in particular.

Limitations, conclusion, and discussion

In this study we aimed to use a unique school-based dataset to identify school resources associated with student performance.

A key limitation of this study is closely related to its main strengths: having one data source means that the data are not nationally representative or even representative of the state they are gathered from. The data represent schools from one block in India where the foundation is active. Nonetheless, to the extent that these data record school circumstances of relatively disadvantaged children, they provide some insights into such schools and may inform future large-scale analyses.

The other key limitation of our work is our inability to establish causal relationships. Our ability to control for prior test scores helped us to remove some bias in our analysis. But to establish causality, researchers ideally need to benefit from a well-designed experimental or a quasi-experimental study. Our approach using retrospective, cross-sectional data does not allow us to infer such cause-and-effect relationships.

Despite these limitations, the study provides important insights into factors associated with improved student performance. The first is the importance of controlling for prior performance, where possible, which improves the overall explanatory power of the model. Even controlling for prior performance, we found that certain student background and home-level factors were invariably important. If students did not read outside school, they consistently performed lower in both math and EVS, regardless of what other controls we introduced. Reading outside of school potentially indicates the overall learning environment available to a child once he or she leaves school; it means having access to reading material, having quiet time (i.e., not needing to work), and either having sufficient basic reading skills or having an adult at home who can read. Similarly, the importance of library use, receiving homework support at home, and fathers’ literacy levels in different models are all noteworthy. Data from these factors indicate what we noted earlier: students’ own backgrounds and their home circumstances are crucial to understanding their performance.

Noteworthy and disappointing is the consistent underperformance of female children after accounting for other variables. In India, girls’ lesser access to educational opportunities is well documented, but the phenomenon of female underperformance in tests is not well studied. Our results show a need for greater research in this direction.

Within the school, average teacher attributes do not seem to be important. This finding generally aligns with research acknowledging that teachers are important; but asserting that there is less consensus on what aspects of teacher backgrounds matter for improving student performance (e.g., Luschei and Chudgar 2011). Within the school, however, we find that studying in a larger school, or with more low-SES children, is associated with generally poor performance. The literature supports this latter association; and while a separate study by the foundation team in the same block has also revealed a potential association between lower performance and large school sizes, this relationship warrants more attention.

Returning to the question that motivated this study: What school resources may be associated with improved student performance? First, our findings lend support to prior research that notes that exploring “simple” or “traditional” school resources may not be enough to understand improved student performance. This is not to argue that basic infrastructure such as clean water, functioning toilets, or hygienic learning environments are not important—they are important preconditions for a well-functioning school. But the presence of these infrastructural resources alone does not explain variation in learning performance across schools in this data.

Schools with more learning-specific facilities (library, laboratory, and other resources) and with more co-curricular activities (sports, cultural events) are schools where all else equal, children perform well in math. Similar model specifications for EVS do not yield any significant school resource variables. This may be partly explained by the differences, discussed earlier, in the nature of these two subjects.

An equally, or even more, important “finding” is the challenge of satisfactorily measuring school environments that support learning improvement. This is difficult, even with the benefit of a nuanced, context-specific instrument that has evolved with much thought from experts in the field. So, what lessons emerge from this work for future research, policy, and practice?

For researchers, the perils of looking inside schools with large-scale, standardized datasets are clear. It is not surprising that researchers using standard, large-scale data with traditional school resource measures are unable to identify important school factors for improved performance. Yet, one should not ignore the potential importance of school environments altogether, especially in poor and less developed parts of the world. Rather, this “finding” is a call to improve the current measures of school resources. From a policy and practice perspective, the foundation team’s extensive efforts offer potential guidance for gathering nuanced measures of school resources that go beyond the traditional ones. In the Indian context, the existing CRC system provides a cluster-level infrastructure to gather this form of nuanced school data. Thus, the potential exists to develop and refine similar school-based surveys that can provide valuable data for scholars and practitioners in their efforts to improve student performance.

Finally, perhaps, this study serves as an example of a fruitful and productive partnership between scholars and practitioners. The foundation’s willingness to engage in additional research efforts alongside its existing practices and the technical knowledge of the scholars based in the US created a unique learning opportunity for both groups and has added value to our collective knowledge.