Skip to main content

An IERI – International Educational Research Institute Journal

Inequality in remote learning quality during COVID-19: student perspectives and mitigating factors

Abstract

Background

Remote learning, or synchronous or asynchronous instruction provided to students outside the classroom, was a common strategy used by schools to ensure learning continuity for their students when many school buildings were forced to shut down due to the COVID-19 pandemic. Differences in technology infrastructures, digital competencies of students and teachers, and home supports for learning likely led to inequalities in the way remote learning reached and was perceived by students. This study seeks to understand how student perspectives on remote learning varied across and within several countries.

Methods

Building off a conceptual framework developed to understand remote learning success and using data from the Responses to Education Disruption Survey (REDS) student questionnaire from seven countries, we construct measures of student perceptions of three essential components of successful remote learning: Access to Suitable Technology, Effective Teachers, and Engaged Students. We then compare values on these scales across and within countries to identify inequalities in remote learning quality during school closures. We also investigate the extent to which schools implemented supports for remote learning across countries.

Results

We find evidence of across country variation in remote learning quality with certain countries having much lower values on our remote learning quality scales compared to other countries in our sample. Furthermore, we identify within-country inequalities in access to and confidence in using technology with low-SES students, girls, and those living in rural areas having lower values on these measures. Furthermore, we find some evidence of within-country inequalities in student engagement across socioeconomic groups. In contrast, we do not find as many inequalities in our measures of effective teachers. In most countries, schools provided several supports to improve remote learning.

Conclusions

While inequalities in remote learning experiences were anticipated and confirmed by our results, we find it promising that, in some countries, inequalities in access to and confidence in using technology as well as student engagement did not extend to inequalities in perceptions of teacher effectiveness and support. Schools’ efforts to support remote learning, regardless of student background, should be seen as a positive and illustrate their resilience in the face of many challenges.

Introduction

In response to concerns over the COVID-19 pandemic, educational systems around the globe made the decision to close schools to halt the spread of the coronavirus. According to UNESCO (2020), this decision impacted more than 90% of students globally. As schools were no longer able to offer face-to-face instruction, educational leaders implemented remote learning strategies to maintain learning continuity for their students. Remote learning can be defined as

Synchronous or asynchronous instruction provided in a place outside the classroom. Synchronous learning means that students are connected to learning experiences where a teachers’ immediate feedback is possible. Asynchronous or self-directed learning means that students can learn at their own pace and chosen time. Remote learning takes an array of forms ranging from paper-based take-home packages to online platforms. Remote learning is also possible through a variety of different channels, such as mobile phones, television, radio, and tutors. (Muñoz-Najar et al., 2021, p. 16).

Since most forms of remote learning require access to and the use of several learning or communication technologies that have not been widely adopted in the educational context, it is likely that inequalities in remote learning quality existed. This paper seeks to understand how student perspectives on remote learning varied across and within countries.

This study builds off of a conceptual framework constructed to understand remote learning during the COVID-19 pandemic. In the framework, the World Bank (Muñoz-Najar et al., 2021) identifies three essential components for remote learning take-up and effectiveness. The first component is access to technology appropriate for the remote learning needs. Learning technology not only has to be made available to students to access remote learning, but must be appropriate for the context in which it is deployed. For instance, in areas with low internet coverage, alternative strategies to online methods for providing students access to learning are necessary (e.g., paper-based materials, television or radio broadcast lessons). The second component is teacher knowledge and skills for remote teaching. This not only requires teachers to be equipped with in-depth content knowledge and strong pedagogical skills, but also to have the ability to use and integrate technology into their teaching practices. The final component is student engagement. Keeping students engaged in learning requires interesting and engaging content as well as frequent interactions with teachers (e.g., through feedback on work). According to the framework, all three components, Available and Suitable Technology, Effective Teachers, and Engaged Students are necessary in order to provide a quality remote learning experience for students (Muñoz-Najar et al., 2021).

In this paper, we define remote learning quality as the successful take-up and effectiveness of remote learning. Due to variation in existing technology infrastructures, digital competencies of students and teachers, and home resources for learning, we hypothesize that there will be inequalities in remote learning quality, both across- and within-countries. However, we also hypothesize that certain school-level supports might have been implemented to mitigate these inequalities. Specifically, we seek to answer the following research questions:

  1. 1.

    What inequalities exist in perceived remote learning quality across- and within-countries?

  2. 2.

    How did countries differ in their use of school-level supports/resources aimed at improving remote learning?

To answer our research questions, we use data from the Responses to Educational Disruptions Survey (REDS) study (Meinck et al., 2022). REDS gathers perspectives from students, teachers, and principals in secondary education (grade 8) on both the impacts of and responses to the educational disruption caused by the COVID-19 pandemic. REDS collected data from 11 countries spanning Africa, Asia, the Arab Region, Europe, and Latin America. Using data from the student questionnaire, we construct measures of the essential components of remote learning to capture student perspectives on the quality of their remote learning experience. With these measures, we examine the extent to which inequalities exist within and across countries in perceived remote learning quality. We also examine responses from the school questionnaire to identify supports and resources that were provided by schools to improve the quality of their students’ remote learning. Our study focuses on seven countries: Denmark, Ethiopia, Kenya, the Russian Federation, Slovenia, the United Arab Emirates, and Uzbekistan.

Literature review

Here, we review the literature and elaborate on and more clearly define each of the essential components of remote learning. We also explore the extent to which the literature has identified existing inequalities in these measures. We end by noting some of the supports or resources that schools or national ministries provided to educators to support each of these components and help them deliver high quality remote learning to all students.

Available and suitable technology

The first essential component of remote learning covers two aspects of technology: availability and suitability. Before determining whether the technology is suitable, it is necessary to assess the access to technology in the first place. Access to technology is frequently seen through the lens of the so-called digital divide, which refers to the gap between those who do and those who do not have access to information and communication technologies (ICTs) (Scheerder et al., 2017; van Dijk, 2006). This includes access to reliable and affordable physical devices (e.g., computers, laptops) and Internet connectivity, which are basic requirements for remote learning that takes place online (ITU, 2021a).

According to the International Telecommunication Union (ITU, 2019), prior to the COVID-19 outbreak, 93% of the world population could, in theory, access the Internet, as they were within range of a mobile broadband or Internet service. However, there are large inequalities in access both between and within countries. For example, only 23% and 11% of households in Africa and the Commonwealth of Independent States, respectively, have access to a mobile broadband network (ITU, 2020b). Within countries, gaps exist between urban and rural areas, as Internet access at home is twice as high in urban areas (72% compared to 37%). A similar situation is seen for computer access, for which 63% of households in urban areas have access, compared to only 25% in rural areas, with developing countries’ rural areas being the most disadvantaged. The differences in Internet and computer access between urban and rural areas also vary across countries. In Africa, only 28% and 17% of urban households have access to the Internet and to computers, respectively, and these figures drop to 6% and 2% for households in rural areas. In contrast, 88% and 82% of urban households in Europe have access to the Internet and computer access, while 78% and 66% of rural households have such access. Within countries, data also shows that those without access are typically the most vulnerable groups, such as women, children and youth from disadvantaged socioeconomic backgrounds, people with disabilities, and indigenous and marginalized groups (ITU, 2020a, ITU, UNESCO, & UNICEF, 2020).

Beyond access: the second-level digital divide

Although 93% of the population could access the Internet, just over 53% actually use it (ITU, 2019), indicating that having the opportunity to access computers and the Internet does not necessarily mean that they will be used. Therefore, the digital divide is also seen in terms of differences in digital skills or actual use, which is refered to as the second-level digital divide (Hargittai, 2002; ITU, 2021a van Dijk, 2005; Scheerder et al., 2017; Wei et al., 2011).

According to the ITU (2020a), skills are considered one of the main barriers to Internet use. For instance, data from a representative household survey compiled by the ITU show that around 65% of the population in developing countries are not using the Internet because they do not know what it is or they do not know how to use it (ITU, 2020a). The results of ICILS 2018 show that, on average, 18% of students did not have a functional knowledge of computers (Fraillon et al., 2020), putting into question students’ readiness for the shift to remote learning brought by the pandemic. Data from ICILS 2018 also shows inequalities in students’ computer and information literacy, particularly within countries (Fraillon et al., 2020). The review by Scheerder and van Dijk (2017) shows that the main determinants of the second-level digital divide are sociodemographic (e.g., gender, urban/rural dimension, residency) and economic (e.g., SES, education level, parental education) in nature, more so than material factors (e.g., ICT access at home, access quality). Besides actual skills, some researchers use self-reported skills (i.e., self-efficacy) as a proxy variable to study the second-level digital divide (Hargittai, 2005; Zhong, 2011).

The suitability of technology for remote learning

For effective remote learning, the availability or access to technology is necessary but not a sufficient condition, as it also needs to be suitable for the context in which it is used (Muñoz-Najar et al., 2021). During the COVID-19 school closures, there was a variety of remote learning strategies that educational systems could implement, such as online learning solutions, radio or TV broadcasts, or even paper-based take-home packages (Meinck et al., 2022; Muñoz-Najar et al., 2021). However, some countries found themselves in a remote learning paradox when they implemented unimodal strategies for remote learning that did not match the possibilities and needs of the majority of the students (Muñoz-Najar et al., 2021). For instance, when online learning was deployed but a majority of students lacked devices, had connectivity constraints, or did not have sufficient digital skills for this type of remote learning strategy.

Effective teachers

It is well-known from previous research that teachers matter for student learning outcomes. Although measuring the impact of teachers is not straightforward, there seems to be some consensus about the importance of instructional quality. In fact, instructional quality has been regarded as the most important variable for student learning outcomes (Creemers & Kyriakides, 2008; Hattie, 2009; Scherer & Nilsen, 2016), more so than other teacher characteristics such as their qualifications or background. In other words, what matters more for student outcomes is what teachers actually do in the classrooms, as it is inside the classroom where teachers and students interact and where learning ultimately takes place (Creemers & Kyriakides, 2008; Goe, 2007). Therefore, we focus on instructional quality as a measure of effective teachers.

Instructional quality is regarded as a multidimensional construct (e.g., Fauth et al., 2014; Klieme et al., 2009; Nilsen et al., 2016), and one of its key aspects is supportive climate. A supportive classroom climate covers aspects of teacher-student interaction, such as providing positive and constructive feedback, having a positive approach to students’ errors and misconceptions, providing extra help when needed, listening and respecting students’ ideas and questions, and caring about and encouraging students (Blömeke et al., 2016; Klieme et al., 2009). Across the literature, there is evidence that a supportive climate affects student engagement and, consequently, student achievement (Berkowitz et al., 2017). The review by Berkowitz et al. (2017) also shows that a positive school climate—which includes classroom supportive climate—decreases the correlation between SES and student achievement. Bergem et al. (2020), based on Norwegian data, also find that instructional quality, including aspects related to teacher support, is important for students’ motivation to learn mathematics, particularly for low-SES students.

Engaged students

Effective remote learning also requires student engagement. According to Skinner et al. (2009), in general, engagement refers to “the quality of a student’s connection or involvement with the endeavor of schooling and hence with the people, activities, goals, values, and place that compose it” (p. 494). Across the literature, student engagement includes aspects such as motivation to learn, emotional responses to the learning activities and the learning environment (e.g., towards teachers, classmates, school), absence of disruptive behavior, and involvement in learning tasks (e.g., paying attention, asking questions, contributing to class discussion) (Fredricks et al., 2004; Fredricks et al., 2012; Salas-Pilco et al., 2022).Footnote 1

Student engagement is considered an important educational outcome by itself (Moore et al., 2015) but research also shows that it is a crucial factor in predicting academic achievement (Alrashidi et al., 2016) and influencing other outcomes, such as emotional well-being (Subramainan & Mahmoud, 2020). Engagement is recognized to be highly influenced by contextual factors, representing an attractive and potentially malleable influence for student learning (Fredricks et al., 2004). Examples of such factors are good-quality teacher-student relationships (Quin, 2017), and teacher and parent support (Lam et al., 2016; Wang & Eccles, 2012). In the context of the pandemic, the World Bank highlights access to suitable remote learning technology, access to engaging content, regular feedback and motivation from teachers, and a suitable home learning environment as factors that can influence the engagement of students (Muñoz-Najar et al., 2021). Additionally, researchers have highlighted the role of self-efficacy for motivation and engagement (Schunk & Mullen, 2013).

Regarding equity in student engagement, specific subgroups of students are at higher risk of suffering low levels of engagement (OECD, 2013). For example, previous research suggests boys tend to show lower engagement (Lam et al., 2016; Lietaert et al., 2015; OECD, 2013). However, educational barriers disproportionately affect girls in developing countries due to increased domestic responsibilities and gender bias, which could also affect their behavioral, emotional and cognitive engagement. Results from the OECD (2013) also suggest that students from low socioeconomic status are also more likely to show lower levels of engagement.

School supports during the COVID-19 educational disruption

As the COVID-19 pandemic led to widespread closures of schools, educational systems were faced with the unprecedented challenge of continuing learning without the typical resources that are available in in-person learning environments. In response, national ministries or local governments overseeing schools provided guidelines for strategies to support learning continuity during the educational disruption (American Institutes for Research, 2021; Barron Rodriguez et al., 2021; Stancel-Piatak et al., 2022). Many of the strategies were explicitly or implicitly aimed at supporting the three essential components for remote learning success (Muñoz-Najar et al., 2021).

Providing access to suitable technology

Schools were advised to provide students and teachers with the tools necessary to access remote learning spaces. To reach this goal, governments would often partner with the private sector to facilitate student access to remote learning. For instance, many countries noted that they heavily subsidized the cost of internet or data plans (UNESCO, UNICEF, the World Bank, & OECD, 2021). While many schools distributed technology (e.g., laptops, broadband internet) to homes, several countries or areas within countries had limited internet coverage leaving it impossible to hold lessons in an online format. In these locations, it was necessary to make sure that other forms of instruction were made available (e.g., television and radio broadcasts, paper-based materials). For example, in Uzbekistan, teachers were encouraged to prepare video lessons that would be broadcast across six TV channels of the National TV and Radio Company. In Kenya, the Ministry of Education provided radio lessons to approximately 40 community and church radio stations to be broadcast to students. In instances where these video or radio lessons were not accessible, paper learning materials were encouraged to be distributed to homes. In all countries included in our analysis, national plans or policies for remote learning either required or recommended the physical distribution of learning materials to students (Stancel-Piatak et al., 2022).

School supports for effective teaching

In supporting effective teaching practices during remote learning, many schools were advised to provide professional development opportunities targeted at supporting teachers in their use of technology and delivering teaching remotely. Prior to the pandemic, several countries had begun to integrate the use of technology in teaching and learning (Fraillon et al., 2020). Therefore, some countries already had existing infrastructure in place to support the transition from in-person to remote learning. For example, in Uruguay, Plan Ceibal had been in place since 2007 aiming to increase and promote the use of digital technologies in education. During the pandemic, Plan Ceibal was adapted and strengthened its services provided to educators, including training and support in the use of virtual learning environments (Stancel-Piatak et al., 2022). A large majority of teachers in Uruguay were satisfied with the training provided by the program (Muñoz-Najar et al., 2021).

In addition to professional learning opportunities, several national guidelines explicitly stated that teachers were to be provided more opportunities to meet and collaborate with their peers (Stancel-Piatak et al., 2022). Collaboration among teachers in a school can facilitate learning and support innovation in teaching methods (Vangrieken et al., 2015). In the United Arab Emirates, educational institutions saw the potential benefits of facilitating collaboration among not just teachers, but schools. In their efforts to support remote teaching and learning during school closures, they established a peer pairing program for struggling schools to work with others who were seeing successes. Furthermore, an online platform was used for schools and teachers to share remote teaching and learning resources.

Supporting student engagement

Synchronous approaches to remote learning, those that allow for real-time interactions between students and teachers during lessons, provide a mode of instruction most similar to in-person classroom instruction that can promote student engagement (Muñoz-Najar et al., 2021). These types of approaches are best used in online formats. However, in situations where synchronous delivery methods are not available (i.e., those without access to online resources), schools had to find alternative strategies to keep students engaged. Schools could use multiple communication channels to engage students and families in the learning process. For example, in Peru, it was hypothesized that communication and feedback were key to ensure student engagement in learning (Muñoz-Najar et al., 2021). Teachers were encouraged to keep in touch with their students and students’ families using phone calls, text messages and social media (Accinelli, 2020). Frequent interaction between students and teachers resulted in moderate to high satisfaction rates for remote learning using asynchronous learning approaches (e.g., pre-recorded online lessons, television and radio broadcasts) (Muñoz-Najar et al., 2021).

As most students participated in their lessons from home, schools recognized the important partner parents/guardians could be in supporting student engagement in remote learning. Several organizations such as the Sesame Workshop and Lifelong Health made resources available to parents/guardians to support their child in learning from home Muñoz-Najar et al. (2021). Additionally, the majority (>50%) of teachers or teacher respondents in Denmark, Ethiopia, the Russian Federation, Slovenia, the United Arab Emirates, and Uzbekistan reported that the time they spent communicating with parents increased during the period of school closures (Chen et al., 2022).

Methods

Table 1 Sample size by country

Data

For this study, we use data from the Responses to Educational Disruption Survey (REDS) (Meinck et al., 2022). REDS utilized a two-stage stratified cluster sampling designFootnote 2 to gather national samples of students, teachers, and schools. The student target population was defined as “all students enrolled in the grade that represents eight years of schooling counting from the first year of ISCED level 1” (Meyer et al., 2022, p. 22). Teacher and school target populations were those that taught or served students from the student target population.Footnote 3 Questionnaires from REDS ask students, teachers, and principals to reflect on their experiences before, during, and after the “COVID-19 disruption”Footnote 4, defined as the time period in which “[most] schools in a [country] closed for [the majority of students] in the last school year” (Fraillon & Stancel-Piatak, 2022, p. 11). Data for REDS was collected during the end of 2020 or the first half of 2021, therefore the “last school year” referred to the school year that included the start of 2020, when most schools in REDS participating countries decided to close down (Meinck et al., 2022).

With the exception of Burkina Faso, our sample includes all countries that administered the student questionnaire in REDS: Denmark, Ethiopia, Kenya, the Russian Federation, Slovenia, United Arab Emirates, and Uzbekistan. We removed students from our sample that responded that they did not do any schoolwork during the COVID-19 disruption; if the response was missing, we imputed the value using the modal response of students in that school, except in the case of Slovenia where the question was not administered. We dropped Burkina Faso from the analysis, since the number of students in the final sample was particularly low (16%, N=389 of the original 2474), meaning that very few students reported participating in remote learning. In REDS, several issues led to some countries not reaching the data quality standards. Denmark had a particularly low response rate (38%), while Ethiopia and Kenya were not able to collect fully random samples within schools and lacked information on the target population. Therefore, analyses with Denmark, Ethiopia, and Kenya are not considered representative of their populations and do not use sampling weights. The remaining countries—the Russian Federation, Slovenia, the United Arab Emirates, and Uzbekistan—present usual levels of data quality, with minor caveats (see Meinck et al., 2022). In addition, for our analyses we only worked with complete cases of Remote Learning Quality (RLQ) scales (see below), so analyses are comparable between scales. The original sample N, the filtered sample N and the analytical sample N for each country can be found in Table 1.

Measures of remote learning quality

Factor analysis scales

All the measures of remote learning quality indicators were built by us, based on the original responses available in the REDS dataset. For this, we identified several items from REDS that map onto essential components for remote learning quality (see Table 8). We developed a measure of students’ perspectives on remote learning quality (RLQ), so we only took items from the student questionnaire. We only report measures included in the final analyses. Our exploratory factor analyses found suitability for several constructs of remote learning quality that fit within the World Bank conceptual framework. Among these items, we used the following scales:

  • Effective teachers: Scale of perception of support provided by teachers during disruption.

  • Self-efficacy in technology use: Scale of self-reported skills on using learning-related technology, as a measure of access to skills to use technology effectively.

  • Engaged students: Scale of students’ experiences completing schoolwork remotely.

Within each scale, we applied a single imputation on the original items using the MICE package in R (van Buuren & Groothuis-Oudshoorn, 2011). We imputed those cases only where the student had a valid response in at least half of the items within a scale. If the student presented valid response in less than half of the items, the scale was marked as missing. For the predictive model we used the respective scale items and sociodemographic variables, specifically: gender, socioeconomic status, urbanicity, and language at home. The original N of complete cases, the number of cases imputed and the N of complete cases after-imputation can be found in Table 9.

For each of the three scales, we estimated scores based on a one-factor confirmatory factor analysis (CFA). We re-scaled the CFA scores to an international mean of 0 with a standard deviation of 1 using senate weights (i.e., giving the same weight to each country for the mean estimation). Using multi-group CFA, we tested the measurement invariance of these scales across countries. All CFA and measurement invariance estimations were done using the lavaan package in R (Rosseel, 2012), while exploratory analysis was done using Stata 16. The reliability (Cronbach’s Alpha) and measurement invariance models fit are available in Table 10. Effective Teachers and Engaged Students present scalar invariance between countries, meaning that the scales are comparable between countries. Self-efficacy in Technology Use present only metric invariance between countries, meaning that we cannot compare differences on latent means between countries. Self-reported measures of skills are prone to bias, as some students may undervalue their skills and others may overvalue them (Litt, 2013); it may be the case that students in some countries systematically underrate or overrate their skills, explaining the lack of scalar invariance.

Dichotomous indicator

In addition, on the dimension of Available and Suitable Technology, we constructed a dichotomous indicator of Access to Quality Technology; students with a positive response on this indicator means that they declared to have access to a device and that it worked well most of the time. Table 2 presents summary statistics of these Remote Learning Quality scales.

Table 2 Average remote learning quality scales and indicator by country

Measures of school-provided supports/resources for remote learning

The REDS school questionnaire includes several items that ask principals about the types of supports or resources that they provided to students, parents/guardians, and teachers during the disruption period. We identify items regarding the supports or resources that were tied specifically to the key components of successful remote learning: Access to Quality Technology, Effective Teachers, and Student Engagement. In our review, we did not locate any items that we believed to be related with improving Self-efficacy in Technology Use.

Each of the items was coded into a binary indicator specifying whether the school did (“Yes”) or did not (“No”) provide the support or resource. Details on how each of the items was coded can be found in Table 11. The support we found related with Access to Quality Technology was whether the school distributed suitable technology for remote learning access. That is, whether the school provided internet or devices to students during the disruption period. In the case of Ethiopia and Kenya, where access to technology was generally low (10% and 12%, respectively), we checked whether the school made paper materials available to students or supported students in accessing broadcast lessons over the television or radio.

We identify several items that ask whether schools increased resources to support teachers in implementing remote teaching. These resources include online teaching tools (e.g., online platforms, remote teaching tools), professional learning (e.g., professional development, resources for effective remote teaching pedagogy), and peer collaboration opportunities. Each item was coded so that it focuses only on schools that responded that they “Increased” resources. Therefore, schools that had already been investing in these types of resources prior to the disruption would not be included among the schools that we indicate as providing the support. Finally, we select several items that ask schools whether they provided supports for Student Engagement. The first set of items describe supports that were provided to encourage live interactions between students and teachers (e.g., offering live virtual lessons or support sessions). We also examine whether schools expected their teachers to provide students feedback on their work during the disruption period. Finally, we check whether schools provided any advice and/or materials to parents/guardians to support their child’s home learning environment.

In discussing school-provided supports or resources, we believe it is important to note that schools may have been constrained in their ability to provide certain resources for their students, teachers, and communities based on the priorities of higher level governments in their educational systems. For example, some schools may have only been able to provide technology for learning to their students as a result of Ministries of Education or local governments providing schools with the funds or resources to provide these services. Therefore, we note the importance of other actors in schools’ ability to provide these supports/resources for remote learning.

Sociodemographic variables

We also use several student and school background variables in our analysis. First, we use a measure of students’ socioeconomic status (SES). The scale, available in the REDS dataset, was constructed using information about the number of learning resources (e.g., a quiet place to work, ICT devices) and books in the home, parents’ level of education and occupation, and language spoken at home (UNESCO & IEA, 2022). The scale is divided into three categories based on tertiles on the continuous scale across countries: high SES (top tertile), medium SES (middle tertile), and low SES (bottom tertile). Second, we use information about where the school is located. Specifically, we distinguish between schools in a “Big town or city” (located in a town with \(\ge\) 15,000 people) and those in a small town (located in town with < 15,000 people). This was a similar way to how school locations were categorized in Strietholt and Süttmann (2022). Throughout the report we also use “urban” and “rural” to distinguish between schools located in a big town or city and a small town, respectively. Third, we look at student-reported gender (i.e., Boy, Girl).Footnote 5 Finally, we look at student home language and whether it differs from the language of instruction. Specifically, students were asked “What language do you speak at home most of the time?” In relation to this item’s responses, it is important to note that, in Kenya, the language of instruction is English for most students, yet this often does not match with the language most often spoken in the catchment area (Dexis Consulting Group, 2021). Hence, there is a large proportion of students speaking a language other than the language of instruction. This is important for interpretation of results by language in Kenya. Table 3 shows a summary of these variables for our analytic sample by country.

Table 3 Overview statistics of students by country

Analytic approach

We follow the suggested procedures for analyzing data from the REDS international database (UNESCO & IEA, 2022). That is, we apply probability weights and estimate variance using the jackknife repeated replication (JRR) method, where possible. JRR incorporates the sampling variance within the standard error estimations; this step is needed due the complex sampling and the clustering of students within schools (González & Foy, 2009). Due to some countries not reaching data quality standards, the use of weights and JRR was not possible in three of our seven selected countries (Denmark, Ethiopia, and Kenya). In regression models, standard errors for these countries have been calculated accounting for clustering of observations at the school level (Rogers, 1994). However, weights have not been applied so results should not be interpreted as representative of the target population.

To explore inequalities in perceived remote learning quality within countries, we test differences in our RLQ scales by several student characteristics: socioeconomic status, urbanicity, gender, and home language. We conduct these tests in a regression framework. Specifically, we estimate, separately for each country, regression models of the following form:

$$\begin{aligned} RLQ_{i} = \alpha + \beta Grp_{i} + \varepsilon _i \end{aligned},$$
(1)

where \(RLQ_{i}\) is one of our measures of perceived remote learning quality for student i and \(Grp_{i}\) is some categorical variable representing a characteristics of student i (i.e., girl vs. boy, high SES vs. low SES, speaks language of instruction vs. other language at home) or student i’s school (located in an big town or city vs. small town). The coefficient \(\beta\) tests the difference in average perceived remote learning quality across student groups within a country. A significant coefficient would be indicative of inequalities in perceived remote learning quality within a country.

All analyses were done using Stata 16, taking advantage of its support of complex samples on survey analyses (i.e., jackknife).

Results

Across-country inequalities in student perceptions of remote learning quality

We begin our exploration by examining cross-country differences in our remote learning quality (RLQ) measures. In interpreting the results, it is important to note that data from Denmark, Ethiopia, and Kenya may not be representative of the target population and only reflect findings for student respondents on the REDS questionnaire. Furthermore, given a lack of scalar invariance in our Self-Efficacy in Technology Use scale, we do not present means for this scale in Fig. 1 as they cannot be compared across countries.

Fig. 1
figure 1

Cross-country Comparisons of Remote Learning Quality Scales and Indicator.

aData may not be representative of target population. Notes: 95% confidence intervals are presented as error bars around mean estimates. The Access to Quality Technology indicator (top panel) shows the share of students with a positive value on the measure. Each scale (second and third panel) has an international average of 0 (represented by the horizontal line) and a standard deviation of 1 after applying senate weights. Probability weights have been applied and jackknife standard errors have been calculated in all countries except in Denmark, Ethiopia, and Kenya. Standard errors calculated for Denmark, Ethiopia, and Kenya account for clustering of observations within schools. Due to lack of scalar invariance, averages on the Self-Efficacy in Technology Use scale are not presented. DNK = Denmark, ETH = Ethiopia, KEN = Kenya, RUS = Russian Federation, SVN = Slovenia, ARE = United Arab Emirates, UZB = Uzbekistan

A key component for successful remote learning is giving students access to the devices or content needed to continue learning outside of a school setting. The Access to Quality Technology indicator captures the share of students who report that they had both a good quality device (e.g., a laptop or tablet) as well as reliable internet access. The top panel of Fig. 1 shows the summary of the Access to Quality Technology indicator across countries. Unlike the scales, which are continuous, the mean values should be interpreted as the share of students responding that they had access to quality technology for remote learning. While most countries had at least three quarters of students or student respondents saying that they had access to quality devices, both Ethiopia and Kenya stood out, with only 10% and 12% of student respondents having access to quality devices or internet, respectively. It should be noted that, in results not shown here, despite these low levels of access, many students respondents in Ethiopia and Kenya reported that schools provided other means for them to access content, either through paper-based worksheets or lessons broadcast on the television or radio (84% and 92%, respectively). So while it appears that Ethiopia and Kenya did not have the technology to provide remote learning online, other methods were used to reach their students.

In addition to having access to technology, students also need effective teachers that can support them in their learning in a remote setting. The second panel of Fig. 1 compares means across countries on the Effective Teachers scale. The horizontal line in this panel represents the international average for the included countries. Uzbekistan and the United Arab Emirates had the highest values on the scale (0.65 and 0.43, respectively), indicating that students in these countries felt more supported by their teacher during the disruption than students in other countries. Ethiopia and Kenya stand out with particularly low values on the Effective Teachers scale (\(-0.29\) and \(-0.78\), respectively).

As well as access to suitable technology and effective teachers, students need to be engaged during remote learning in order for it to be successful. The bottom right panel of Fig. 1 compares means on the Engaged Students scale across countries. The Russian Federation, Slovenia, the United Arab Emirates, and Uzbekistan all had averages significantly above the international average on this measure of student engagement. Ethiopia and Kenya had the lowest average values on the scale, − 0.73 and − 0.83, respectively.

Within-country inequalities in student perceptions of remote learning quality

Table 4 Inequalities in self-efficacy in technology use scale

In this part of the results, we explore within-country inequalities in the remote learning quality measures across several different sociodemographic groups. Table 4 displays estimated differences in the Self-efficacy in Technology Use scale across several groupings, within each country. Inequalities between medium or high SES and low SES students in self-efficacy in technology use is evident across six of the seven countries. The scale is standardized so the size of the differences can be interpreted in standard deviations. The inequalities range in size from 0.21 to 1.47 standard deviations on the scale. The largest gaps appear for responding students in Ethiopia (0.98 difference between high and low SES, 0.53 difference between medium and low SES) and Kenya (1.47 difference between high and low SES, 0.40 difference between medium and low SES). The gaps indicate that student confidence in their ability to use technology is significantly higher for students in families with more resources. In contrast, in Denmark and Slovenia, we find no significant differences between medium or high SES students and low SES students. That would indicate that student respondents across Denmark and Slovenia tended to be confident in using technology for learning, regardless of their home resources.

In comparing the Self-efficacy in Technology Use scale between students attending school in a big town or city versus a school in a small town, we observe significant differences in favor of students in more urban schools for four of the seven countries. The size of the inequalities ranged between 0.20 to 0.57 standard deviations. The largest differences appear in Ethiopia and Kenya. While most student respondents were in schools that reported being located in a small town, it appears that the few students living in more urban areas had greater confidence in their abilities to use technology entering the pandemic (see Table 3). In contrast, Denmark, Slovenia, and Uzbekistan did not show results that indicated inequalities in the Self-efficacy in Technology Use scale between urban and rural schools.

Girls tended to report lower levels of confidence in their ability to use technology than boys across all countries. The size of the inequality ranged between 0.12 and 0.37 standard deviations. The largest differences are found in Denmark and Slovenia, where boys, on average, score about 0.37 and 0.36 points higher than girls, respectively, on the Self-efficacy in Technology Use scale.

In exploring differences by students’ home language, we find significant gaps in three of the seven countries. The largest differences show that students who speak a different language than the language of instruction at home tended to have lower levels of confidence in using technology. These gaps were largest and significant in Kenya and the Russian Federation (\(-0.26\) and \(-0.24\), respectively). However it is important to recall that the majority of student respondents in Kenya (93%) reported that they spoke a language other than the language of instruction (English) at home showing that the few students who did speak English at home had higher confidence than those who did not. The opposite pattern was found in the United Arab Emirates, where students speaking the language of instruction scored 0.21 points lower on the Self-efficacy in Technology Use scale than their counterparts. The difference was not significant in the other countries.

Table 5 Inequalities in access to quality technology indicator

We next examine inequalities in access to quality technology within each country. Table 5 shows the estimated differences in the Access to Quality Technology indicator across student and school groups. When comparing high or medium SES students with low SES students, inequalities were found in five of the seven countries. In the five countries, high or medium SES students were significantly more likely to respond to having a quality device and reliable internet access (an exception was in the United Arab Emirates where there was no significant difference between medium SES and low SES students). Student respondents in Kenya and Ethiopia reported having generally low levels of quality technology access compared to the other countries (10% and 12%, respectively) and the observed gap may reflect limited resources. That is, schools in these countries may not have had the ability to provide nearly 90% of their students with a good quality device or internet coverage. However, in results not shown here, we still observe inequalities in student access to other types of remote learning materials (e.g., paper-based assignments or television/radio broadcast lessons) in Ethiopia and Kenya. Specifically, high and medium SES students were significantly more likely to have access to these alternative resources for remote learning compared to low SES students (with one exception in Kenya, where high SES students did not have significantly higher access to these materials compared to low SES students). However, higher levels of technology access did not necessarily mean that all students had access as illustrated by the observed inequalities in the Russian Federation and Uzbekistan. In contrast, however, Denmark and Slovenia, also with high levels of technology access, showed no significant gaps in technology access across socioeconomic levels.

Schools located in a big town or city had higher shares of students reporting having access to quality device in the United Arab Emirates and Uzbekistan. However, it could be argued that there were large differences across school locations in Ethiopia, Kenya, and the Russian Federation (although the findings are not significant at the 5% level, only at the 10%). Findings may be a result of generally poor internet coverage in rural areas (ITU, 2020b; 2021b). If these countries have the resources to provide technology to students, the findings indicate that it would be worth investing in schools located in small towns as they were less likely to have students with good quality devices or internet coverage.

When comparing the Access to Quality Technology indicator by gender and home language, relatively fewer inequalities were observed. One exception was in Slovenia, where a higher share of girls had access to technology than boys did. Another exception was in Kenya where the share of students speaking a language other than the language of instruction at home (the majority) was 17 percentage points lower than their counterparts. That is, those who spoke English at home in Kenya were more likely to have access to quality technology as well as have confidence in their ability to use technology (see Table 4).

Table 6 Inequalities in effective teachers scale

Inequalities in student perceptions of their teachers’ effectiveness during the disruption period are examined in Table 6. Across the socioeconomic status, school urbanicity, gender, and home language groupings, most countries do not show significant inequalities in the perceived effectiveness of teachers. That is, students had the same level of perceived support from their teachers across most countries, regardless of their socioeconomic background, gender, language status, or school location. Exceptions to this pattern were Kenya and Uzbekistan when comparing differences across different SES levels. In these countries, high or medium SES students had higher values on the Effective Teachers scale compared to low SES students. Kenya was the exception when examining differences by school location, having higher perceptions of teacher effectiveness in schools located in a big town or city. Denmark, the Russian Federation, and Uzbekistan all showed inequalities by gender, with boys having higher scores on the Effective Teachers scale than girls in Denmark and the Russian Federation and the opposite being true in Uzbekistan. Finally, Ethiopia and the United Arab Emirates had significant inequalities across home language groups. In the United Arab Emirates, students speaking the language of instruction had higher perceptions of their teacher’s effectiveness than their counterparts. The opposite pattern was observed in Ethiopia. Despite the observed inequalities in the Self-Efficacy in Technology Use scale, it appears the perceptions of effective teaching were mostly the same within countries with the noted exceptions.

Table 7 Inequalities in student engagement scale

We finally explore within-country inequalities in student engagement. Table 7 presents tests for inequalities in the Student Engagement scale. In comparing high or medium SES with low SES students, we find significant differences in four out of seven countries. In those four countries, high or medium SES students tended to have higher values on the Student Engagement scale compared to low SES students. Inequalities ranged from 0.17 to 1.12 standard deviations. The largest differences were observed in Ethiopia and Kenya, where high SES student respondents had averages on the Student Engagement scale 0.91 and 1.12 standard deviations higher than low SES students, respectively (or 0.64 and 0.56 standard deviation differences when comparing medium SES to low SES students, respectively). Further differences in student engagement were found when comparing by gender. Three out of seven countries had significant differences between boys and girls. In Denmark and Slovenia, boys had higher values on the Student Engagement scales than girls. The opposite was true in Ethiopia, where girls had higher engagement measures than boys.

Relatively fewer differences were found when comparing across home language and school location. When comparing across home language groups, student respondents in Ethiopia and Kenya had significantly lower reported levels of student engagement among students speaking a language other than the language of instruction. Finally, schools located in a big town or city had students with higher levels of engagement than those in schools located in a small town. However, these were exceptions. Most differences were not found to be significant when comparing across these student groupings.

School-provided supports/resources for remote learning

We finally examine the extent to which resources were provided by schools to students, teachers, or parents/guardians in order to support remote learning implementation. The first row of Fig. 2 shows the share of students in schools that provided at least some students with access to remote learning. That is, these are schools that provided laptops and internet access to at least some of their students. In Ethiopia and Kenya, given the low levels of technology access, we look at schools who made paper-based materials or television/radio broadcast lessons available. While near to over half of students were in schools that reported providing some form of suitable technology during the disruption period in five of the seven countries, the distribution of technology is not a necessary component to ensure access. For instance, in a country like Denmark, over 90% of students respondents indicated that they had a quality device and reliable internet access. It would not be necessary for schools to provide devices or internet to students when most students already had access to a device. The distribution of suitable technology would be most appropriate in areas where there are already low levels of access or inequalities are present so that schools can target their resources to students most in need. Therefore, it is promising to see so many schools in Ethiopia making materials available to students so that they are able to continue to access lessons outside the classroom. Furthermore, we highlight the relatively high level of distribution that occurred in Slovenia’s schools (71.6%). In comparing with our other findings, data from Slovenia did not reveal any significant inequalities in our Access to Quality Technology indicator. However, the same pattern did not appear in the data for Uzbekistan, where the high use of this strategy (69.7%) was not paired with findings of no inequalities in our technology access indicator.

Fig. 2
figure 2

Cross-country Comparisons of School Supports and Resources for Remote Learning.

aData may not be representative of target population. Notes: Standard errors are presented in parentheses. Probability weights have been applied and jackknife standard errors are presented for all countries except in Denmark, Ethiopia, and Kenya. Standard errors reported for Denmark, Ethiopia, and Kenya account for clustering of observations within schools. DNK = Denmark, ETH = Ethiopia, KEN = Kenya, RUS = Russian Federation, SVN = Slovenia, ARE = United Arab Emirates, UZB = Uzbekistan

The next section of Fig. 2 shows whether resources were increased to support teachers during the disruption period. Given the wording of these questions, it is important to note that some schools might have provided sufficient resources for these teacher supports prior to the disruption period. Therefore, a school not increasing the use of these supports during the disruption period does not necessarily mean that it was ignored, it just highlights the schools where extra efforts were made to focus on such practices. The findings show a good range of variation in whether schools increased or did not increase such supports. Schools in Slovenia and the United Arab Emirates generally responded that they increased the access and level of specific supports or resources for teachers in their schools (above 50% across the listed teacher supports). When comparing this with the inequalities in the Effective Teachers scale, few inequalities were found in Slovenia and the United Arab Emirates (the only exception being students who speak another language than the language of instruction at home had significantly lower perceptions of their teachers’ effectiveness in the United Arab Emirates).

The final section of Fig. 2 shows the use of several strategies used to promote student engagement. The first set of strategies, offering live lessons or support, tended to be used in most countries where online remote learning was possible (i.e., all countries except in Ethiopia and Kenya where there were very low levels of technology access). Expecting teachers to provide feedback to students was generally found to be the case in all countries except Kenya (all above 85%). Finally, schools also provided resources to parents/guardians to help them in supporting their child with remote learning. The majority of students across all countries were in schools that provided this support to parents. One exception was in Kenya, where only 44% of student respondents were in schools that reported that they provided teaching materials and worksheets to parents and guardians and only 57% of student respondents were in schools that reported that they provided advice to parents/guardians on how to help with specific aspects of schoolwork.

Additional analyses on the use of school-provided supports/resources for remote learning

We performed additional analyses estimating inequalities in the use of these supports within countries. However, results did not reveal many significant or meaningful differences. In addition, regression analysis was done to test the association between a school’s use of these supports and our measures of remote learning quality. However, we did not find many significant associations. Given the lack of significant results and the fact that we believe the analysis was outside of the scope of the current study, we do not report these results.

Discussion and conclusion

Using student perceptions of their remote learning experience, we develop several measures of remote learning quality proposed by the World Bank to assess remote learning success: Self-efficacy in Technology Use, Access to Quality Technology, Effective Teachers, and Engaged Students (Muñoz-Najar et al., 2021). Findings reveal several inequalities across countries in our sample. Specifically, in the African countries, Ethiopia and Kenya, student respondents reported much lower levels of access to quality technology needed to access online lessons. Not only that, but a considerably low proportion of student respondents in the three African countries that participated in REDS reported that they did not participate in remote learning during the disruption (see Table 1), especially in Ethiopia and Burkina Faso (which had to be dropped from the analysis because of this). This result already indicates some disparities across countries in the access to remote learning, reflecting the challenges in the implementation of remote learning in some countries. These results show that not all countries were prepared to offer online instruction to students. However, we also find that Ethiopia and Kenya found other ways to provide access to learning to their students (i.e., by providing paper-based worksheets and learning materials or broadcasting lessons over the television or radio). Despite this, Ethiopia and Kenya, still had relatively lower scores on our Effective Teachers and Engaged Students scales. While this might be concerning, it likely reveals the advantages of online synchronous instruction that allows teachers to provide more tailored and personal support to students and gives more opportunities to engage students in lessons. Unfortunately, it seems from our findings, that countries where students were extremely limited in their access to technology were unable to offer the type of remote learning experience that would engage students and showcase the effectiveness of teachers. However, we also find that having access to technology was not sufficient on its own to ensure more effective teachers or higher student engagement as illustrated in data from Denmark.

Within countries, our findings reveal that, in many of the countries examined, inequalities in student access to and confidence in using technology presented challenges in implementing remote learning for certain groups of students. In line with the literature on the digital divide, these groups were, typically, low-SES students, girlsFootnote 6, or students living in more rural areas. This was the case even in some countries that had high overall levels of access to technology (e.g., the Russian Federation, the United Arab Emirates, and Uzbekistan). These findings indicate that, even in countries that may have high levels of access to learning technologies, there were still some groups of students that could use additional support in gaining access to technology as well as help on how to use it. Targeted distribution could be one way to address these concerns of within-country inequalities. That is, schools could identify student technological needs and make devices and internet coverage available to them as well as resources and help on how to use them. For this to be most effective, baseline measures in technology access or technological skills would assist in identifying students most in need of support.

These inequalities in countries’ ability to offer online remote learning may have carried over into inequalities in student engagement. As referenced earlier, synchronous activities (usually facilitated by online interaction) were one mechanism to keep students engaged even outside the classroom (Muñoz-Najar et al., 2021). A lot of the inequalities in student engagement found within countries were associated with inequalities in access to quality technology, illustrating a possible consequence of the digital divide in educational systems forced to transition learning into a remote setting.

Despite the challenges observed within countries, student perceptions of the effectiveness of their teachers, generally, did not differ significantly across these groups (with some exceptions noted in the Results section). Taken together, these findings imply that, within-countries, schools were not hindered (nor helped) in their efforts to prepare teachers for managing classrooms remotely by limitations in technology access or use. Within countries, students with lower levels of access and confidence in their skills to use technology, perceived their teachers in similar ways to their counterparts. Whether this is because of the additional effort made by teachers in these countries to ensure the continuity of learning in a fair and equitable way or because students, who are closer to teachers, more recognized their efforts, rather than the supports provided at higher levels (i.e., school or institutional policies) cannot be answered from our analysis. We feel that future research could further explore the implications of this finding.

The relatively lower number of inequalities in our Effective Teachers measure may be related to the efforts that schools made to support their students’ remote learning experience. In our review of several school-provided supports or resources for remote learning, we find that increased efforts to support teachers evident in Slovenia and the United Arab Emirates was paired with fewer inequalities in these measures. In addition, many schools implemented supports aimed at improving student engagement (e.g., live virtual lessons and support, expectations of teachers to provide feedback to students, and materials and advice to parents/guardians to improve the home learning environment). Unfortunately, it does not appear that these supports were able to overcome the challenges of inequalities in technology use and access. Schools, with little time to prepare for the transition to remote learning, appear to have made significant efforts to support their students, even if their students had difficulty accessing online content.

It was clear that the learning experiences of students globally would be impacted by the pandemic. Schools were forced to close and educational leaders were confronted with several challenges in how they would maintain learning continuity with students not able to attend class in-person. In response, many schools implemented remote learning strategies. Given differences in schools’ preparedness to offer such remote learning options, many anticipated the transition to remote learning to exacerbate existing educational inequalities. Our findings partly support this hypothesis. Existing inequalities in access to and confidence in using technology across and within countries likely significantly altered the learning experience of certain students in a negative way, keeping them less engaged in the learning material. However, school efforts in some countries to prepare teachers to support students in a remote setting might have reduced some of the inequalities that might have been expected. The resilience of schools in responding to the COVID-19 disruption possibly limited some of the negative consequences to students, regardless of their background.

Availability of data and materials

The datasets supporting the conclusions of this article are openly accessible and can be downloaded as public use files from the IEA’s website: https://www.iea.nl/data-tools/repository/reds.

Notes

  1. We also acknowledge the vital role that self-regulation skills play in keeping students engaged and motivated to learn while outside the physical classroom without the supervision of the teacher (Fredricks et al., 2004). Furthermore, learning routines taught in the classroom likely supported students in continuing their learning during asynchronous sections (Kauffman, 2015).

  2. Schools were the first sampling stage and students and teachers were the second sampling stage.

  3. More details on the REDS sampling design and methodology can be found in Meyer et al. (2022).

  4. This time period is also referred to as the “reference period” or “disruption period” in Meinck et al. (2022). We use these terms interchangeably throughout this paper.

  5. The option for “Other” in the gender question was administered in only two countries in our sample (Denmark and Kenya). In these two countries, only a small proportion of students chose this category with the highest being 4% in Denmark. Because of this, we do not report findings for this group in the results.

  6. In regards to gender differences, it is important to note that results from the International Computer and Information Literacy Study (ICILS) 2013 show that girls tend to report lower levels of self-efficacy for advanced or specialized ICT tasks, even when girls outperform boys in the computer and information literacy test (Fraillon et al., 2014). In contrast, REDS only asked students about basic ICT-related tasks, not advanced or specialized tasks. ICILS 2013 did not find a significant gender gap when examining basic ICT self-efficacy measures; however, in our analysis, we do find significant gender differences in multiple countries.

References

  • Accinelli, A. (2020). Going the distance. Peru’s two-week remote education response to COVID-19.

  • Alrashidi, O., Phan, H. P., & Ngu, B. H. (2016). Academic engagement: An overview of its definitions, dimensions, and major conceptualisations. International Education Studies, 9(12), 41. https://doi.org/10.5539/ies.v9n12p41.

    Article  Google Scholar 

  • American Institutes for Research. (2021). Promising Practices for Addressing Common Challenges: Remote Instruction Practices (Tech. Rep.).

  • Barron Rodriguez, M., Cobo, C., Munoz-Najar, A., & Sánchez Ciarrusta, I. (2021). Remote learning during the global school lockdown: Multi-country lessons. World Bank. https://doi.org/10.1596/36141.

    Article  Google Scholar 

  • Bergem, O. K., Nilsen, T., Mittal, O., & Rader, H. G. (2020). Can teachers’ instruction increase low-SES students’ motivation to learn mathematics? In T. S. Frones, A. Pettersen, J. Radišić, & N. Buchholtz (Eds.), Equity, Equality and Diversity in the Nordic Model of Education (pp. 251–272). Cham: Springer International Publishing. https://doi.org/10.1007/978-3-030-61648-9_10

  • Berkowitz, R., Moore, H., Astor, R. A., & Benbenishty, R. (2017). A research synthesis of the associations between socioeconomic background, inequality, school climate, and academic achievement. Review of Educational Research, 87(2), 425–469. https://doi.org/10.3102/0034654316669821.

    Article  Google Scholar 

  • Blömeke, S., Olsen, R. V., & Suhl, U. (2016). Relation of Student Achievement to the Quality of Their Teachers and Instructional Quality. In T. Nilsen & J.-E. Gustafsson (Eds.), Teacher Quality, Instructional Quality and Student Outcomes (vol. 2, pp. 21–50). Cham: Springer International Publishing. https://doi.org/10.1007/978-3-319-41252-8_2

  • Chen, M., Kennedy, A. I., Meinck, S., & Rožman, M. (2022). Communication, feedback, and assessment. In S. Meinck, J. Fraillon, & R. Strietholt (Eds.), The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (pp. 84–106). 7, place de Fontenoy, 75007 Paris, France and 311 Keizersgracht 1016 EE Amsterdam Netherlands: UNESCO and IEA. https://doi.org/10.1007/978-1-4614-2018-7_13

  • Creemers, B., & Kyriakides, L. (2008). The dynamics of educational effectiveness: A contribution to policy, practice and theory in contemporary schools. Routledge.

  • Dexis Consulting Group. (2021). Language of Instruction Country Profile: Kenya (Tech. Rep.). Retrieved from https://pdf.usaid.gov/pdf docs/PA00XH25.pdf

  • Fauth, B., Decristan, J., Rieser, S., Klieme, E., & Büttner, G. (2014). Student ratings of teaching quality in primary school: Dimensions and prediction of student outcomes. Learning and Instruction, 29, 1–9. https://doi.org/10.1016/j.learninstruc.2013.07.001.

    Article  Google Scholar 

  • Fraillon, J., Ainley, J., Schulz, W., Friedman, T., & Duckworth, D. (2020). Preparing for Life in a Digital World: IEA International Computer and Information Literacy Study 2018 International Report. Cham: Springer International Publishing. https://doi.org/10.1007/978-3-030-38781-5.

    Book  Google Scholar 

  • Fraillon, J., Ainley, J., Schulz, W., Friedman, T., & Gebhardt, E. (2014). Preparing for Life in a Digital Age: The IEA International Computer and Information Literacy Study International Report. Springer Cham. https://doi.org/10.1007/978-3-319-14222-7

  • Fraillon, J., & Stancel-Piatak, A. (2022). The conceptual background of reds. In S. Meinck, J. Fraillon, & R. Strietholt (Eds.), The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (pp. 7–18). 7, place de Fontenoy, 75007 Paris, France and 311 Keizersgracht 1016 EE Amsterdam Netherlands: UNESCO and IEA. https://doi.org/10.1007/978-1-4614-2018-7_13

  • Fredricks, J. A., Blumenfeld, P. C., & Paris, A. H. (2004). School engagement: Potential of the concept, state of the evidence. Review of Educational Research, 74(1), 59–109. https://doi.org/10.3102/00346543074001059.

    Article  Google Scholar 

  • Fredricks, J. A., & McColskey, W. (2012). The Measurement of Student Engagement: A Comparative Analysis of Various Methods and Student Self-report Instruments. In S. L. Christenson, A. L. Reschly, & C. Wylie (Eds.), Handbook of Research on Student Engagement (pp. 763–782). Boston, MA: Springer US. https://doi.org/10.1007/978-1-4614-2018-7_37

  • Goe, L. (2007). The Link Between Teacher Quality and Student Outcomes: A Research Synthesis. National comprehensive center for teacher quality. 76.

  • González, E. J., & Foy, P. (2000). Chapter 12. estimation of sampling variance. In M. O. Martin, K. D. Gregory, & S. E. Stemler (Eds.), Timss 1999 technical report. International Study Center, Lynch School of Educationm Boston College.

  • Hargittai, E. (2002). Second-level digital divide: Differences in people’s online skills. First Monday. https://doi.org/10.5210/fm.v7i4.942.

    Article  Google Scholar 

  • Hargittai, E. (2005). Survey measures of web-oriented digital literacy. Social Science Computer Review, 23(3), 371–379. https://doi.org/10.1177/0894439305275911.

    Article  Google Scholar 

  • Hattie, J. (2009). Visible learning: A synthesis of over 800 meta-analyses relating to achievement (Reprinted). London: Routledge.

    Google Scholar 

  • ITU. (2019). Measuring digital development: Facts and Figures 2019 (Tech. Rep.).

  • ITU. (2020a). Digital Skills Insights 2020.

  • ITU. (2020b). Measuring digital development: Facts and Figures 2020 (Tech. Rep.).

  • ITU. (2021a). Digital Skills Insights 2021. , 156.

  • ITU. (2021b). Measuring digital development—facts and figures 2021 (Tech. Rep.).

  • ITU, UNESCO, & UNICEF. (2020). The Digital Transformation of Education: Connecting Schools, Empowering Learners (Tech. Rep.).

  • Kauffman, H. (2015). A review of predictive factors of student success in and satisfaction with online learning. Research in Learning Technology, 23.

  • Klieme, E., Pauli, C., & Reusser, K. (2009). The Pythagoras study: Investigating effects of teaching and learning in Swiss and German mathematics classrooms. In The Power of Video Studies in Investigating Teaching and Learning in the Classroom (pp. 137–160). BoD-Books on Demand

  • Lam, S.-F., Jimerson, S., Shin, H., Cefai, C., Veiga, F. H., Hatzichristou, C., & Zollneritsch, J. (2016). Cultural universality and specificity of student engagement in school: The results of an international study from 12 countries. British Journal of Educational Psychology, 86(1), 137–153. https://doi.org/10.1111/bjep.12079.

    Article  Google Scholar 

  • Lietaert, S., Roorda, D., Laevers, F., Verschueren, K., & De Fraine, B. (2015). The gender gap in student engagement: The role of teachers’ autonomy support, structure, and involvement. British Journal of Educational Psychology, 85(4), 498–518. https://doi.org/10.1111/bjep.12095.

    Article  Google Scholar 

  • Litt, E. (2013). Measuring users’ internet skills: A review of past assessments and a look toward the future. New Media and Society, 15, 612–630. https://doi.org/10.1177/1461444813475424.

    Article  Google Scholar 

  • Meinck, S., Fraillon, J., & Strietholt, R. (2022). The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (Tech. Rep.).

  • Meyer, S., Penon, K., Whilsher Beyer, C., Meinck, S., & Waschk, A. (2022). Methods, procedures, and data. In S. Meinck, J. Fraillon, & R. Strietholt (Eds.), The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (pp. 34–201). 7, place de Fontenoy, 75007 Paris, France and 311 Keizersgracht 1016 EE Amsterdam Netherlands: UNESCO and IEA. https://doi.org/10.1007/978-1-4614-2018-7_13

  • Moore, K. A., Lippman, L. H., & Ryberg, R. (2015). Improving outcome measures other than achievement. AERA Open, 1(2), 2332858415579676. https://doi.org/10.1177/2332858415579676.

    Article  Google Scholar 

  • Muñoz-Najar, A., Gilberto, A., Hasan, A., Cobo, C., Azevedo, J. P., & Akmal, M. (2021). Remote learning during covid-19: Lessons from today, principles for tomorrow. World Bank, Washington, DChttps://doi.org/10.1596/36665

  • Nilsen, T., Gustafsson, J.-E., & Blömeke, S. (2016). Conceptual Framework and Methodology of This Report. In T. Nilsen & J.-E. Gustafsson (Eds.), Teacher Quality, Instructional Quality and Student Outcomes: Relationships Across Countries, Cohorts and Time (pp. 1–19). Cham: Springer International Publishing. https://doi.org/10.1007/978-3-319-41252-8_1

  • OECD. (2013). PISA 2012 Results: Ready to Learn (Volume III): Students’ Engagement, Drive and Self-Beliefs. Author. https://doi.org/10.1787/9789264201170-en

  • Quin, D. (2017). Longitudinal and contextual associations between teacher-student relationships and student engagement: A systematic review. Review of Educational Research, 87(2), 345–387. https://doi.org/10.3102/0034654316669434.

    Article  Google Scholar 

  • Rogers, W. (1994). Regression standard errors in clustered samples. Stata Technical Bulletin 3(13).

  • Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48(2), 1–36. https://doi.org/10.18637/jss.v048.i02.

    Article  Google Scholar 

  • Salas-Pilco, S. Z., Yang, Y., & Zhang, Z. (2022). Student engagement in online learning in Latin American higher education during the COVID-19 pandemic: A systematic review. British Journal of Educational Technology, 53(3), 593–619. https://doi.org/10.1111/bjet.13190.

    Article  Google Scholar 

  • Scheerder, A., & van Dijk, J. (2017). Determinants of Internet skills, uses and outcomes. A systematic review of the second- and third-level digital divide. Telematics and Informatics, 34(8), 1607–1624. https://doi.org/10.1016/j.tele.2017.07.007.

    Article  Google Scholar 

  • Scherer, R., & Nilsen, T. (2016). The Relations Among School Climate, Instructional Quality, and Achievement Motivation in Mathematics. In T. Nilsen & J.-E. Gustafsson (Eds.), Teacher Quality, Instructional Quality and Student Outcomes: Relationships Across Countries, Cohorts and Time (pp. 51–80). Cham: Springer International Publishing. https://doi.org/10.1007/978-3-319-41252-8_3

  • Schunk, D. H., & Mullen, C. A. (2013). Motivation. In J. Hattie & E. M. Anderman (Eds.), International Guide to Student Achievement (pp. 67–69). Routledge. https://doi.org/10.4324/9780203850398-24

  • Skinner, E. A., Kindermann, T. A., & Furrer, C. J. (2009). A motivational perspective on engagement and disaffection: Conceptualization and assessment of children’s behavioral and emotional participation in academic activities in the classroom. Educational and Psychological Measurement, 69(3), 493–525. https://doi.org/10.1177/0013164408323233.

    Article  Google Scholar 

  • Stancel-Piatak, A., Franck, E., & Kennedy, A. I. (2022). International findings. In S. Meinck, J. Fraillon, & R. Strietholt (Eds.), The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (pp. 34–201). 7, place de Fontenoy, 75007 Paris, France and 311 Keizersgracht 1016 EE Amsterdam Netherlands: UNESCO and IEA. https://doi.org/10.1007/978-1-4614-2018-7_13

  • Strietholt, R., & Süttmann, F. (2022). Inequalities in teaching and learning during the pandemic. In S. Meinck, J. Fraillon, & R. Strietholt (Eds.), The impact of the covid-19 pandemic on education: International evidence from the responses to educational disruption survey (reds) (pp. 184–201). 7, place de Fontenoy, 75007 Paris, France and 311 Keizersgracht 1016 EE Amsterdam Netherlands: UNESCO and IEA. https://doi.org/10.1007/978-1-4614-2018-7_13

  • Subramainan, L., & Mahmoud, M. A. (2020). A systematic review on students’ engagement in classroom: indicators, challenges and computational techniques. International Journal of Advanced Computer Science and Applications, 11(1). https://doi.org/10.14569/IJACSA.2020.0110113

  • UNESCO. (2020). School closures caused by coronavirus (COVID-19). Retrieved March 10, 2022, from https://en.unesco.org/covid19/educationresponse

  • UNESCO, & IEA. (2022). Responses to educational disruption survey: User guide for the international database. Retrieved from https://unesdoc.unesco.org/ark:/48223/pf0000380397/PDF/380397eng.pdf.multi

  • UNESCO, UNICEF, the World Bank, & OECD. (2021). What’s Next? Lessons on Education Recovery: Findings from a Survey of Ministries of Education amid the COVID-19 pandemic (Tech. Rep.). Paris, New York, Washington D.C.: UNESCO, UNICEF, World Bank.

  • van Buuren, S., & Groothuis-Oudshoorn, K. (2011). Mice: Multivariate imputation by chained equations in r. Journal of Statistical Software, 45(3), 1–67. https://doi.org/10.18637/jss.v045.i03.

    Article  Google Scholar 

  • van Dijk, J. (2005). The Deepening Divide: Inequality in the Information Society. Thousand Oaks, California. https://doi.org/10.4135/9781452229812

  • van Dijk, J. (2006). Digital divide research, achievements and shortcomings. Poetics, 34(4–5), 221–235. https://doi.org/10.1016/j.poetic.2006.05.004.

    Article  Google Scholar 

  • Vangrieken, K., Dochy, F., Raes, E., & Kyndt, E. (2015). Teacher collaboration: A systematic review. Educational Research Review, 15, 17–40. https://doi.org/10.1016/j.edurev.2015.04.002.

    Article  Google Scholar 

  • Wang, M.-T., & Eccles, J. S. (2012). Social support matters: Longitudinal effects of social support on three dimensions of school engagement from middle to high school. Child Development, 83(3), 877–895. https://doi.org/10.1111/j.1467-8624.2012.01745.x.

    Article  Google Scholar 

  • Wei, K.-K., Teo, H.-H., Chan, H. C., & Tan, B. C. Y. (2011). Conceptualizing and testing a social cognitive model of the digital divide. Information Systems Research, 22(1), 170–187. https://doi.org/10.1287/isre.1090.0273.

    Article  Google Scholar 

  • Zhong, Z.-J. (2011). From access to usage: The divide of self-reported digital skills among adolescents. Computers and Education, 56(3), 736–746. https://doi.org/10.1016/j.compedu.2010.10.016.

    Article  Google Scholar 

Download references

Acknowledgements

We would like to thank Dr. Sabine Meinck for providing comments on an early draft of this manuscript, Dr. Andrés Christiansen for his helpful advice and consultation on scaling, and two anonymous reviewers for their feedback on this paper.

Funding

No funding was used in support of this paper.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the conceptualization of the paper. AIK carried out the analysis and wrote the results and discussion sections. AMMR prepared the literature review. AS worked on the scaling and data preparation. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Alec I. Kennedy.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

All authors provide consent for publication.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

See Tables 8, 910 and 11.

Table 8 Items used for constructing remote learning quality scales
Table 9 N of complete cases before- and after-imputation
Table 10 Reliability and measurement invariance of scales
Table 11 Items used for school-provided supports/resources for remote learning

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kennedy, A.I., Mejía-Rodríguez, A.M. & Strello, A. Inequality in remote learning quality during COVID-19: student perspectives and mitigating factors. Large-scale Assess Educ 10, 29 (2022). https://doi.org/10.1186/s40536-022-00143-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40536-022-00143-7

Keywords