 Research
 Open Access
 Published:
The relation between teachers’ emphasis on the development of students’ digital information and communication skills and computer selfefficacy: the moderating roles of age and gender
Largescale Assessments in Education volume 4, Article number: 17 (2016)
Abstract
Teachers’ integration of information and communication technology (ICT) has been widely studied, given that digital competence is considered to be a crucial outcome of twenty first century education. In this context, research highlighted teachers’ computer selfefficacy (CSE) as one of the most important determinants of their ICT integration into teaching practices. Whereas previous research mainly focused on the relation between CSE and ICT integration from a frequencybased point of view, recent research suggests to investigate this relation using more qualitative measures of ICT integration such as the degree to which teachers emphasize developing students’ digital information and communication skills (TEDDICS). Consequently, the present study investigates the relations between these two constructs: teachers’ emphasis on developing students’ digital skills and their computer selfefficacy, taking into account the moderating roles of age and gender. We used a representative sample of 1071 Norwegian secondary school teachers who participated in the international computer and information literacy study (ICILS) in 2013. Our results provide evidence on the positive relation between CSE and TEDDICS. Furthermore, age positively moderated this relation between some factors of the two constructs, indicating that computer selfefficacy plays an even more important role for teachers of higher age in the context of emphasizing ICT skills in classrooms. The unique effect of gender was present for one correlation between CSE and TEDDICS, indicating that moderation by gender was apparent to a limited extent, and related to use of computers for instructional purposes. The interaction between age and gender did not reveal significant moderation effects. We discuss these results in light of the potential consequences for teacher training.
Introduction
The role and use of information and communication technology (ICT) in education has changed profoundly over the last decade. This change is evident at many levels in education, for instance, with respect to the availability of ICT resources at schools, the access to internet, and the transition from paperandpencil to computerbased exams (Scherer and Siddiq 2015a; Scheuermann and Pedró 2009). Furthermore, students’ digital competence has gained substantial attention and is considered to be an important twenty first century skill (Griffin et al. 2012). As a consequence, a first line of research studied the determinants of teachers’ integration of ICT into classroom activities (Tondeur et al. 2008), given that the teachers play a key role in developing students’ digital skills (Schibeci et al. 2008). Specifically, teachers’ computer selfefficacy (CSE) has been identified as one of the most important determinants for teachers’ integration of ICT in teaching and learning practices (Kreijns et al. 2013; Mumtaz 2000). Existing research identified positive relations between teachers’ CSE and their use of ICT (e.g., Kreijns et al. 2013; Sang et al. 2010). It therefore seems, as if the degree to which teachers integrate ICT into their teaching depends on the beliefs in their capabilities of using ICT (i.e., selfefficacy). But these beliefs may depend on a number of factors. For instance, some research showed that teachers’ age and CSE are negatively related, indicating that older teachers are less selfefficacious than their younger colleagues (e.g., O’Bannon and Thomas 2014; Vanderlinde et al. 2014). Regarding the relation between teachers’ CSE and gender, there has been less consistent results depending on how CSE was measured (Ong and Lai 2006; Sang et al. 2010; Scherer and Siddiq 2015b; Sieverding and Koch 2009).
A second line of research focused on providing more finegrained conceptualizations of ICT use that not only reflect teachers’ bare use of ICT in classrooms, but also adds value by linking it to students’ digital skills (Siddiq et al. 2016). For instance, teachers’ emphasis on developing students’ digital information and communication skills (TEDDICS) was introduced as a goaloriented measure which combines teachers’ use of ICT and teaching practices with their beliefs about which digital skills are considered important (Fraillon et al. 2014). In a recent study, Siddiq and Scherer (2015) showed that teachers’ selfefficacy in using computers for instructional purposes and aspects of TEDDICS were positively related. However, an indepth view concerning this relation, which accounts for further factors of CSE on the one hand and for the potential effects of age and gender on the other hand, is still lacking. Such a view may provide detailed information on how the TEDDICSCSE relation operates in different age groups and across gender, and may help us identify potential needs for strengthening teachers’ CSE and TEDDICS.
On the basis of the findings described above, we first investigate the relation between different factors of teachers’ CSE and TEDDICS, and secondly, examine the moderating roles of teachers’ age, gender, and their interaction for this relation. Drawing on the Norwegian sample of lower secondary school teachers who participated in the international computer and information literacy study (ICILS) in 2013, we apply structural equation modelling and moderation analyses to examine these relations.
Background
Teachers’ computer selfefficacy (CSE)
Selfefficacy is defined as an individual’s beliefs about his or her capabilities and levels of performance related to a course of action (Bandura 1997). In educational research, teachers’ selfefficacy has been shown to play an important role in influencing their teaching practices and furthermore their students’ achievement and motivation (Skaalvik and Skaalvik 2007). Teachers’ computer selfefficacy was defined by Compeau and Higgins (1995), and refers to “an individual’s perception of his or her ability to use computers in the accomplishment of a task” (p. 191). Many researchers have taken a general approach toward studying this construct, assuming that there exists a general CSE factor only, which focuses on teachers’ general perceptions of their capabilities in using ICT (e.g., Durndell and Haag 2002; Teo 2014). Scherer and Siddiq (2015b) pointed out that this unidimensional view on CSE may have caused the somehow puzzling and contradictory results on the determining factors of teachers’ intentions toward technology usage. Together with Lee et al. (2009), they further argued that one way to solve this conundrum is to assume that teachers do not make general decisions about the use of technology in their classrooms, but rather individual judgments about specific uses. Therefore, the ways teachers make these decisions may vary according to the different types of ICT use. This view is supported by other researchers that consider the nature of selfefficacy to be specific to situations and domains (e.g., Dicke et al. 2014; O’Mara et al. 2006). They suggested using specific CSE measures that reflect the targeted performance rather than global assessments (Bong and Skaalvik 2003; Pajares and Schunk 2001).
As a consequence, a limited number of recent studies have adopted this view by operationalizing CSE as a multidimensional construct according to the different uses of computers for specific teaching and learning purposes (Scherer and Siddiq 2015b). This approach is in line with the requirements in national curricula that relate to students’ digital literacy as being composed of several facets (e.g., Aesaert et al. 2014; Claro et al. 2012; Ferrari 2013). We therefore consider CSE to be multidimensional and aligned with the specific facets of digital literacy.
Teachers’ emphasis on developing students’ digital information and communication skills (TEDDICS)
The construct ‘TEDDICS’ was developed in the context of ICILS 2013 (Fraillon et al. 2014). TEDDICS aims to gauge to what extent teachers’ emphasize the development of students’ ICTrelated skills. In contrast to existing measures of teachers’ use of ICT, which were mostly derived from indicators of the quantity, technology specificity, and the duration of ICT use (e.g., Akarsu and Akbiyik 2012; Hsiao et al. 2010; Yildirim 2000), TEDDICS represents a more qualitative aspect of ICT use (Fraillon et al. 2013; Siddiq et al. 2016). Furthermore, it bring together curricular demands and teachers’ beliefs about the importance of digital skills, further linking it to the development of students’ competence in this area (Fraillon et al. 2013).
In the twenty first century, managing digital information is regarded as a vital competence (Griffin et al. 2012). Frameworks on students’ digital competence comprise several facets, and most of the frameworks share common dimensions focusing on different activities of handling digital information (e.g., searching, accessing, evaluating, sharing and communicating digital information; Claro et al. 2012; Ferrari 2013; GallardoEchenique et al. 2015). Moreover, since a number of studies indicated that students struggle within this area and lack skills related to information retrieval and information processing (e.g., Aesaert et al. 2014; Kuiper et al. 2005), there is a pressing need for fostering these skills in the classroom. As a consequence, focusing on TEDDICS may provide information on potential opportunities to address this need.
The assessment of TEDDICS in ICILS 2013 captured the extent to which teachers’ emphasize the development of students’ competencies of handling digital information (i.e., accessing, evaluating, and sharing and communicating digital information). This measure was scrutinized by Siddiq et al. (2016) with respect to its internal and external validity. In fact, three TEDDICS factors, each representing one of the facets of dealing with digital information were identified. Furthermore, positive relations between TEDDICS, teachers’ use of ICT, and CSE in instruction were found (Ainley et al. 2015; Siddiq et al. 2016). On the basis of these findings, we argue that taking a multidimensional perspective on both CSE and TEDDICS can provide detailed information on the relation between specific ICTrelated selfbeliefs and the emphasis on developing specific digital and information skills in classrooms.
The roles of teachers’ age and gender as potential moderators
The existing body of research identified age and gender differences in the context of technology adoption (Morris and Venkatesh 2000), information technology acceptance (Teo 2014), computer experience (Hsiao et al. 2010), and ICT integration (Cassidy and Eachus 2002). These differences may also affect the relation between TEDDICS, a construct closely related to ICT use and integration, and CSE, a construct determining the use and integration of ICT. We thus provide a brief summary of existing findings on age and gender effects in the following subsections.
Teachers’ age
In the context of technology acceptance and integration of ICT into classrooms, teachers’ age was identified as a potential source of variation in the constructs involved. For instance, older teachers tend to express lower levels of perceived usefulness of ICT, computer selfefficacy, and perceived ease of use than their younger colleagues (O’Bannon and Thomas 2014; Vanderlinde et al. 2014; Venkatesh et al. 2003). In line with these observations, older teachers also display higher ICT anxiety (Mac Callum et al. 2014) and emphasize problems and obstacles created by the use of ICT for teaching and learning more than younger teachers (Scherer et al. 2015).
Gender differences
Gender differences in ICTrelated constructs have gained considerable attention. One reason for this attention may lie in the fact that existing studies have provided conflicting findings on both, the direction and significance of the gender effects. For instance, significant gender effects were reported for constructs such as teachers’ ICT use, CSE, and perceived usefulness (e.g., Scherer and Siddiq 2015b; Volman and van Eck 2001). On the contrary, a number of studies could not identify gender differences in these constructs (Antonietti and Giorgetti 2006; Shapka and Ferrari 2003; Teo 2008). Hence, these conflicting findings require a continued focus on whether differences across gender exist for the specific constructs and samples under investigation.
Potential moderation effects
In light of the above mentioned findings on age and gender differences in ICTrelated constructs, it is currently unclear whether or not the relation between CSE and TEDDICS is affected by teachers’ age and gender. In research on general selfefficacy, teachers’ age, gender, and main subject have been integrated as moderators. Specifically in the context of instructional selfefficacy, age and gender are considered to be moderators of different relations among classroom management, teaching effectiveness, and job satisfaction (e.g., Dicke et al. 2014; Klassen and Tze 2014). However, in the context of teachers’ ICT integration in classroom practice, moderation effects of age and gender on the relation between ICTrelated constructs have rarely been explored in detail (Schepers and Wetzels 2007). Thus, we cannot be certain if the TEDDICSCSE relation is also subject to age and gender differences. In other words, potential age or gender differences in each of the two constructs may not necessarily imply differences in their relation.
The present study
This study attempts to provide a detailed view on the TEDDICSCSE relation by using multidimensional measures of both constructs. Moreover, since it has been unclear whether this relation is robust against age and gender differences, we include these variables along with their interaction as potential moderators (see Fig. 1). Specifically, we address two research questions:

1.
How does teachers’ emphasis on developing students’ digital information and communication skills relate to teachers’ computer selfefficacy?

2.
To what extent do age, gender, and their interaction (age × gender) moderate the relation between TEDDICS and CSE?
Methods
Sample and procedure
The current study is based on the Norwegian sample of secondary school teachers who participated in ICILS 2013 (Fraillon et al. 2014). In total, 1071 teachers responded to both, the TEDDICS and CSE scales, and provided information on their background (e.g., age, gender, and main subjects). Norwegian teachers were randomly sampled in a twostep procedure (step 1: sampling of schools, step 2: sampling of teachers within schools), and were based in 132 secondary schools in different municipalities across Norway. The sampling accounted for schools’ composition, background, and socioeconomic characteristics. Teachers’ mean age was 44.3 years (SD = 11.2) and ranged between 23 and 71 years (64.2 % female teachers).
We estimated the reliability of each TEDDICS and CSE factor as McDonald’s ω (Yang and Green 2011). All analyses were employed in the statistical package Mplus 7.3 (Muthén and Muthén 1998–2015).
Measures
Teachers’ emphasis on developing students’ digital information and communication skills (TEDDICS)
Since students’ skills in accessing, evaluating, and sharing and communicating digital information are considered to be crucial factors of digital competence (Fraillon et al. 2013), we used the multidimensional measure of TEDDICS that was used in ICILS 2013 (Jung and Carstens 2015; Siddiq et al. 2016). This measure distinguishes between three factors of the construct: ‘Accessing digital information’ (ω = .79, 3 items), ‘Evaluating digital information’ (ω = .90, 4 items), and ‘Sharing and communicating digital information’ (ω = .80, 5 items). Teachers were asked to rate the degree to which they emphasize the development of these skills in their lessons on a 4point scale ranging from ‘0 = no emphasis’ to ‘3 = strong emphasis’. Please review the supplementary material for the item wordings and labels of this scale (see Additional file 1: A1).
Teachers’ computer selfefficacy (CSE)
The assessment of teachers’ CSE comprised the beliefs in their capabilities of performing specific operational tasks with the help of computers on the one hand, and using computers for instructional purposes on the other hand (Fraillon et al. 2014; Jung and Carstens 2015; Scherer and Siddiq 2015b). Specifically, teachers were asked to rate the degree to which they perceived their capabilities of performing 14 computer tasks on a threepoint rating scale (0 = I do not think I can do this, 1 = I could work out how to do this, 2 = I know how to do this). Based on Bandura’s (1997) recommendations on measuring selfefficacy, the item stimulus referred to the degree to which they believed they can do these tasks. In total, 14 items were used to measure three factors of the construct with sufficient reliabilities: Selfefficacy in basic operational skills (ω = .79, 6 items), selfefficacy in advanced operational and collaborative skills (ω = .72, 4 items), and selfefficacy in using computers for instructional purposes (ω = .76, 4 items). The distinction between these three CSE factors has recently been confirmed empirically, and sufficient evidence on the validity of the CSE assessment was obtained (Scherer and Siddiq 2015b). Item wordings and labels used in ICILS 2013 can be found in the Additional file 1: A2.
Statistical analyses
Research question 1
In order to address our first research question on the relation between TEDDICS and CSE, we specified correlatedtraits models of confirmatory factor analysis for both constructs and studied the correlation between the latent variables (Brown 2013). These models distinguished between the three TEDDICS factors (Accessing, evaluating, sharing & communicating digital information) and the three CSE factors (CSE in basic operational ICT tasks, CSE in advanced operational ICT and collaboration tasks, and CSE in using ICT for instructional purposes), and resulted in nine correlations. In these analyses, we treated teachers’ responses categorically and applied weighted least squares means and variance adjusted (WLSMV) estimation (Rhemtulla et al. 2012; Sass et al. 2014).
In order to evaluate the goodnessoffit of the models, we examined model fit statistics such as the χ^{2} value, the root mean square error of approximation (RMSEA), the comparative fit index (CFI), and the Tucker Lewis index (TLI), and applied common guidelines for an acceptable model fit: RMSEA ≤ .08, CFI ≥ .95, and TLI ≥ .95 (Marsh et al. 2005). We note that a significant χ^{2} value indicates substantial deviations of the empirically implied model from the model that is based on the actual data. Nevertheless, this statistic might show a significant value although the model fits the data, given the relatively large sample size. As a consequence, we did not base our decision for or against a model solely on this statistic.
Research question 2
Our second research question was concerned with the moderating effects of age, gender, and their interaction on the TEDDICSCSE relations. This question was approached in a sequence of modelling steps: First, we examined whether or not the measurement models of TEDDICS and CSE provided representations of the constructs that are invariant across gender. This step was necessary to ensure that potential gender differences in the relations between TEDDICS and CSE were not due to differences in the measurement of the two constructs (Millsap 2011). Specifically, we tested the three models of configural, metric, and scalar invariance. In the configural invariance model, the same factor structure is specified for female and male teachers, assuming that the same number of factors and links between manifest and latent variables are present. This model is the least restrictive and forms the basis for further invariance testing and model comparisons. Subsequently, item factor loadings were constrained to be equal across the gender groups, resulting in a model of metric invariance. If this model can be accepted, the relations among latent variables and to external variables can be compared. Finally, the item thresholds were constrained in the scalar invariance model. Establishing scalar invariance is considered to be the prerequisite for meaningful comparisons among the means of the latent variables (Byrne et al. 1989). However, for comparing the TEDDICSCSE relations across gender, metric invariance is sufficient.
In order to decide on which level of invariance was achieved, model comparisons were conducted on the basis of χ^{2} difference testing, and the differences in the goodnessof fit statistics (i.e., CFI, TLI, and RMSEA) between two invariance models were taken into account. In particular, we compared the metric and scalar model with the configural model and regarded changes of ΔCFI ≤ .010, ΔTLI ≤ .010, and ΔRMSEA ≤ .015 as insignificant (Cheung and Rensvold 2002). Hence, if the changes in these statistics were within the suggested cutoffs, the changes in the χ^{2} statistics were rather low or insignificant. If the model also showed an acceptable fit, the more restricted invariance model was accepted. Gender differences in the resulting correlations were tested with the help of Wald’s χ^{2} test (Van de Schoot et al. 2012). Significant differences in the TEDDICSCSE correlations point to the moderating role of gender. Please find a sample Mplus code for the invariance testing in Additional file 1: B1.
Second, we investigated the moderation effects of age by establishing latent regression models with the TEDDICS factors as outcome variables, teachers’ age, the CSE factors, and their interaction (Age × CSE) as predictors. In these analyses, age was zstandardized to avoid nonessential multicollinearity (Dalal and Zickar 2012; Marsh et al. 2014). The interaction between a latent CSE variable and the manifest age variable was established using the ‘XWITH’ and ‘TYPE = RANDOM’ options in Mplus (Muthén and Muthén 1998–2015). These options are typically used to define interactions between either two latent variables or a latent and a manifest variable (e.g., Little et al. 2006). For specifying the interaction models, we used the robust maximum likelihood estimator (MLR) with corrected standard errors and χ^{2} statistics in conjunction with Monte Carlo integration and 500 integration points. Given that this numerical integration method becomes computationally very demanding if a number of correlated latent variables are used simultaneously to create interaction terms, we decided to run the age moderation models for each of the three TEDDICS and CSE factors separately. Another argument supporting this decision is that the CSE factors are highly correlated resulting in multicollinearity when used as predictors in regression models. Although theoretically possible, we did not use the factor scores obtained from the TEDDICS and CSE measurement models to estimate the moderation effects. This approach could have resulted in heavily biased regression coefficients (Skrondal and Laake 2001). We notice that teachers’ responses were treated categorically in all moderation analyses. Please find an example code for these models in Additional file 1: B2. If the 95 % confidence interval of the regression coefficient of the interaction Age × CSE did not contain zero, moderation was indicated (Marsh et al. 2014).
Third, teachers’ gender was added to the moderation analyses, resulting in models with three single predictors (Age, gender, and CSE), three twoway interactions (Age × gender, Age × CSE, and gender × CSE), and a threeway interaction term (Age × gender × CSE). To decide on whether or not Age × gender moderated the TEDDICSCSE relations, we inspected the 95 % confidence interval of the corresponding regression coefficient of the threeway interaction term.
Handling clustered and missing data
Due to the clustered data structure in ICILS 2013 (i.e., teachers are nested in schools), we adjusted the standard errors of the model parameters and the χ^{2} statistics, using the MLR estimator and the ‘TYPE = COMPLEX’ option in Mplus for the moderation analyses. Furthermore, differences in the probabilities of being sampled as a teacher were accounted for by using teachers’ sampling weights (Mplus option ‘WEIGHT = TEACHWT’; Asparouhov 2005). As mentioned earlier, teachers’ responses were treated categorically using the WLSMV estimator for establishing the measurement models and testing for invariance across gender. This treatment also allows for the incorporation of the ‘TYPE = COMPLEX’ and weight options.
Among the teachers who responded to the TEDDICS and CSE scales, low proportions of missing values at the item level occurred (less than 1 %). Since these missing values were not due to the design of the study, we assumed that they were ‘missing at random’ and applied the fullinformationmaximumlikelihood procedure to handle them in the moderation models (Enders 2010). In the cases of using the WLSMV estimator, missing data were handled with the help of the pairwise deletion method (Asparouhov and Muthén 2010).
Results
Descriptive statistics and measurement models
The item descriptive statistics for both the TEDDICS and CSE scales are shown in Table 1. It is noteworthy that the teachers reported high levels of computer selfefficacy for most of the CSE items, as suggested by the means and the medians. Hence, statistical models that are based on a perfect normal distribution of the manifest indicators may not apply to CSE. We consequently decided to account for this deviation in subsequent analyses. For items belonging to the CSE factor of advanced operational and collaboration skills, the means of responses were lower than for the others. This result indicated that this factor may, indeed, present skills that are more demanding and difficult for teachers than others. However, these differences were by and large statistically insignificant, except for the most extreme mean differences (e.g., between items IT1G07A and IT1G07 M, t [1066] = –36.5, p < .001, r = .07), and only point to tendencies. In contrast, potential ceiling effects were not identified for the TEDDICS scale, as the means and medians were lower than the maximum scores of items. Nevertheless, we decided to be consistent in treating the data categorically and accounted for deviations from normal distributions. After inspecting the descriptive statistics, we established the measurement models of TEDDICS and CSE.
TEDDICS measurement model
In a recently published study, which examined the validity of the TEDDICS scale in ICILS 2013 (Siddiq et al. 2016), it was shown that this scale comprised three correlated factors of teachers’ emphasis on developing students’ skills in accessing (factor 1), evaluating (factor 2), and sharing and communicating digital information (factor 3). We therefore based our assumptions on the structure of the construct on this finding, establishing a correlatedtraits confirmatory factoranalytic model with three factors. This model fitted the data well, χ^{2} (51) = 368.6, p < .001, RMSEA = .076, 90 % CI RMSEA = [.069, .084], CFI = .984, TLI = .980, and indicated sufficiently high factor loadings for each of the factors (TEDDICS factor 1: standardized λ = .74–.83, TEDDICS factor 2: standardized λ = .84–.97, and TEDDICS factor 3: standardized λ = .67–.78). Although the factor correlations were rather high (ρ = .86–.90; see Table 2), and a unidimensional model fitted the data only slightly worse, χ^{2} (54) = 479.9, p < .001, RMSEA = .086, 90 % CI RMSEA = [.079, .093], CFI = .979, TLI = .974, ∆χ^{2} (3, N = 1071) = 132.4, p < .001, we decided to keep the distinction between the three factors for substantive reasons. Specifically, we wanted to see how different aspects of TEDDICS related to CSE rather than examining this relation for an overall emphasis on developing students’ skills in the context of ICT. In addition to establishing a threefactor measurement model for the total sample, we fitted the same model to the subsample of female and male teachers. As for the total sample, the model showed an acceptable fit for both females, χ^{2} (51) = 251.5, p < .001, RMSEA = .076, 90 % CI RMSEA = [.066, .085], CFI = .987, TLI = .983, and males, χ^{2} (51) = 179.8, p < .001, RMSEA = .081, 90 % CI RMSEA = [.069, .094], CFI = .980, TLI = .974. Hence, it can be used to study measurement invariance across gender and potential moderation effects of gender in subsequent analyses.
CSE measurement model
Following the same procedure, we specified a threefactor model for teachers’ computer selfefficacy, assuming that CSE in basic operational ICT skills (factor 1), CSE in advanced operational and collaboration skills (factor 2), and, finally, CSE in using computers for instructional purposes (factor 3) can be distinguished. This distinction was also based on prior research (e.g., Scherer and Siddiq 2015b). The resulting confirmatory factoranalytic model showed an excellent fit for the total sample, χ^{2} [74] = 167.1, p < .001, RMSEA = .034, 90 % CI RMSEA = [.027, .041], CFI = .979, TLI = .974. As for the TEDDICS model, correlations among the latent variables were rather high (ρ = .74–.77; see Table 2); however, a unidimensional model fitted the data significantly worse, χ^{2} [77] = 280.2, p < .001, RMSEA = .050, 90 % CI RMSEA = [.043, .056], CFI = .953, TLI = .945, Δχ^{2} [3, N = 1071] = 124.4, p < .001. Hence, we accepted the threefactor model as a measurement model of CSE, also because the loadings for each factor were reasonably high (CSE factor 1: standardized λ = .80–.99, CSE factor 2: standardized λ = .64–.79, and CSE factor 3: standardized λ = .83–.92). This model fitted the data well for females, χ^{2} [74] = 136.0, p < .001, RMSEA = .035, 90 % CI RMSEA = [.026, .044], CFI = .968, TLI = .961, and males, χ^{2} [74] = 121.6, p < .001, RMSEA = .041, 90 % CI RMSEA = [.027, .054], CFI = .988, TLI = .985. As a consequence, this model formed the baseline for further invariance testing across gender.
Correlations among the TEDDICS and CSE factors (Research Question 1)
To address Research Question 1, we combined the measurement models of TEDDICS and CSE, and examined the correlations among the latent variables. The combined model had an acceptable fit, χ^{2} (284) = 504.1, p < .001, RMSEA = .027, 90 % CI RMSEA = [.023, .031], CFI = .987, TLI = .985, and indicated low to moderate correlations (ρ = .15–.41; Table 2). Since the resulting factor correlations were positive and significant (see Table 2), it can be concluded that higher levels of computer selfefficacy are associated with higher levels of emphasis on developing students’ digital information and communication skills, and vice versa. The highest correlations occurred between the CSE factor of ‘Selfefficacy in using computers for instructional purposes’ and all TEDDICS factors (ρ = .37–.41). The lowest correlation was found between CSE in basic operational ICT skills and the TEDDICS factor of ‘accessing digital information’ (ρ = .15). In light of these findings, our response to Research Question 1 is: The factors of TEDDICS and CSE are positively correlated.
Moderation analyses (Research Question 2)
Moderation by gender
As mentioned earlier, measurement invariance is considered to be a prerequisite for comparing the TEDDICSCSE correlations across gender. Since the baseline measurement models for both TEDDICS and CSE have been established successfully, further invariance models could be specified using multigroup confirmatory factor analysis. The results of invariance testing were clearcut and suggested that the three invariance levels (configural, metric, and scalar) could be established (see Table 3). This was evident, because these models showed an acceptable fit to the data on the one hand, and indicated only small changes in the fit statistics, as compared to the configural model, on the other hand. In sum, comparing the relations between TEDDICS and CSE across gender was legitimate.
In order to investigate potential differences in the TEDDICSCSE relations, we established a multigroup model that combined TEDDICS and CSE under the scalar invariance assumptions. The model fitted the data very well, χ^{2} (626) = 898.9, p < .001, RMSEA = .029, 90 % CI RMSEA = [.024, .033], CFI = .985, TLI = .985, and was therefore accepted. To rule out that potential differences in the correlations were not due to differences in the factor correlations within the TEDDICS and CSE measurement models or differences in factor variances, we constrained these parameters in addition to the scalar invariance assumptions. These constraints led to a wellfitting multigroup model, χ^{2} (632) = 908.3, p < .001, RMSEA = .029, 90 % CI RMSEA = [.024, .033], CFI = .985, TLI = .985, which was used for comparisons among the correlations. We compared the correlations among the TEDDICS and CSE factors by performing the Wald χ^{2} test (Van de Schoot et al. 2012). Specifically, we first tested whether or not any differences in the correlations existed (overall test) and examined which specific correlations differed (local test) in a second step.
The pattern of relations for female and male teachers, by and large, corresponded (see Table 4). More specifically, all correlations except for the one between ‘TEDDICS: Accessing digital information’ and ‘CSE: Basic operational skills’ were positive and statistically significant. The overall Wald χ^{2} test indicated that differences in the TEDDICSCSE correlations, χ^{2} (9) = 19.2, p < .05. Testing the differences in correlations with a stepwise procedure (i.e., local test) revealed that only the correlation between ‘Sharing and communicating digital information’ and ‘selfefficacy in using computers for instructional purposes’ was subject to gender differences in favour of female teachers, χ^{2} (1) = 6.0, p < .05. The remaining correlations were similar across gender (see Additional file 1: C1). Although significant for only one correlation, there was a tendency toward stronger relations between the two TEDDICS factors of evaluating and sharing & communicating digital information and CSE for female teachers. As a consequence, given that only one of the TEDDICSCSE correlations showed gender differences, moderation by gender was apparent to a limited extent.
Moderation by age
Investigating the moderation by teachers’ age, we specified a series of models with an interaction between a CSE factor and age as a predictor of a TEDDICS factor (see Method section). The resulting information criteria of these nine models (3 TEDDICS factors × 3 CSE factors) are presented in Additional file 1: C2. Regarding the regression coefficients of the interaction term CSE × Age in these models, only two out of nine coefficients showed statistical significance, as their confidence intervals did not contain zero (see Table 5). This applied to the prediction of the TEDDICS factor ‘Accessing digital information’ and ‘Sharing & communicating digital information’ by ‘CSE: Advanced operational and collaboration skills’. In these two cases, moderation by age was present; the coefficients were positive and therefore indicated that the relation between CSE and TEDDICS was stronger as age increased. Alternatively, it may also be concluded that the relation between age and TEDDICS was stronger for teachers’ with high CSE than for teachers with low CSE.
Moderation by age × gender
Finally, we tested for threeway interaction effects by adding gender to the age moderation models. The corresponding information criteria of the nine models can be found in Additional file 1: C3. In none of the full models containing all possible interaction terms, it was possible to identify significant moderation by age × gender (see Table 6). Most of the confidence intervals were rather large and contained zero. In addition to this finding, the information criteria of the age × gender moderation models (see Additional file 1: C3) were by and large higher than those of the age moderation models (see Additional file 1: C2), suggesting that adding gender and further interaction terms may not necessarily improve the fit of the model. Hence, we conclude that there is not enough evidence to argue for an age × gender moderation of the TEDDICSCSE relations.
Taken together, with respect to Research Question 2, our findings suggested that gender and age moderation were present for some of the TEDDICSCSE relations; yet, the age × gender moderation could not be identified.
Discussion
The aim of the current study was to deepen the understanding of how teachers’ selfefficacy in using computers is related to their emphasis on developing students’ digital skills (Research Question 1), and to what extent age, gender, and their interaction moderate this relation (Research Question 2). Applying structural equation modelling, we found support for positive and significant relations between the three factors of TEDDICS and the three CSE factors. Furthermore, for comparing the TEDDICS and CSE relations across gender, scalar invariance was established. On the basis of the invariant model, we provided evidence for the moderating role of gender, indicating at least one significantly higher TEDDICSCSE correlation favouring female teachers. Further analyses showed moderation effects of age on two TEDDICSCSE correlations, indicating a stronger relation as teachers’ age increases. Finally, the moderation effects of age × gender could not be identified.
The relations between TEDDICS and CSE factors (Research Question 1)
An indepth view was provided by examining the TEDDICSCSE relations for the three facets of TEDDICS and the three facets of CSE. The results support our assumptions of positive relations between the two constructs, meaning that teachers who believe in their competences related to use of computers also emphasize developing their students’ digital skills in their classroom more. Interestingly, the highest correlations were identified between all three TEDDICS factors and the CSE factor ‘Using computers for instructional purposes’. This CSE factor is related to teachers’ beliefs in their competence of using computers in classroom settings (Scherer and Siddiq 2015b). Whereas the other two factors of CSE refer to operating computers at different levels of competence (basic operational, and advanced operational and collaboration skills), the instructional CSE factor reflects the embedment of computers in instructional settings and for teaching purposes. As a consequence, the significant correlations between this factor and the three TEDDICS factors may be due to their commonalities in focusing on instructional activities. Nevertheless, the correlations are moderate, suggesting that TEDDICS and CSE are still distinct and take different perspectives on teaching and learning with computers (Siddiq et al. 2016). This finding supports the notion that CSE should not be studied as a general construct but rather refer to more specific capabilities of using computers, for instance, in classroom settings (Dicke et al. 2014).
The correlations between the three factors of CSE were moderate, whereas the correlations between the three factors of TEDDICS were rather high (Table 2). As a consequence, the three TEDDICS factors showed similar correlations with the CSE factors. This finding indicates that the differentiation of TEDDICS is not clearly evident in this sample of Norwegian teachers. One explanation may be that teachers who emphasize the development of students’ ICT skills in one of the three hypothesized factors may out emphasis on the other factors to the same extent. In fact, the digital skills proposed in the TEDDICS framework are closely related and might reflect a process rather than a set of skills (Siddiq et al. 2016). Another explanation may lie in the fact that each of the factors contained only a limited number of items, which may not necessarily provide enough indicators in order to distinguish between the three TEDDICS factors. We therefore suggest developing and empirically investigating alternative and more extended measures of the TEDDICS construct.
It must be noticed that the positive TEDDICSCSE relations advocate that if teachers are expected to instruct students in order to improve their digital skills, selfconfidence in their own digital skills may be beneficial in order to meet these instructional expectations (e.g., Niederhauser and Perkmen 2010). Henceforth, teachers that do not see themselves as competent in these matters are less likely to emphasize the development of students’ digital and information skills. This finding can be discussed generally in the context of teachers’ selfefficacy and their instructional practices. Specifically, Holzberger, Philipp, and Kunter (2013) showed that teachers’ general selfefficacy outside the context of ICT is related to their instructional behaviour, even in a longitudinal perspective. TschannenMoran and Woolfolk Hoy (2007) present a slightly different perspective on this relation: They propose a number of sources of selfefficacy, of which the most important one refers to the mastery experience people make. As such, positive (mastery) experience in specific tasks may increase people’s selfefficacy in these tasks. Transferring this general argumentation into the ICT context, we argue that teachers who design instructional settings in order to emphasize the development of students’ digital skills may make mastery experience in such scenarios, which in turn could strengthen their selfefficacy in using computers for instructional purposes in the future.
Given the undeniable importance of selfefficacy even in the context of ICT, one may stress the necessity of teacher training programs being closely related to handson teaching practice in their subject domains in order to strengthen their computer selfefficacy (Hennessy et al. 2005; Scherer and Siddiq 2015b). Finally, our findings are in line with previous research on CSE as a significant predictor of teachers’ use and implementation of ICT in classrooms (Akarsu and Akbiyik 2012; Chen 2010; Teo 2008).
The moderating roles of age, gender and their interaction (Research Question 2)
Gender effects
The premise of an invariant measurement model was met and facilitated further analyses for comparing male and female teachers. Acquiring evidence of measurement invariance is vital for assuring that the measures do not act differently across gender groups (Scherer and Siddiq 2015b). Based on this premise, significant gender differences in favour of female teachers were identified for only one out of nine TEDDICSCSE correlations. Since gender effects were not found for all nine TEDDICSCSE relations, our findings suggest that male and female teachers may differ in some matters related to ICT to a limited extent. Nevertheless, previous research in ICTrelated investigations provided contradicting findings on the existence of gender differences (Durndell and Haag 2002; Pamuk and Peker 2009; Shapka and Ferrari 2003; Sieverding and Koch 2009). Furthermore, previous research did not find evidence on differences across gender for the TEDDICS construct (Siddiq et al. 2016), and only partly for CSE (Scherer and Siddiq 2015b). Accordingly, our results provide only limited evidence of gender effects in particular ICT contexts.
Specifically, the relation between the TEDDICS factor ‘Sharing & communicating digital information’ and the CSE factor ‘Using ICT for instructional purposes’ was stronger for female teachers than for male teachers. This result points toward the belief that female teachers may lack confidence in their competences in using computers for teaching (Scherer and Siddiq 2015b; Sieverding and Koch 2009), and consequently put less emphasis on developing students’ ICTrelated skills. The other two CSE factors in this study are to a larger degree related to teachers’ use of computers for personal matters. Hence, these findings suggest that the gender gap related to CSE in general is narrowing. Although, it is apparent that there are differences between female and male teachers regarding to what extent they feel confident to integrate ICT in their teaching practices, namely their technological pedagogical content knowledge (Koehler and Mishra 2009). However, it may also mirror results from existing studies which revealed that male and female teachers’ respond differently when evaluating their ICT competences. Male teachers tend to regard themselves as more competent and female teachers are more inclined to underestimate their own competence (Cooper 2006; Ong and Lai 2006). Our findings indicate that the influence of computer selfbeliefs is, to some extent, subject to gender differences.
Age effects
The age effects identified in our study indicate that some of the TEDDICSCSE relations tend to be stronger for older teachers than for their younger colleagues. In other words, the influence of the CSE factor related to advancedoperational and collaboration skills on the two TEDDICS factors ‘Accessing digital information’ and ‘Sharing & communicating digital information’ is more important for teachers of higher age. However, since the ICILS 2013 data do not allow causal interpretations of the direction of these relations, alternative explanations may exist. For example, we may also conclude that the relations between age and the two TEDDICS factors were stronger for teachers with high CSE than for teachers with low CSE. Either ways, our results agree with prior research by showing that teachers’ age plays a noteworthy role in their ICT use and selfbeliefs (O’Bannon and Thomas 2014; Scherer et al. 2015; Vanderlinde et al. 2014).
Age × gender effects
Finally, as the interaction effects between teachers’ age and gender did not moderate the TEDDICSCSE relations, we do not have evidence that the moderation by gender was specific to certain age groups, and the moderation by age was not sensitive to gender differences. It therefore seems as if the standalone effects of age and gender dominate the moderation. Nevertheless, the identification of such complex moderations is often subject to high standard errors and broad confidence intervals (Afshartous and Preston 2011). Moreover, the incorporation of further interaction terms (e.g., CSE × gender, CSE × age) increases the complexity of the regression model and may introduce essential multicollinearity (Marsh et al. 2014). As a consequence, we need to consider these findings in light of the methodological complexities.
Limitations and future directions
The present study has a number of limitations that point to future research: First, we only investigated the relations between CSE and TEDDICS. Future research may study these constructs as part of a bigger framework such as the technology acceptance model (Ong and Lai 2006), in which further measures related to ICT attitudes, use, and beliefs are included (e.g., perceived usefulness of ICT; Scherer et al. 2015). Second, we restricted our analyses to the Norwegian context, in which ICT plays an important role in school curricula (Norwegian Directorate for Education and Training 2012). It would therefore be interesting to examine the generalizability of our findings across further countries and educational contexts. In fact, taking an international perspective on the measures of and relations between CSE and TEDDICS may provide information on their differences and similarities. Finally, only a limited number of items were assigned to the three facets of TEDDICS; this design issue may have caused the considerable high correlations among the TEDDICS factors. We therefore suggest putting further effort into the development of items and in investigating the extent to which a broader TEDDICS assessment is able to differentiate between the three hypothesized factors. Moreover, it still needs to be disentangled how well the TEDDICS facets can be used to inform teacher professional development and practice.
Conclusions
In light of the findings the present study has revealed, we first conclude that teachers’ computer selfefficacy plays a significant role for their emphasis on developing students’ digital and information skills in classroom settings. This finding suggests that feeling competent in using ICT for instructional purposes may be regarded as a prerequisite for emphasizing the development of students’ ICT skills. Hence, teacher training intuitions may emphasize the development of teachers’ technological pedagogical content knowledge to enable and strengthen their competence of ICT integration in classroom activities. Second, we showed that the TEDDICSCSE relations are, to some extent, subject to gender and age effects. This finding suggests that the importance of CSE for TEDDICS does not distribute equally between males and females, and across age groups. This may point to the need for designing teacher training programs that are aimed at fostering CSE and specifically take into account gender and age variation. We conclude that this study provides knowledge that could benefit teacher training programs, and may be further useful for designing teacher development material which takes in account that female teachers may have lower confidence in their technological pedagogical content knowledge.
Abbreviations
 AIC:

Akaike’s Information Criterion
 BIC:

Bayesian Information Criterion
 CFI:

comparative fit index
 CI:

confidence interval
 CSE:

computer selfefficacy
 df :

degrees of freedom
 ICILS:

international computer and information literacy study
 ICT:

information and communication technology
 MLR:

robust maximum likelihood estimator
 RMSEA:

root mean square error of approximation
 TEDDICS:

teachers’ emphasis on developing students’ digital information and communication skills
 TLI:

Tucker Lewis index
 ω:

McDonald’s ω (reliability coefficient)
 WLSMV:

weighted least squares means and variance adjusted estimator
References
Aesaert, K., van Nijlen, D., Vanderlinde, R., & van Braak, J. (2014). Direct measures of digital information processing and communication skills in primary education: using item response theory for the development and validation of an ICT competence scale. Computers Education, 76, 168–181. doi:10.1016/j.compedu.2014.03.013.
Afshartous, D., & Preston, R. A. (2011). Key results of interaction models with centering. Journal of Statistics Education, 19(3), 1–24. http://www.amstat.org/publications/jse/v19n3/afshartous.pdf. Accessed 27 Oct 2015
Ainley, J., Friedman, T., Gebhardt, E., & Macaskill, G. (2015, June). Teaching About Information and Communication Technologies. Paper presented at the 6th IEA International Research Conference (IRC), Capetown, South Africa.
Akarsu, B., & Akbıyık, C. (2012). Relationships among perceived computer literacy skills, computer attitudes, and computer selfefficacy levels. Journal of European Education, 2(2). http://jee.erciyes.edu.tr/article/view/5000121646/5000112096. Accessed 20 Oct 2015
Antonietti, A., & Giorgetti, M. (2006). Teachers’ beliefs about learning from multimedia. Computers in Human Behavior, 22, 267–282. doi:10.1016/j.chb.2004.06.002.
Asparouhov, T. (2005). Sampling weights in latent variable modeling. Structural Equation Modeling: A Multidisciplinary Journal, 12, 411–434. doi:10.1207/s15328007sem1203_4.
Asparouhov, T., & Muthén, B. (2010). Weighted least squares estimation with missing data. https://www.statmodel.com/download/GstrucMissingRevision.pdf. Accessed 12 Oct 2015
Bandura, A. (1997). Selfefficacy: The exercise of control. New York: Freeman.
Bong, M., & Skaalvik, E. M. (2003). Academic selfconcept and selfefficacy: How different are they really? Educational Psychology Review, 15, 1–40. doi:10.1023/A:1021302408382.
Brown, T. A. (2013). Latent variable measurement models. In T. D. Little (Ed.), The Oxford handbook of quantitative methods (Vol. 2, pp. 257–280). New York: Oxford University Press.
Byrne, B. M., Shavelson, R. J., & Muthén, B. (1989). Testing for the equivalence of factor covariance and mean structures: The issue of partial measurement invariance. Psychological Bulletin, 105(3), 456–466.
Cassidy, S., & Eachus, P. (2002). Developing the computer user selfefficacy (CUSE) scale: Investigating the relationship between computer selfefficacy, gender and experience with computers. Journal of Educational Computing Research, 26(2), 169–189.
Chen, R. J. (2010). Investigating models for preservice teachers’ use of technology to support studentcentered learning. Computers Education, 55, 32–42. doi:10.1016/j.compedu.2009.11.015.
Cheung, G. W., & Rensvold, R. B. (2002). Evaluating goodnessoffit indexes for testing measurement invariance. Structural Equation Modeling: A Multidisciplinary Journal, 9, 233–255. doi:10.1207/S15328007SEM0902_5.
Claro, M., Preiss, D. D., San Martín, E., Jara, I., Hinostroza, J. E., Valenzuela, S., et al. (2012). Assessment of 21st century ICT skills in Chile: Test design and results from high school level students. Computers Education, 59(3), 1042–1053.
Compeau, D. R., & Higgins, C. A. (1995). Application of social cognitive theory to training for computer skills. Information Systems Research, 6(2), 118–143. doi:10.1287/isre.6.2.118.
Cooper, J. (2006). The digital divide: The special case of gender. Journal of Computer Assisted Learning, 22, 320–334. doi:10.1111/j.13652729.2006.00185.x.
Dalal, D. K., & Zickar, M. J. (2012). Some common myths about centering predictor variables in moderated multiple regression and polynomial regression. Organizational Research Methods, 15(3), 339–362.
Dicke, T., Parker, P. D., Marsh, H. W., Kunter, M., Schmeck, A., & Leutner, D. (2014). Selfefficacy in classroom management, classroom disturbances, and emotional exhaustion: A moderated mediation analysis of teacher candidates. Journal of Educational Psychology, 106(2), 569–583. doi:10.1037/a0035504.
Durndell, A., & Haag, Z. (2002). Computer selfefficacy, computer anxiety, attitudes towards the Internet and reported experience with the Internet, by gender, in an East European sample. Computers in Human Behavior, 18, 521–535. doi:10.1016/S07475632(02)000067.
Enders, C. (2010). Applied missing data analysis. New York: Guilford Press.
Ferrari, A. (2013). DIGCOMP: A framework for developing and understanding digital competence in Europe. Luxembourg: Publications Office of the European Union. doi:10.2788/52966.
Fraillon, J., Ainley, J., Schulz, W., Friedman, T., & Gebhardt, E. (2014). Preparing for life in a digital age—the IEA international computer and information literacy study. International report. Amsterdam: IEA.
Fraillon, J., Schulz, W., & Ainley, J. (2013). International computer and information literacy study: Assessment framework. Amsterdam: IEA.
GallardoEchenique, E. E., de Oliveira M. J., MarquésMolias, L., & EsteveMon, F. (2015). Digital competence in the knowledge society. MERLOT Journal of Online Learning and Teaching, 11(1), 1. jolt.merlot.org/vol11no1/GallardoEchenique_0315.pdf. Accessed 20 Oct 2015
Griffin, P., Care, E., & McGraw, B. (2012). The changing role of education and schools. In P. Griffin, B. McGraw, & E. Care (Eds.), Assessment and teaching of 21st century skills (pp. 1–15). New York: Springer Science + Business.
Hennessy, S., Ruthven, K., & Brindley, S. (2005). Teacher perspectives on integrating ICT into subject teaching: Commitment, constraints, caution, and change. Journal of Curriculum Studies, 37, 155–192. doi:10.1080/0022027032000276961.
Holzberger, D., Philipp, A., & Kunter, M. (2013). How teachers’ selfefficacy is related to instructional quality: A longitudinal analysis. Journal of Educational Psychology, 105(3), 774–786. doi:10.1037/a0032198.
Hsiao, L., Lin, Y.R., & Tu, Y.L. (2010, December). Gender differences in computer experience and computer selfefficacy among high school teachers. Paper presented at The Second Asian Conference on Education 2010, Osaka, Japan. http://120.107.180.177/1832/9901/099207p.pdf. Accessed 20 Oct 2015
Jung, M., & Carstens, R. (Eds.). (2015). ICILS 2013 user guide for the international database. Amsterdam: IEA. doi:10.15478/uuid:73a9f0187b644299affcdc33fe57f3e1.
Klassen, R. M., & Tze, V. M. C. (2014). Teachers’ selfefficacy, personality, and teaching effectiveness: A metaanalysis. Educational Research Review, 12, 59–76. doi:10.1016/j.edurev.2014.06.001.
Koehler, M., & Mishra, P. (2009). What is technological pedagogical content knowledge (TPACK)? Contempary Issues in Technology and Teacher Education, 9(1), 60–70.
Kreijns, K., Vermeulen, M., Kirschner, P. A., van Buuren, H., & Van Acker, F. (2013). Adopting the integrative model of behavior prediction to explain teachers’ willingness to integrate ICT in their pedagogical practices: A perspective for research on teachers’ ICT usage in pedagogical practices. Technology Pedagogy and Education, 22, 55–71. doi:10.1080/1475939X.2012.754371.
Kuiper, E., Volman, M., & Terwel, J. (2005). The web as an information resource in K12 education: Strategies for supporting students in searching and processing information. Review of Educational Research, 75, 285–328. doi:10.3102/00346543075003285.
Lee, J., Cerreto, F., & Lee, J. (2009). Teachers’ intentions toward technology usage: Do different uses lead to different determinants? In I. Gibson, R. Weber, K. McFerrin, R. Carlsen, & D. Willis (Eds.), Proceedings of the society for information technology & teacher education international conference 2009 (pp. 979–986). Chesapeake: Association for the Advancement of Computing in Education (AACE).
Little, T. D., Bovaird, J. A., & Widaman, K. F. (2006). On the merits of orthogonalizing powered and product terms: Implications for modeling interactions among latent variables. Structural Equation Modeling: A Multidisciplinary Journal, 13(4), 497–519.
Mac Callum, K., Jeffrey, L., & Kinshuk, A (2014). Comparing the role of ICT literacy and anxiety in the adoption of mobile learning. Computers in Human Behavior, 39, 8–19. doi:10.1016/j.chb.2014.05.024.
Marsh, H. W., Hau, K., & Grayson, D. (2005). Goodness of fit evaluation in structural equation modeling. In A. MaydeuOlivares & J. McArdle (Eds.), Contemporary psychometrics (pp. 275–340). Mahwah: Erlbaum.
Marsh, H. W., Hau, K.T., Wen, Z., Nagengast, B., & Morin, A. J. S. (2014). Moderation. In T. D. Little (Ed.), The Oxford handbook of quantitative methods (Vol. 2, pp. 361–386). New York: Oxford University Press.
Millsap, R. E. (2011). Statistical approaches to measurement invariance. New York: Routledge.
Morris, M. G., & Venkatesh, V. (2000). Age differences in technology adoption decisions: Implications for a changing workforce. Personnel Psychology, 53, 375–403. doi:10.1111/j.17446570.2000.tb00206.x.
Mumtaz, S. (2000). Factors affecting teachers’ use of information and communications technology: A review of the literature. Journal of Information Technology for Teacher Education, 9, 319–341. doi:10.1080/14759390000200096.
Muthén, B., & Muthén, L. (1998–2015). Mplus 7.3 [Computer Software]. Los Angeles, CA: Muthén & Muthén.
Niederhauser, D. S., & Perkmen, S. (2010). Beyond selfefficacy: Measuring preservice teachers’ instructional technology outcome expectations. Computers in Human Behavior, 26(3), 436–442. doi:10.1016/j.chb.2009.12.002.
Norwegian Directorate for Education and Training. (2012.) Framework for basic skills. http://www.udir.no/PageFiles/66463/FRAMEWORK_FOR_BASIC_SKILLS.pdf?epslanguage=no. Accessed 28 Oct 2015
O’Bannon, B. W., & Thomas, K. (2014). Teacher perceptions of using mobile phones in the classroom: Age matters! Computers Education, 74, 15–25. doi:10.1016/j.compedu.2014.01.006.
O’Mara, A. J., Marsh, H. W., Craven, R. G., & Debus, R. (2006). Do selfconcept interventions make a difference? A synergistic blend of construct validation and metaanalysis. Educational Psychologist, 41, 181–206. doi:10.1207/s15326985ep4103_4.
Ong, C.S., & Lai, J.Y. (2006). Gender differences in perceptions and relationships among dominants of elearning acceptance. Computers in Human Behavior, 22, 816–829. doi:10.1016/j.chb.2004.03.006.
Pajares, F., & Schunk, D. H. (2001). Selfbeliefs and school success: Selfefficacy, selfconcept, and school achievement. In R. Riding & S. Rayner (Eds.), Selfperception (pp. 239–266). London: Ablex.
Pamuk, S., & Peker, D. (2009). Turkish preservice science and mathematics teachers’ computer related selfefficacies, attitudes, and the relationship between these variables. Computers Education, 53, 454–461. doi:10.1016/j.compedu.2009.03.004.
Rhemtulla, M., BrosseauLiard, P. E., & Savalei, V. (2012). When can categorical variables be treated as continuous? A comparison of robust continuous and categorical SEM estimation methods under suboptimal conditions. Psychological Methods, 17, 354–373.
Sang, G., Valcke, M., van Braak, J., & Tondeur, J. (2010). Student teachers’ thinking processes and ICT integration: Predictors of prospective teaching behaviors with educational technology. Computers Education, 54, 103–112. doi:10.1016/j.compedu.2009.07.010.
Sass, D. A., Schmitt, T. A., & Marsh, H. W. (2014). Evaluating model fit with ordered categorical data within a measurement invariance framework: A comparison of estimators. Structural Equation Modeling: A Multidisciplinary Journal, 21, 167–180.
Schepers, J., & Wetzels, M. (2007). A metaanalysis of the technology acceptance model: Investigating subjective norm and moderation effects. Information Management, 44, 90–103. doi:10.1016/j.im.2006.10.007.
Scherer, R., & Siddiq, F. (2015a). The BigFishe LittlePondEffect revisited: Do different types of assessments matter? Computers Education, 80, 198–210. doi:10.1016/j.compedu.2014.09.003.
Scherer, R., & Siddiq, F. (2015b). Revisiting teachers’ computer selfefficacy: A differentiated view on gender differences. Computers in Human Behavior, 53, 48–57. doi:10.1016/j.chb.2015.06.038.
Scherer, R., Siddiq, F., & Teo, T. (2015). Becoming more specific: Measuring and modeling teachers’ perceived usefulness of ICT in the context of teaching and learning. Computers Education, 88, 202–214. doi:10.1016/j.compedu.2015.05.005.
Scheuermann, F., & Pedró, F. (Eds.). (2009). Assessing the effects of ICT in education—indicators, criteria and benchmarks for international comparisons. Luxembourg: Publications Office of the European Union. doi: 10.2788/27419.
Schibeci, R., Lake, D., Phillips, R., Lowe, K., Cummings, R., & Miller, E. (2008). Evaluating the use of learning objects in Australian and New Zealand schools. Computers Education, 50, 271–283. doi:10.1016/j.compedu.2006.05.006.
Shapka, J. D., & Ferrari, M. (2003). Computerrelated attitudes and actions of teacher candidates. Computers in Human Behavior, 19, 319–334. doi:10.1016/S07475632(02)000596.
Siddiq, F., & Scherer, R. (2015, June). How teachers emphasize the development of students’ digital information and communication skills: Looking at 21st Century education. Paper presented at the 6th IEA International Research Conference (IRC), Capetown, South Africa.
Siddiq, F., Scherer, R., & Tondeur, J. (2016). Teachers’ emphasis on developing students’ digital information and communication skills (TEDDICS): A new construct in 21st century education. Computers Education, 92–93, 1–14. doi:10.1016/j.compedu.2015.10.006.
Sieverding, M., & Koch, S. C. (2009). (Self) Evaluation of computer competence: How gender matters. Computers Education, 52, 696–701. doi:10.1016/j.compedu.2008.11.016.
Skaalvik, E., & Skaalvik, S. (2007). Dimensions of teacher selfefficacy and relations with strain factors, perceived collective selfefficacy, and teacher burnout. Journal of Educational Psychology, 99, 611–625. doi:10.1037/00220663.99.3.611.
Skrondal, A., & Laake, P. (2001). Regression among factor scores. Psychometrika, 66, 563–575.
Teo, T. (2008). Preservice teachers’ attitudes towards computer use: A Singapore survey. Australasian Journal of Educational Technology, 24, 413–424.
Teo, T. (2014). Unpacking teachers’ acceptance of technology: Tests of measurement invariance and latent mean differences. Computers Education, 75, 127–135. doi:10.1016/j.compedu.2014.01.014.
Tondeur, J., van Keer, H., van Braak, J., & Valcke, M. (2008). ICT integration in the classroom: Challenging the potential of a school policy. Computers Education, 51, 212–223. doi:10.1016/j.compedu.2007.05.003.
TschannenMoran, M., & Hoy, A. W. (2007). The differential antecedents of selfefficacy beliefs of novice and experienced teachers. Teaching and Teacher Education, 23(6), 944–956. doi:10.1016/j.tate.2006.05.003.
Van de Schoot, R., Lugtig, P., & Hox, J. (2012). A checklist for testing measurement invariance. European Journal of Developmental Psychology, 9(4), 486–492.
Vanderlinde, R., Aesaert, K., & van Braak, J. (2014). Institutionalised ICT use in primary education: A multilevel analysis. Computers Education, 72, 1–10. doi:10.1016/j.compedu.2013.10.007.
Venkatesh, V., Morris, M. G., Davis, M. G., & Davis, F. D. (2003). User acceptance of information technology: Toward a unified theory. MIS Quartely, 27, 425–478.
Volman, M., & van Eck, E. (2001). Gender equity and information technology in education: The second decade. Review of Educational Research, 71, 613–634. doi:10.3102/00346543071004613.
Yang, Y., & Green, S. (2011). Coefficient alpha: A reliability coefficient for the 21st century? Journal of Psychoeducational Assessment, 29, 377–392. doi:10.1177/0734282911406668.
Yildirim, S. (2000). Effects of an educational computing course on preservice and inservice teachers: A discussion and analysis of attitudes and use. Journal of Research on Computing in Education, 32(3), 479–495.
Authors’ contributions
FS prepared the data, participated in the process of developing a rationale, drafted the background and discussion sections of the manuscript, and revised earlier versions of the manuscript. RS led the modeling process, drafted the methods and results sections, and revised earlier versions of the manuscript. Both authors read and approved the final manuscript.
Acknowledgements
Special thanks to the Norwegian ICILS 2013 Group for their support in providing the data.
Competing interests
We confirm that this manuscript has not yet been published elsewhere and is not under consideration by another journal. All authors have approved the manuscript and agree with its submission to Largescale Assessments in Education. Furthermore, the authors accept the copyright information and the Springer author’s rights. The authors declare that they have no competing interests.
Author information
Additional information
Fazilat Siddiq and Ronny Scherer contributed equally to this work
Additional file
40536_2016_32_MOESM1_ESM.docx
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Received
Accepted
Published
DOI
Keywords
 Age and gender differences
 Computer selfefficacy
 Emphasis on digital skills [TEDDICS]
 ICILS 2013
 Moderation
 Teachers’ ICT integration