Abstract
Training programs on digital competences for teachers are often carried out without a thorough diagnosis, so it is difficult to accurately identify which digital competences need to be developed in teachers. Best-case scenario, questionnaires are proposed as the only diagnostic tool, and the answers are often based on the subjectivity of respondents which inevitably led to contradictions as revealed by the literature. In this study, a machine learning algorithm has been used to determine the teachers’ needs in terms of detecting which and how much they need to improve their digital competences. 181 teachers took a 4-week-virtual course containing two modules (Content Creation and School Management) to assess specific teacher competences.
Results showed that the type of question used in the evaluations influences the academic performance of the participants. Data were analyzed based on 1) three types of questions (matching, multiple choice and true/false) and 2) scores obtained by the participants at the end of the course.
Firstly, matching questions reflect a greater distribution in the results, which indicates that this type of question may be more challenging, followed by the multiple-choice questions. While true or false questions should not be included as they leave too much room for chance.
Secondly, performance level reflects that demographic information (e.g. age, type of institution, and level at which they teach) is associated with final performance of digital competences. This type of analytics can provide a much more detailed overview in terms of detecting strengths and weaknesses in the development of digital competences of teachers.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
It is widely acknowledged that digital literacy is a must in current times. Yet, this field has received scant attention despite that teachers must constantly be upgrading their competences to meet the evolving demands of digital education systems [1]. Among all professions, training of digital competences for preservice teachers is considered the lowest of all [2]. Consequently, it is becoming increasingly difficult to ignore that teachers need the necessary set of skills to integrate technology in their daily practices, mainly when students are developing their own digital expertise and learning from early ages, see e.g. [3,4,5]. Trends show us an increasing number of technologies which are waiting for well-trained teachers to be implemented in the 21st century classroom [6]. In the coming section, a review of the literature is presented to scope what has been done in the field of teachers’ digital competences.
1.1 Digital Competence
The term digital competence is as elusive as evolving due to, among others, the rapid social and technological changes. Apart from some confusion about its conceptualization (competence, competency and competencies) [7], no one can deny that they are relevant to the educational field nowadays. Thus, Oberländer et al. [8] refer to digital competencies as “a set of basic knowledge, skills, abilities, and other characteristics that enable people…to efficiently and successfully accomplish their job tasks regarding digital media at work” whereas Ferrari [9] conceptualized digital competence as ‘the set of knowledge, skills, attitudes, abilities, strategies and awareness that are required when using ICT [information and communication technologies] and digital media to perform tasks; solve problems; communicate; manage information; collaborate; create and share content; and build knowledge effectively, efficiently, appropriately, critically, creatively, autonomously, flexibly, ethically, reflectively for work, leisure, participation, learning and socialising’ (p. 30).
Yet, there are few mechanisms to differentiate which digital competences teachers need to focus on. The most common tools are perception surveys and some frameworks intended to capture the nature of indicators in the complex field of digital competences. Although useful, both mechanisms heavily rely on personal opinions and, as such, perceptions regarding one’s own competencies might be prone to subjectivity. Moreover, respondents may be confronted to choose between what they want instead of what they need to develop around their digital competences. In consequence, there are divergent results in early and primary education teachers who are the focus of the current study. Betancourt-Odio et al. [10] reported that 427 teachers from 15 Ibero-american countries found their digital competences ineffective to cope with changing situations (such as those of the Covid-19 pandemic). Even more, when digital competences are compared between levels, early childhood educators seem to have a better performance than their counterparts in primary education [11]. Sometimes student teachers attribute digital skills deficiency to their professors [12] or the outdated university programs [13]. That judgement may be because, in some contexts, teaching the basics of a program (e.g. a word processor) is considered to be enough, denying thus the utilization of other interactive resources that may help in the teaching tasks [14].
Even if digital competencies were identified, teachers usually use them inconsistently throughout the teaching process or with more emphasis in some subjects than others [15] which, again, widen the existing gaps in the field of teachers’ digital competencies. In this regard, some local attempts, e.g. in Slovakia, to improve curricula are usually made, but again, they are based on subjectiveness [16].
On the other hand, there are several frameworks which have tried to standardize digital competences of teachers. Of the most seven used, the European framework for the digital competence of educators or DigCompEdu [17] has been considered the most favored by specialists which, of course, does not disregard the usefulness of other frameworks [18]. In addition to the obvious bias of self-perception, a common disadvantage is that each framework do not always point to the same indicators which has led researchers to take their own pathway through scale validations, see [19].
When applying DigCompEdu, it is usually the case that early childhood educators obtain the lowest scores [20] which contradicts what other studies found, e.g. [2, 7]. But this is not the only contradiction. There are more inconsistencies regarding the digital competences of early childhood educators. Despite acknowledging the importance of digital competences, student teachers in early childhood education believe that little these competences can help to their career and, although they recognize that ICTs may have a motivator effect on children, they do not always want to be trained in digital competences [21]. This perception is similar when other types of surveys are applied to early childhood education students [22]. However, in countries such as Portugal, the level of digital competences in teachers at primary and secondary level seems to be moderate while the pedagogic and student management competences are considered to be low [23, 24].
Another less frequent framework is the Common Framework for Teachers’ Digital Competence (MCCDD, acronym in Spanish) which is supported by the Spain’s National Institute of Educational Technologies and Teacher Training (INTEF). Supporters of this framework assure that it improves the digital competences in teachers mainly when it is combined with active methods [25]. Nevertheless, regardless which framework is taken into account, they all either use self-reporting instruments or are not adapted to countries in Latin America but the specific European contexts.
An additional oversighted component of every training program is the type of assessment used to evaluate digital literacy in early childhood educators. The next section deals with the type of assessment questions used in training programs for digital competences in education.
1.2 Types of Assessment Questions
The type of assessment may play a significant role in the teacher training of digital competences. Given the wide variety of questions to be used during a training program, stakeholders need to choose the best options. Some argue that formative assessment guided by a system can help trainees to better learn [26] as it has a positive influence on students due to its engaging nature [27]. Likewise, Draskovic et al. [28] calls our attention to the fact that students benefit from a more efficient method of performing tasks when using digital assessment items compared to pen and paper. Although, teachers may initially feel a little overloaded, the authors assert that teachers found this type of items easier to administrate and score than those of pen and paper. While there is some discussion about the fact that the type of question does not greatly affect the learning outcomes [29] others give special importance to certain types of questions. In this way, contemporary pedagogical models and their implementation in ICT mediating tools show a tendency towards quantitative and summative evaluation [30]. These researchers found that commonly used learning management systems (LMS) include a variety of quantitative assessment questions, with multiple choice questions being the most common.
On this subject, Mafinejad et al. [31] found that scores for multiple choice questions are significantly related to total skill scores while others, e.g. [32], see that multiple choice questions have the drawback that the student can choose the correct answer by guessing, without following the proper procedure to solve a problem. Where most research seems to agree is on the fact that there is practically no difference between text-based questions and multiple choice questions [33].
As described above, this section has outlined the importance of digital competences in teachers and how these competences are assessed. The evidence presented thus far suggests that discrepancies found among studies in the field of teachers’ digital competences may emerge from the subjective nature of the instruments to assess. As a way to find another alternative outside the realm of self-reporting instruments, we launched a training course to assess digital competences in the very field through learning analytics. Based on the literature review, we hypothesized that assessment of digital competences is also threatened by the type of questions used to assess those digital competences. Accordingly, this study set out to apply a highly reliable algorithm 1) to identify how the type of question affected assessment of digital competences. Then, 2) to determine digital competences of teachers based on their performance.
2 Methodology
2.1 Research Design
This study adopted a quantitative non-experimental cross-sectional design without any manipulation of the variables which were measured at a given time [34]. Based on this, our research followed four differentiated and sequential stages: (a) introduction, which aimed at analyzing previous gaps; (b) methodological design, in which we selected research tools and techniques, study sample, and the data analysis method; (c) results, which showcased figures and tables of findings; finally (d) discussion, to compare and contrast our findings with previous studies.
2.2 Participants
181 early childhood educators from different provinces in Ecuador were enrolled in our course. Their age ranged from 22–59 years old. Of them, the largest subgroup (59%) was made of teachers between 22–38 years old followed by a smaller subgroup (36%) with ages between 39–53 years and finally the smallest subgroup (3%) with an age between 54 -59 years old.
They all came from 12 out of the 24 provinces of Ecuador, being Pichincha the most representative (94%), followed by Azuay, Loja, Santo Domingo (3%) and finally the other provinces (1%). Concerning the type of institution, 32% reported to work in public schools, 29% in private institutions, 34% were not working at that time and 5% worked in charter schools.
Regarding the educational level in which they taught, 36% worked with children from 5–6 years old, 24% with children from 4–5 years old, 22% with children from 3–4 years old, 11% with children from 7 -8 years and 6% work with children between 5–6 years. For this study, the results of the first 59 teachers who finished the course were taken to avoid noise in the results.
2.3 Instruments and Procedures
For data collection, a virtual training course was planned, designed and implemented in an institutional Learning Management System (LMS). This course was called How to teach children through the Internet (in Spanish), and aimed at early childhood education teachers. The course was made of 3 blocks whose contents are detailed below (Table 1):
For the purposes of this study, only blocks 2 and 3 were taken into account since block 1 was introductory and focused on capturing the attention of the participants rather than evaluating their technological skills. Hereafter, block 2 will be named Module 1 (Digital Content Creation) and block 3 will be named Module 2 (School Management).
Advertising of the virtual course was carried out through the institutional Facebook and registrations were open for two weeks during the height of the Covid-19 pandemic in 2021. At the end of the registrations, 181 teachers from different provinces of Ecuador were enrolled. Once the course was finished, the records of the grades of each student were retrieved from the online platform database.
2.4 Data Analysis
For data analysis we used a data mining approach. Data mining combines several technologies and theories (e.g. artificial intelligence, mathematical statistics, data visualization) to extract and link useful information to make scientific decisions more efficient [35].
Therefore, analyses were carried out in Python 3 [36] along with Pandas [37] (for tabular data management), numpy [38] (for array management), seaborn [39] (to make graphics) and scikit-learn [40] (for clusterization). Data were cleaned prior to descriptive and multivariate analyses by removing non-relevant information. Additionally, demographic data were coded according to province, city, institution in which teachers worked and the educational level in which they taught.
Data Analysis 1: How the Type of Question Affected Assessment
Data related to assessment were classified according to the type of question (multiple choice, true/false, matching). To determine which questions were easy or difficult for teachers, an analysis was carried out regarding the type of questions. To this end, the average for each participant grouped according to the type of question was obtained by following these steps: 1) data preprocessing, words were all turned to lowercase, spaces were changed to hyphens, special characters and accents were removed. 2) All scores were scaled to avoid extreme values. One of the main purposes of scaling is to make all datapoints be comparable, so we used the MaxAbsScaler method. 3) In order to compare the impact of the three types of questions: MC (multiple choice), T/F (True/False) and M (matching), the mean of the scores grouped by question type was obtained, to create three variables: MC_score, T/F_score and M_score (See Table 2).
Data Analysis 2: Digital Competences of Teachers
To determine the teachers' academic performance in terms of their digital competences, we applied k-means clustering which is an unsupervised classification method that groups individuals into subgroups based on their characteristics [41]. Additionally, the “k-means clustering algorithm is considered one of the most powerful and popular data mining algorithms in the research community” [42] (p. 1) and it has proven its value in grouping large amounts of data in virtual courses [43]. The value of k is fundamental in carrying out an adequate clustering, so the elbow method was used to determine the value of k. To get k-number of groups, it was necessary to compute the square of the distance between the sample points in each group and the center of the cluster to give a series of k values [41] (p. 2). Thus, the optimal number of groups was 3. In addition, the XGBoost regression model was used to determine the attributes that contributed the most to the final score of participants. i.e. we used XGBoost to make a feature selection of the most important characteristics of individuals.
3 Results
3.1 How the Type of Question Affected Assessment
The first reported results are based on the scores grouped by question type: MC (mul-tiple choice), T/F (True/False) and M (matching).
Table 2 shows the mean of the scores obtained by the participants with respect to the type of question. It can be seen that each score is slightly different, with TF_score being the highest, followed by MC_score and M_score.
Figure 1 and Table 2 show that matching questions have a greater distribution of scores, which tells us that, for students, this type of question could be a little more complicated or that, by having a greater rating range, the information is more variable.
Additionally, the Pearson correlation coefficient was used to determine a linear relationship between age, module 1, module 2, final score, MC score and M_score.
Table 3 indicates that there is a slight negative correlation (-0.2063) between age and module 1 (Digital Content Creation), which suggests that the older the age the lower the score in module 1 (Digital Content Creation). Conversely, a slight positive correlation (0.1711) between age and module 2 (School Management) which makes sense as the older the age the higher the score in module 2 (School Management). Likewise, a marginal positive correlation between module 1 (Creation of Digital Content) and module 2 (School Management) (0.3042) meaning that the higher the score in module 1 (Creation of Digital Content), the higher the score in module 2 (School Management).
3.2 Digital Competences of Teachers
As there were multiple digital competences assessed throughout the two modules of the course, a rigorous procedure that objectively grouped teachers according to the mastery of those competences was necessary. Therefore, the k-means algorithm was used as explained in Sect. 2.4. From such analyses, we derived 3 groups of teachers with similar characteristics.
Henceforth, these 3 groups will be called performance groups, which present the following characteristics according to the type of institution and educational level to which teachers belong (Figs. 2 and 3).
Group 0: It is mostly made of teachers who work in public institutions and teach at first grade level (children 5–6 years old).
Group 1: It is mostly made of teachers who are not currently working but teach at first grade level (children 5–6 years old).
Group 2: It is mostly made of teachers who work in private institutions and teach at early level 1 (children 3–4 years old).
Performance Groups According to Their Final Scores
Table 4 shows that group 0 (public teachers-teaching children 5–6 years old) obtained the lowest mean (51.08 points) and approximately 50% of the participants obtained a score of 52.00 points out of 60. Group 1 (teachers who are not currently working - teach children 5–6 years old) obtained the highest mean (56.98 points) and approximately 50% of them obtained a score of 56.67 points out of 60. Group 2 (private teachers - teaching children 3–4 years old) obtained an average of 52.15 and approximately 50% of the participants obtained a score of 52.33 points out of 60.
The maximum score for each module was 30 points. So, Fig. 4 shows that, in Module 1 (Digital Content Creation), the scores of group 0 (public teachers- teaching children of 5–6 years) are more dispersed compared to group 1 (teachers who are not working- teach to children aged 5–6 years) which are more compact. This means that the level of knowledge/proficiency in group 0 is more heterogeneous which could imply a wider variety in terms of training in contrast to group 1 which has a more homogeneous level of digital competences.
On the other hand, Fig. 5 shows that, in Module 2 (School Management), the grades of group 0 (public teachers- teaching children of 5–6 years) are more scattered than those of group 2 (private teachers - teaching children 3–4 years old). This means that group 0 showed greater heterogeneity in terms of mastering the school management skills, which implies that this group may need a wider variety of training courses in the near future.
Figure 6 reveals that group 1 (teachers who are not currently working and teach children aged 5–6 years) in green performed better academically than the other groups in both module 1 (Digital Content Creation, left) and module 2 (School Management, center) as well as in the final result of the course (right).
Figure 7 shows comparatively, in greater detail, the academic performance in both modules (Digital Content Creation and School Management). Group 0 (in blue) had a higher score in module 1 (Digital Content Creation) but their performance is low in module 2 (School Management) whereas group 1 (in orange) obtained the highest grade in the two modules and group 2 (in green) obtained a low grade in module 1 (Digital Content Creation) but their score went up in module 2 (School Management).
4 Discussion and Conclusion
Assessment of digital competences in teachers is often carried out through self-reporting instruments which may usually induce biased or contradictory outcomes as reported by the literature (see Sect. 1.1). However, our study set out a different approach by applying machine learning techniques on the digital competence behaviors of early childhood educators who took an online training course about how to teach children through the Internet. The course in itself was intended to develop and assess digital competencies in participants at the end of it. Yet we used learning analytics to unveil further details.
The results of this study have shown that groups of teachers sharing similar characteristics also share some common digital competences. The k-means algorithm grouped all participating teachers in three clusters: group 0 (public teachers-teaching children of 5–6 years), group 1 (teachers who are not working-teach to children aged 5–6 years) and group 2 (private teachers - teaching children 3–4 years old). Of all these three groups of teachers, group 1 showed the highest performance in the competences of the two modules of the course: Module 1 (Digital Content Creation) and Module 2 (School Management). While group 0 scored second place in Module 1 (Digital Content Creation) and group 2 obtained second place in Module 2 (School Management). That a group of teachers who are not currently working had obtained the highest performance may indicate that teacher training should not take place during periods of high workload.
It was also found that age seems to be associated with digital competencies i.e. the younger the teachers the higher their digital competencies which, in turn, support those initiatives to include such skills in the curricula of pre-service teachers. Comparing these findings with those of other studies, e.g. [44,45,46,47], confirm age as predictor of the mastery of digital competences. Regarding the level where teachers work (which seems to be another predictor of teacher competencies), our research showed consistency with past studies, e.g. [48, 49].
The second major finding was that the type of question does influence the outcome of assessment, in this case, of teacher digital competences. Matching questions are the ones with the greatest dispersion in their data, followed by the Multiple Choice and T/F questions, which shows that the Matching questions and Multiple-Choice questions are more complex. The max and min score in the T/F questions do not have any variation, which reveals that these questions have much less complexity. This indicates that matching and multiple choice questions seem to be most suitable as they can cover a wider range of knowledge in terms of content as stated by previous studies, e.g. [50, 51], and can discourage students to guess the correct answer as contrarily stated by [32].
In conclusion, teacher training should be organized according to diagnoses that reveal the most important variables to be intervened. This research shows that the type of question used in training courses is key, as well as the level of education to which it is directed and, to a lesser degree, the age of the teachers. In other words, teacher training on technological teacher competencies must be adjusted to the type of institution and educational level in which teachers have worked recently, considering the existing weaknesses. An implication of this is the possibility for institutions to design digital literacy training programs for their teachers according to these specifics.
Taken together, these results suggest that there are much more objective ways to assess teacher digital competences than sole self-reporting instruments. As with all such studies, there are limitations that offer opportunities for further research. Further work containing a larger sample size is required in order to reach more generalizable results.
References
Davydenko, V., Kuznetsova, A., Chichik, N.: Impact of digitalization trends on the development of the Russian system of advanced training and methodological support for teachers (2021)
Senkbeil, M., Ihme, J.M., Schöber, C.: Dissemination of media literacy at school in a digital world: are teacher candidates digitally competent? Psychol. Erzieh. Unterr. 68(1), 4–22 (2021)
Quinga, Y., Pilataxi, N., Carvajal, V., Ocaña, M.: Virtual activities to strengthen basic math skills in children. In: Botto-Tobar, M., Cruz, H., Díaz Cadena, A., Durakovic, B. (eds.) Emerging Research in Intelligent Systems. CIT 2021. Lecture Notes in Networks and Systems, vol. 406. Springer, Cham (2022). https://doi.org/10.1007/978-3-030-96046-9_13
Vega, B., Velasco, M., Ocaña, M., Rebeca, M. (2022). Scratchjr visual programming language for early math skills development in 4–7 years old children. In: Botto-Tobar, M., Cruz, H., Díaz Cadena, A., Durakovic, B. (eds.) Emerging Research in Intelligent Systems. CIT 2021. Lecture Notes in Networks and Systems, vol. 406. Springer, Cham. https://doi.org/10.1007/978-3-030-96046-9_19
Albán Bedoya, I. and M. Ocaña-Garzón. Educational Programming as a Strategy for the Development of Logical-Mathematical Thinking. 2022. Cham: Springer International Publishing
Alexander, B., et al., Educause Horizon report: 2019 Higher Education edition. 2019
Moore, D.R., Cheng, M.I., Dainty, A.R.J.: Competence, competency and competencies: performance assessment in organisations. Work Study 51(6), 314–319 (2002)
Oberländer, M., Beinicke, A., Bipp, T.: Digital competencies: a review of the literature and applications in the workplace. Comput. Educ. 146, 103752 (2020)
Ferrari, A., Digital competence in practice: An analysis of frameworks, in JRC IPTS. 2012, Joint Research Centre of the European Commission: Sevilla. p. 82116
Betancourt-Odio, M.A., et al.: Self-perceptions on digital competences for M-learning and education sustainability: a study with teachers from different countries. Sustainability (Switzerland) 13(1), 1–12 (2021)
Urrea-Solano, M., et al.: The learning of e-sustainability competences: A comparative study between future early childhood and primary school teachers. Educ. Sci. 11(10), 644 (2021)
Gómez-Trigueros, I.M., de Atalaya, S.P.L., Ros, R.D.: Towards an insertion of technologies: the need to train in digital teaching competence. Int. Multi. J. Soc. Sci. 10(3), 64–87 (2021)
Záhorec, J., Hašková, A., Munk, M.: Teachers’ professional digital literacy skills and their upgrade. Eur. J. Contemp. Educ. 8(2), 378–393 (2019)
Záhorec, J., Hašková, A., Munk, M.: Integration of Digital Technology Applications into the Pre-gradual Teacher Training, pp. 892–901 (2020)
Záhorec, J., Nagyová, A., Hašková, A.: Teachers’ attitudes to incorporation digital means in teaching process in relation to the subjects they teach. Int. J. Eng. Pedagogy 9(4), 100–120 (2019)
Záhorec, J., Hašková, A., Munk, M.: Curricula design of teacher training in the area of didactic technological competences. In: Auer, M.E., Guralnick, D., Simonics, I. (eds.) ICL 2017. AISC, vol. 716, pp. 383–393. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-73204-6_43
Redecker, C.: European framework for the digital competence of educators: DigCompEdu., Joint Research Centre (Seville, Spain) (2017)
Cabero-Almenara, J., Romero-Tena, R., Palacios-Rodríguez, A.: Evaluation of teacher digital competence frameworks through expert judgement: the use of the expert competence coefficient 9(2), 19 (2020)
Barragán-Sánchez, R., et al.: Teaching digital competence and eco-responsible use of technologies: development and validation of a scale. Sustainability (Switzerland) 12(18), 7721 (2020)
García-Vandewalle García, J.M., et al.: Analysis of digital competence of educators (DigCompEdu) in teacher trainees: the context of Melilla, Spain. Technology, Knowledge and Learning (2021)
Pérez-Jorge, D., et al.: Training in digital skills in early childhood: education teachers the case of the university of la laguna. Int. J. Interact. Mob. Technol. 14(20), 35–49 (2020)
Romero-Tena, R., et al.: The challenge of initial training for early childhood teachers. A cross sectional study of their digital competences. Sustainability (Switzerland) 12(11), 4782 (2020)
Dias-Trindade, S., Moreira, J.A.: Assessment of high school teachers on their digital competences. Magis 13, 01–21 (2020)
Dias-Trindade, S., Moreira, J.A., Ferreira, A.G.: Evaluation of the teachers’ digital competences in primary and secondary education in Portugal with digcompedu checkin in pandemic times. Acta Scientiarum - Technology 43, e56383–e56383 (2021)
Romero-García, C., Buzón-García, O., de Paz-Lugo, P.: Improving future teachers’ digital competence using active methodologies. Sustainability (Switzerland) 12(18), 7798 2020
Hettiarachchi, E., Huertas, M.A., Mor, E.: E-assessment system for skill and knowledge assessment in computer engineering education. Int. J. Eng. Educ. 31(2), 529–540 (2015)
Hettiarachchi, E., et al.: Improving student performance in high cognitive level courses by using formative e-assessment. Int. J. Technol. Enhanced Learn. 7(2), 116–133 (2015)
Draskovic, D., Misic, M., Stanisavljevic, Z.: Transition from traditional to LMS supported examining: a case study in computer engineering. Comput. Appl. Eng. Educ. 24(5), 775–786 (2016)
Tetteh, G.A., Sarpong, F.A.A.: Infuence of type of assessment and stress on the learning outcome. J. Int. Educ. Bus. 8(2), 125–144 (2015)
Torres-Madroñero, E.M., Torres-Madroñero, M.C., Botero, L.D.R.: Challenges and possibilities of ICT-mediated assessment in virtual teaching and learning processes. Future Internet 12(12), 1–20 (2020)
Mafinejad, M.K., et al.: Use of Multi-Response format test in the assessment of medical students’ critical thinking ability. J. Clin. Diagn. Res. 11(9), LC10-LC13 (2017)
Jaquez, J., et al.: TecEval: an on-line dynamic evaluation system for engineering courses available for web browsers and tablets (2015)
Jahn, G.: Smooth transition from text-based exams to multiple-choice, pp. 448–453 (2022)
Hernández-Sampieri, R., Torres, C.P.M.: Metodología de la investigación, vol. 4 (2018). McGraw-Hill Interamericana México^ eD. F DF
Pan, L.: A big data-based data mining tool for physical education and technical and tactical analysis. Int. J. Emerg. Technol. Learn. 14(22), 220–231 (2019)
Van Rossum, G., Drake, F.L.: Python 3 reference manual (2007). CreateSpace
McKinney, W.: Data structures for statistical computing in python. In: Proceedings of the 9th Python in Science Conference, Austin, TX (2010)
Harris, C.R., et al.: Array programming with NumPy. Nature 585(7825), 357–362 (2020)
Waskom, M.L.: Seaborn: statistical data visualization. J. Open Source Softw. 6(60), 3021 (2021)
Pedregosa, F., et al.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011)
Yuan, C., Yang, H.: Research on k-value selection method of k-means clustering algorithm. J 2(2), 226–235 (2019)
Ahmed, M., Seraj, R., Islam, S.M.S.: The k-means algorithm: a comprehensive survey and performance evaluation. Electronics 9(8), 1295 (2020)
Ocaña, M., Khosravi, H., Bakharia, A.: Profiling language learners in the big data era. In: ASCILITE 2019 - 36th International Conference of Innovation, Practice and Research in the Use of Educational Technologies in Tertiary Education: Personalised Learning. Diverse Goals. One Heart, pp. 237–245 (2019)
Claro, M., et al.: Teaching in a digital environment (TIDE): defining and measuring teachers’ capacity to develop students’ digital information and communication skills. Comput. Educ. 121, 162–174 (2018)
Ertl, B., Csanadi, A., Tarnai, C.: Getting closer to the digital divide: an analysis of impacts on digital competencies based on the German PIAAC sample. Int. J. Educ. Dev. 78, 102259 (2020)
Nguyen, M.H., Hunsaker, A., Hargittai, E.: Older adults’ online social engagement and social capital: the moderating role of Internet skills. Inf. Commun. Soc. 25, 1–17 (2020)
Saikkonen, L., Kaarakainen, M.-T.: Multivariate analysis of teachers’ digital information skills - the importance of available resources. Comput. Educ. 168, 104206 (2021)
Cabero-Almenara, J., et al.: Classification models in the digital competence of higher education teachers based on the DigCompEdu framework: logistic regression and segment tree. J. E-Learn. Knowl. Soc. 17(1), 49–61 (2021)
Záhorec, J., Hašková, A., Munk, M.: First results of a research focused on teachers’ didactic technological competences development. In: Auer, M.E., Tsiatsos, T. (eds.) ICL 2018. AISC, vol. 917, pp. 461–472. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-11935-5_44
Siddiqui, N.I., et al.: Contemplation on marking scheme for Type X multiple choice questions, and an illustration of a practically applicable scheme. Indian J. Pharmacol. 48(2), 114–121 (2016)
Tangianu, F., et al.: Are multiple-choice questions a good tool for the assessment of clinical competence in internal medicine? Italian J. Med. 12(2), 88–96 (2018)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Guallichico, G., Ocaña, M., Tejada, C., Bautista, C. (2023). Assessing Digital Competence Through Teacher Training in Early Education Teachers. In: Botto-Tobar, M., Zambrano Vizuete, M., Montes León, S., Torres-Carrión, P., Durakovic, B. (eds) Applied Technologies. ICAT 2022. Communications in Computer and Information Science, vol 1757. Springer, Cham. https://doi.org/10.1007/978-3-031-24978-5_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-24978-5_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-24977-8
Online ISBN: 978-3-031-24978-5
eBook Packages: Computer ScienceComputer Science (R0)