Abstract
Intelligent Tutoring Systems (ITSs) determine the quality of student responses by means of a diagnostic process, and use this information for providing feedback and determining a student’s progress. This paper studies how ITSs diagnose student responses. In a systematic literature review we compare the diagnostic processes of 40 ITSs in various domains. We investigate what kinds of diagnoses are performed and how they are obtained, and how the processes compare across domains. The analysis identifies eight aspects that ITSs diagnose: correctness, difference, redundancy, type of error, common error, order, preference, and time. All ITSs diagnose correctness of a step. Mathematics tutors diagnose common errors more often than programming tutors, and programming tutors diagnose type of error more often than mathematics tutors. We discuss a general model for representing diagnostic processes.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
More than a decade ago, VanLehn published his paper on the behaviour of Intelligent Tutoring Systems (ITSs) [60]. An ITS consists of an outer loop, which serves tasks to a student matching her progress, and an inner loop, which gives a student feedback and hints about steps she takes towards solving a task. Completing a task in an ITS often requires multiple steps, where “a step is a user interface action that the student takes in order to achieve a task” [60]. An important responsibility of the inner loop is what VanLehn calls step analysis.
Diagnosing student steps is essential for determining progress, and for giving feedback and hints. Feedback and hints are important factors supporting learning [28]. How do different ITSs diagnose a student step? We perform a systematic literature review of available step-based ITSs to classify the diagnostic processes of these systems. We determine the various components that play a role in diagnosing student steps, and study how these components are combined to perform a full diagnosis. Furthermore, we compare the diagnoses of ITSs from different domains (such as mathematics, programming, and physics), and ITSs using different approaches (such as constraint-based tutoring [50], model tracing [5], example tracing [43], and intention-based tutoring [39]). The results of our study inform the design of ITSs, and might in the future be combined with results from effectiveness studies [61] to get a better understanding of what kind of diagnostic processes are likely to be more effective.
The research question we address in this paper is: How do ITSs determine the quality of student responses? To answer this question, we will look at the aspects that can be distinguished in the diagnosis of a student response, how these aspects are combined in various ITSs, and if there are patterns or perhaps even a general scheme that can be identified in the diagnostic processes of the different ITSs. The contributions of this paper are:
-
we distinguish eight aspects that are used in various tutors in diagnosing a student step;
-
we describe patterns in combining these aspects;
-
we compare how diagnosing differs between domains and tutoring approaches.
This paper is organised as follows. Section 2 discusses related work. Section 3 describes the research method, and the resulting diagnostic aspects and processes are presented in Sect. 4. Section 5 concludes.
2 Related Work
We are not aware of research on comparing diagnostic processes of ITSs across various domains and using different tutoring approaches. In the 1970 s and later, research focused on diagnosing a particular aspect of students’ work, namely misconceptions [14]. Diagnosing misconceptions requires collecting and checking for buggy rules, which sometimes leads to overwhelming and impractical numbers of buggy rules, even for simple domains such as fractions [31]. Modern approaches, such as algorithmic debugging [68], automatically distinguish buggy rules. Heeren and Jeuring present an advanced diagnose service, which is used in ITSs for mathematics, logic, and programming [29].
Diagnosis of student steps has been studied extensively in ITSs and assessment systems for mathematics, such as Stack [54] and ActiveMath [24]. El-Kechaï et al. [21] evaluate the diagnosing behaviour of PépiMep, a diagnosis system for algebra that is part of a web-based mathematics platform. This system can distinguish 13 different patterns in student responses. Chanier et al. [16] review how errors are analysed in several ITSs for second-language learning. More related work on diagnosing student steps is described later in this paper.
3 Research Method
For our review we selected papers describing an ITS that is capable of providing feedback at the level of individual steps and that has been used in classrooms, or tested on data from real students. These inclusion criteria ensure that the ITS has an inner loop with a step analysis, and ensure ecological validity, i.e. that the ITS makes realistic diagnoses.
We searched for relevant papers in two ways. First, we considered systems discussed in three relevant reviews. Keuning et al. [41] classify the types of feedback given in programming tutors. Specifically, we included papers describing systems that are labelled as providing feedback on task-processing steps, because these papers are assumed to meet the first two criteria. VanLehn’s review on the effectiveness of tutoring systems [61] classifies systems as answer-based, step-based, or substep-based. The step-based and substep-based systems satisfy our inclusion criteria. Finally, Cheung and Slavin’s review [17] discusses the effectiveness of educational software in mathematics. From these reviews, we included 14, 17, and 0 papers, respectively (i.e. 31 papers in total). The papers in Cheung and Slavin’s review [17] did not meet the inclusion criteria, or lacked a description of the system’s working.
Second, we searched for papers using a literature search. A preliminary search in several search engines (Google Scholar, Scopus, and ERIC) revealed that Scopus produces the most relevant search results. See Van der Bent’s thesis [12] for different search terms and the resulting number of papers. We judged relevance of papers by reading the abstract and, when necessary, by skimming through the article. The search term that produced the most relevant results was
in Scopus, giving 195 papers. Using the same terms in ERIC resulted in fewer papers, largely a subset of documents found in Scopus. Searching in Google Scholar resulted in many more, but less relevant, papers. The papers found in Scopus were also found in Google Scholar. Hence, we used the 195 documents found in Scopus. Note that using the search term (step AND based) OR stepwise may have resulted in finding fewer papers from less-structured domains.
Next, we checked this initial selection of papers for the inclusion criteria. The first author read the abstracts. If the information in the abstract was insufficient to determine whether a system meets all criteria, she read the full paper. If this did not result in a decision, the second author read the paper, and discussed the paper’s relevance with the first author. The literature search resulted in 16 more papers that meet the inclusion criteria.
We categorized the ITSs described in the selected papers by their tutoring approach (model tracing, example tracing, constraint-based, or intention-based: Aleven et al. [2] explain the differences between the first three of these paradigms) and by domain. Then, starting with a small subset of papers (around 10), we iteratively designed a system for labelling the diagnostic processes and diagnosed aspects. With this labelling system we categorized the rest of the selected ITSs.
After labelling the diagnostic processes, we checked whether there are any noticeable differences between approaches or domains, by comparing the frequency at which aspects are diagnosed per approach and per domain. We also described the diagnostic processes in diagrams and tried to abstract a general model from the labelling system.
4 Diagnostic Aspects and Processes
We found 47 papers on 40 ITSs that satisfy our inclusion criteria. Table 1 gives an overview of the ITSs, including references, domain, and tutoring approach. We found 26 model tracing tutors, 8 example tracing tutors, 11 constraint-based tutors, and 1 intention-based tutor. An ITS can make use of multiple approaches, for example, Andes [63] and Mathtutor [1] use constraints in combination with the example tracing approach. We could not determine the approach used by the Technical Troubleshooting tutor [38].
Subsection 4.1 describes the diagnostic aspects we found based on a small sample of papers, which we used to label the rest of the ITSs. Subsection 4.2 describes the frequency of aspects per approach and domain. Subsection 4.3 discusses models representing the diagnostic processes of some tutoring systems, followed by a general model for diagnostic processes in Subsect. 4.4.
4.1 Diagnostic Aspects
We found that ITSs use the following aspects to diagnose a student step: correctness, difference, redundancy, type of error, common error, order, preference, and time. We explain and illustrate these aspects below. Whenever relevant, the running example will be the following algebra problem: “Solve for x: \(5x+6=7x\)”.
-
Correctness (C) determines whether or not a student step matches an expected step, or does not violate any constraint. Possible outcomes are correct and incorrect. For instance, if a student submits \(2x+6=0\), this step is diagnosed as incorrect because it does not match the expected next step \(5x-7x+6=0\). The equation \(5x+6-7x=0\) is considered correct because it is semantically equivalent to the expected answer.
-
Difference (D) is similar to correctness, in that it determines whether or not a step matches an expected step. The result is a measure such as a number or percentage that indicates the edit distance between the student step and an expected step. When the difference is zero, the step is correct. For example, if we use the edit distance, the above incorrect response results in a difference value of 1, since it requires one edit operation (replace “\(+\)” by “−”) to change the incorrect step into the expected step.
-
Redundancy (R) refers to a superfluous step: this includes steps that are too small to be recognized as a meaningful step. Possible outcomes are redundant, not redundant, and unknown. For example, the rewrite step from \(5x-7x+6=0\) into \(-7x+5x+6=0\) can be considered redundant.
-
Type of Error (ToE) refers to a classification of errors. Possible outcomes differ per problem domain or ITS. For example, \(5x-(7x+6=0\) can be classified as a syntax error.
-
Common Errors (CE) or buggy rules are misconceptions that a student may have. Possible outcomes differ per problem domain or ITS. An example of a buggy rule is forgetting to change the sign when moving an expression from one side of the equation to the other side, for instance, rewriting an expression of the form \(5x+6=7x\) into \(5x+6+7x=0\).
-
Order (O) refers to the order in which a student takes steps. Possible outcomes are correct order, incorrect order, and unknown. Note that this is a diagnosis over multiple steps.
-
Preference (P): some solutions may be preferable over others. Possible outcomes are preferred, not preferred, and unknown. For instance, in a programming tutor, a particular algorithm may produce the correct result, but be less efficient than the preferred algorithm. A teacher can express a preference for pedagogical reasons, if she wants students to use a particular approach rather than another.
-
Time (T) refers to the time a student takes to submit a step or solve a problem, measured in (milli)seconds. This aspect was only labelled when time was used for diagnostic purposes. While many systems measure time, only few use it for diagnostic purposes.
Table 1 gives an overview of the diagnosed aspects per ITS.
Of the eight aspects that ITSs diagnose, correctness is the most common aspect, and is used in all systems. Most other aspects depend on its outcome. For example, type of error relies on correctness, because errors can only be found in steps that are known to be incorrect. Likewise, preference also depends on correctness, because it can only determine preference between correct steps. Aside from correctness, the most commonly diagnosed aspects are the type of error and common errors.
Only one ITS [67] diagnoses time with the assumption that the time it takes to answer a question reflects the difficulty of the question. Why are other ITSs not diagnosing time? Most ITSs can be accessed at home, without supervision. This makes it difficult to monitor how much time is actually spent on answering a question. For example, a student might take a long time to answer because she is taking a break or doing something else. Perhaps this is why most ITSs do not use time for diagnosis.
4.2 Diagnostic Aspects per Approach and Domain
We distinguish four ITS approaches: model tracing (mt), example tracing (ex), constraint-based (cb), and intention-based (ib). There is some overlap between these categories: five ITSs combine model tracing and the constraint-based approach, and one ITS (Mathtutor) uses example tracing and the constraint-based approach. Only one ITS uses the intention-based approach. Table 2 (left-hand side) shows the frequency of the occurrence of aspects in the various ITS approaches. The results do not show very different patterns for the approaches.
The ITSs we study deal with tasks in a large variety of problem domains. At an abstract level, we can group them into four domains: mathematics, programming, physics, and other domains. Mathematics includes topics such as algebra, arithmetic, and geometry. Programming includes programming in specific languages, and more general topics such as object-oriented design and data structures. Physics includes qualitative physics and statics. The remaining ITSs involve topics such as botany, foreign language pronunciation, database design, and aircraft engineering. The domains partially overlap. (Why2-)Autotutor is in both the physics and ‘other domains’ category, because it teaches both physics and computer literacy. iList is in both the programming and ‘other domains’ category, because it teaches students about lists, which is an important data structure in programming, but not programming per se. Table 2 (right-hand side) also shows the frequency of the occurrence of aspects in the various ITS domains.
Table 2 shows that ITSs in the domain of mathematics more often diagnose common errors than ITSs in the other domains: 91% of the math tutors diagnose common errors, compared to only 33% of programming tutors, 50% of physics tutors, and 33% of the tutors in other domains. In mathematics, problems typically have a single correct solution, and there are only a few ways to reach that solution. Many errors in student steps can be explained by buggy rules, also because the solution space is relatively small. This partially explains why common errors are relatively often diagnosed in ITSs for mathematics.
In the domain of programming, ITSs diagnose the type of error more often than in other domains: 87% of the programming tutors diagnose the type of error, compared to 46% of the mathematics tutors, 75% of the physics tutors, and 67% of the tutors in other domains. This is perhaps due to the solution space in the domains. In programming tutors, the solution space is usually very large, which makes diagnosing common errors infeasible. Programs may have errors on different levels: syntax, dependency, typing, semantics, and more. This makes type of error a more informative diagnosis than in situations where only syntax and semantics play a role, as is usual in mathematics.
Redundancy is diagnosed in three programming tutors and two other domain tutors, but not in any mathematics or physics tutor. Because of the small sample size, we did not perform a statistical test to determine the significance of these results. The rest of the aspects seem to be diagnosed at a lower frequency across domains.
4.3 Diagnostic Processes
Most ITSs use multiple aspects for diagnosing student responses. How are these aspects combined in a diagnosis? We discuss how the different aspects are combined by the different ITSs to arrive at a diagnosis, and what the commonalities are between these systems. Not all ITSs are covered here, because some papers do not provide enough detail to extract the precise diagnosing process.
Figure 1 shows the most basic diagnostic process. Ovals represent input, grey nodes represent diagnostic ITS components, and rounded rectangles represent a diagnosis. This diagram represents the diagnostic processes in Assistment, Design-a-Plant, and Quantum Accounting. A student step is checked against a single good step. If it matches, the response is correct; if not, the response is incorrect. Although Assistment and Quantum Accounting have an additional diagnostic aspect, namely type of error, this is not shown in the diagram, because it is unclear where the type of error is determined. RMT’s diagnostic process is very similar, except that it uses cosine similarity to check whether a step matches an expected step.
The basic diagram in Fig. 1 can be extended in several ways. The diagnostic processes of the ACT Programming Tutor, LISP Tutor, Geometry Tutor, and PAT2Math add a second diagnostic component (i.e. a grey block) after correctness has determined that the student step does not match a good step. In this second component, common errors are searched for by using a set of buggy rules. Dragoon, on the other hand, extends the diagram with a diagnostic component that determines redundancy before checking correctness.
We give a single example of a more involved diagnostic process, and refer the reader to Van der Bent’s thesis [12] for many more diagrammatic representations of diagnostic processes that were found in ITSs.
The diagnostic process of AITS is illustrated in Fig. 2. AITS calculates the difference using edit distance. This information is used to infer correctness. If the edit distance is zero, the node sequence is correct. Otherwise, AITS checks the number of nodes and the content of the nodes in the submitted answer, and uses this to determine redundancy and type of error: AITS treats redundancy as one type of error. The complete and accurate diagnoses are labelled as types of errors. In AITS, a type of error is a combination of completeness and accuracy, so a step can be complete but inaccurate, incomplete but accurate, or incomplete and inaccurate. The diagnosis complete and accurate never occurs since then the edit distance would be zero, and the step would have been diagnosed as correct.
4.4 Patterns in Diagnostic Processes
Figure 3 illustrates the general diagnostic process. A dashed border indicates that the components are optional. All tutors check whether a step is correct using correctness or difference. Before this is done, however, some tutors check the order of steps or how much time was taken to submit a step. After it has been determined that a step is correct, some tutors check whether the correct step is also a preferred step. Some tutors also check whether a correct step is redundant. For incorrect steps, some tutors check whether the step contains common errors, and what type of error was made. Lastly, some tutors check whether an incorrect step is redundant. Note that, as was mentioned before, some tutors consider redundancy as an error, while others treat it as correct.
Some ITSs make more fine-grained diagnoses than the ones discussed in this study [8, 48]. For example, in Arends’ ITS [8] expressions can be semantically equivalent after an incorrect step. To signal such a step, the system can diagnose expressions that are semantically equivalent while also following a buggy rule, or expressions that are expected by a strategy despite not being semantically equivalent. Since these types of diagnoses only appear in this particular ITS, and seem to be very particular to the domain, we did not include them in our research.
5 Conclusion
As an answer to our research question, we found eight diagnostic aspects of student responses in Intelligent Tutoring Systems: correctness, difference, redundancy, type of error, common error, order, preference, and time. The diagnostic aspects are combined in various ways in the full diagnoses of the ITSs. Although these processes vary widely between systems, we distilled a general, abstract process that is used in all ITSs. All ITSs diagnose correctness, and although there are differences between domains, common errors and the type of error are also often diagnosed. The main difference between domains is that common errors is the second most frequently diagnosed aspect in mathematics tutors, whereas type of error is the second most frequently diagnosed aspect in programming tutors. Our analysis found no difference between four common tutoring approaches.
A limitation of our work is that the analysis of diagnostic processes is based on the information given in the papers written about the ITSs, rather than on the source code of the ITSs. Not all papers provide an in-depth description of how student steps are diagnosed, which made it impossible to describe the diagnostic processes of some systems. Sometimes we had to interpret the text to determine the diagnostic process.
Our analysis of diagnostic processes in ITSs contributes to a better understanding of the diagnosing behaviour of ITSs. For future research, the results of this study could be combined with results from evaluations of the effectiveness of tutoring systems [61]. This would give insight into which diagnostic processes are most effective at improving learning. This insight could then inform the design and development of tutoring systems in the future.
References
Aleven, V., McLaren, B.M., Sewall, J.: Scaling up programming by demonstration for intelligent tutoring systems development: an open-access web site for middle school mathematics learning. IEEE Trans. Learn. Technol. 2(2), 64–78 (2009)
Aleven, V., Mclaren, B.M., Sewall, J., Koedinger, K.R.: A new paradigm for intelligent tutoring systems: example-tracing tutors. Int. J. Artif. Intell. Educ. 19(2), 105–154 (2009)
Aleven, V., Popescu, O., Koedinger, K.: Pilot-testing a tutorial dialogue system that supports self-explanation. In: Cerri, S.A., Gouardères, G., Paraguaçu, F. (eds.) ITS 2002. LNCS, vol. 2363, pp. 344–354. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-47987-2_38
Aleven, V., Popescu, O., Koedinger, K.R.: Towards tutorial dialog to support self-explanation: adding natural language understanding to a cognitive tutor. In: Proceedings of Artificial Intelligence in Education, pp. 246–255. Citeseer (2001)
Anderson, J.R., Boyle, C.F., Reiser, B.J.: Intelligent tutoring systems. Science 228(4698), 456–462 (1985)
Anderson, J.R., Boyle, C.F., Yost, G.: The geometry tutor. In: IJCAI, pp. 1–7 (1985)
Anderson, J.R., Reiser, B.J.: The LISP tutor. Byte 10, 159–175 (1985)
Arends, H., Keuning, H., Heeren, B., Jeuring, J.: An intelligent tutor to learn the evaluation of microcontroller I/O programming expressions. In: Proceedings of the 17th Koli Calling Conference on Computing Education Research, pp. 2–9. ACM (2017)
Arevalillo-Herráez, M., Arnau, D., Marco-Giménez, L.: Domain-specific knowledge representation and inference engine for an intelligent tutoring system. Knowl.-Based Syst. 49, 97–105 (2013)
Arnau, D., Arevalillo-Herráez, M., Puig, L., González-Calero, J.A.: Fundamentals of the design and the operation of an intelligent tutoring system for the learning of the arithmetical and algebraic way of solving word problems. Comput. Educ. 63, 119–130 (2013)
Arnott, E., Hastings, P., Allbritton, D.: Research methods tutor: evaluation of a dialogue-based tutoring system in the classroom. Behav. Res. Methods 40(3), 694–698 (2008)
van der Bent, R.: The diagnosing behaviour of intelligent tutoring systems. Master’s thesis, Universiteit Utrecht (2018)
Blank, G., Parvez, S., Wei, F., Moritz, S.: A web-based ITS for OO design. In: Proceedings of Workshop on Adaptive Systems for Web-based Education at 12th International Conference on Artificial Intelligence in Education (AIED 2005), Amsterdam, the Netherlands, pp. 59–64 (2005)
Brown, J.S., Burton, R.R.: Diagnostic models for procedural bugs in basic mathematical skills. Cogn. Sci. 2(2), 155–192 (1978)
Burton, R.R., Brown, J.S.: A tutoring and student modelling paradigm for gaming environments. ACM SIGCUE Outlook 10(SI), 236–246 (1976)
Chanier, T., Pengelly, M., Twidale, M., Self, J.: Conceptual modelling in error analysis in computer-assisted language learning systems. In: Swartz, M.L., Yazdani, M. (eds.) Intelligent Tutoring Systems for Foreign Language Learning. NATO ASI Series, pp. 125–150. Springer, Heidelberg (1992). https://doi.org/10.1007/978-3-642-77202-3_9
Cheung, A.C., Slavin, R.E.: The effectiveness of educational technology applications for enhancing mathematics achievement in k-12 classrooms: a meta-analysis. Educ. Res. Rev. 9, 88–113 (2013)
Corbett, A.T., Anderson, J.R.: Locus of feedback control in computer-based tutoring: impact on learning rate, achievement and attitudes. In: Proceedings of the SIGCHI conference on Human factors in computing systems, pp. 245–252. ACM (2001)
Corbett, A.T., Anderson, J.R., Patterson, E.G.: Student modeling and tutoring flexibility in the Lisp intelligent tutoring system. In: Gauthier, G., Frasson, C. (eds.) Intelligent Tutoring Systems: at the crossroads of artificial intelligence and education, pp. 83–106. Intellect Ltd. (1990)
Demenko, G., Wagner, A., Cylwik, N.: The use of speech technology in foreign language pronunciation training. Arch. Acoust. 35(3), 309–329 (2010)
El-Kechaï, N., Delozanne, É., Prévit, D., Grugeon, B., Chenevotot, F.: Evaluating the performance of a diagnosis system in school algebra. In: Leung, H., Popescu, E., Cao, Y., Lau, R.W.H., Nejdl, W. (eds.) ICWL 2011. LNCS, vol. 7048, pp. 263–272. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25813-8_28
Fossati, D., Di Eugenio, B., Ohlsson, S., Brown, C., Chen, L.: Data driven automatic feedback generation in the ilist intelligent tutoring system. Technol. Instr. Cogn. Learn. 10(1), 5–26 (2015)
Glass, M.: Some phenomena handled by the circsim-tutor version 3 input understander. In: Proceedings of the Tenth Florida Artificial Intelligence Research Symposium, Daytona Beach, pp. 21–25 (1997)
Goguadze, G., Melis, E.: Combining evaluative and generative diagnosis in activemath. In: Proceedings of the 2009 Conference on Artificial Intelligence in Education: Building Learning Systems That Care: From Knowledge Representation to Affective Modelling, pp. 668–670 (2009)
Graesser, A.C., et al.: Autotutor: a tutor with dialogue in natural language. Behav. Res. Methods Instrum. Comput. 36(2), 180–192 (2004)
Graesser, A.C., Wiemer-Hastings, P., Wiemer-Hastings, K., Harter, D., Tutoring Research Group, T.R.G., Person, N.: Using latent semantic analysis to evaluate the contributions of students in autotutor. Interact. Learn. Environ. 8(2), 129–147 (2000)
Grivokostopoulou, F., Perikos, I., Hatzilygeroudis, I.: An educational system for learning search algorithms and automatically assessing student performance. Int. J. Artif. Intell. Educ. 27(1), 207–240 (2017)
Hattie, J., Timperley, H.: The power of feedback. Rev. Educ. Res. 77(1), 81–112 (2007)
Heeren, B., Jeuring, J.: Feedback services for stepwise exercises. Sci. Comput. Program. 88, 110–129 (2014)
Heffernan, N.T., Koedinger, K.R.: An intelligent tutoring system incorporating a model of an experienced human tutor. In: Cerri, S.A., Gouardères, G., Paraguaçu, F. (eds.) ITS 2002. LNCS, vol. 2363, pp. 596–608. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-47987-2_61
Hennecke, M.: Online Diagnose in intelligenten mathematischen Lehr-Lern-Systemen. Ph.D. thesis, Hildesheim University (1999). in German
Hong, J.: Guided programming and automated error analysis in an intelligent prolog tutor. Int. J. Hum Comput Stud. 61(4), 505–534 (2004)
Jaques, P.A., et al.: Rule-based expert systems to support step-by-step guidance in algebraic problem solving: the case of the tutor pat2math. Expert Syst. Appl. 40(14), 5456–5465 (2013)
Jeuring, J., Gerdes, A., Heeren, B.: A programming tutor for haskell. In: Zsók, V., Horváth, Z., Plasmeijer, R. (eds.) CEFP 2011. LNCS, vol. 7241, pp. 1–45. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-32096-5_1
Jin, W., Barnes, T., Stamper, J., Eagle, M.J., Johnson, M.W., Lehmann, L.: Program representation for automatic hint generation for a data-driven novice programming tutor. In: Cerri, S.A., Clancey, W.J., Papadourakis, G., Panourgia, K. (eds.) ITS 2012. LNCS, vol. 7315, pp. 304–309. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-30950-2_40
Jin, W., Corbett, A., Lloyd, W., Baumstark, L., Rolka, C.: Evaluation of guided-planning and assisted-coding with task relevant dynamic hinting. In: Trausan-Matu, S., Boyer, K.E., Crosby, M., Panourgia, K. (eds.) ITS 2014. LNCS, vol. 8474, pp. 318–328. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-07221-0_40
Johnson, B.G., Phillips, F., Chase, L.G.: An intelligent tutoring system for the accounting cycle: enhancing textbook homework with artificial intelligence. J. Account. Educ. 27(1), 30–39 (2009)
Johnson, S.D., et al.: Application of cognitive theory to the design, development, and implementation of a computer-based troubleshooting tutor (1992)
Johnson, W.L.: Intention-Based Diagnosis of Novice Programming Errors. MorganKaufmann, Los Altos (1986)
Keuning, H., Heeren, B., Jeuring, J.: Strategy-based feedback in a programming tutor. In: Proceedings of the Computer Science Education Research Conference, pp. 43–54. ACM (2014)
Keuning, H., Jeuring, J., Heeren, B.: A systematic literature review of automated feedback generation for programming exercises. ACM Trans. Comput. Educ. 19(1), 3:1–3:43 (2018)
Kim, N., Evens, M., Michael, J.A., Rovick, A.A.: Circsim-tutor: an intelligent tutoring system for circulatory physiology. In: Maurer, H. (ed.) ICCAL 1989. LNCS, vol. 360, pp. 254–266. Springer, Heidelberg (1989). https://doi.org/10.1007/3-540-51142-3_64
Koedinger, K.R., Aleven, V., Heffernan, N., McLaren, B., Hockenberry, M.: Opening the door to non-programmers: authoring intelligent tutor behavior by demonstration. In: Lester, J.C., Vicari, R.M., Paraguaçu, F. (eds.) ITS 2004. LNCS, vol. 3220, pp. 162–174. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-30139-4_16
Koedinger, K.R., Anderson, J.R.: Reifying implicit planning in geometry guidelines for model-based intelligent. In: Lajoie, S., Derry, S. (eds.) Computers as Cognitive Tools. Erlbaum, Hillsdale (2013)
Lee, W., de Silva, R., Peterson, E.J., Calfee, R.C., Stahovich, T.F.: Newton’s pen: a pen-based tutoring system for statics. Comput. Graph. 32(5), 511–524 (2008)
Lester, J.C., Stone, B.A., O’Leary, M.A., Stevenson, R.B.: Focusing problem solving in design-centered learning environments. In: Frasson, C., Gauthier, G., Lesgold, A. (eds.) ITS 1996. LNCS, vol. 1086, pp. 475–483. Springer, Heidelberg (1996). https://doi.org/10.1007/3-540-61327-7_146
Lester, J.C., Stone, B.A., Stelling, G.D.: Lifelike pedagogical agents for mixed-initiative problem solving in constructivist learning environments. User Model. User-Adap. Inter. 9(1–2), 1–44 (1999)
Lodder, J., Heeren, B., Jeuring, J.: A domain reasoner for propositional logic. J. Univ. Comput. Sci. 22(8), 1097–1122 (2016)
Looi, C.K.: Automatic debugging of prolog programs in a prolog intelligent tutoring system. Instr. Sci. 20(2–3), 215–263 (1991)
Mitrovic, A., Suraweera, P., Martin, B.: Intelligent tutors for all: the constraint-based approach. IEEE Intell. Syst. 22(4), 38–45 (2007)
Razzaq, L.M., et al.: Blending assessment and instructional assisting. In: AIED, pp. 555–562 (2005)
Rivers, K., Koedinger, K.R.: Data-driven hint generation in vast solution spaces: a self-improving python programming tutor. Int. J. Artif. Intell. Educ. 27(1), 37–64 (2017)
Roll, I., Aleven, V., Koedinger, K.R.: The invention lab: using a hybrid of model tracing and constraint-based modeling to offer intelligent support in inquiry environments. In: Aleven, V., Kay, J., Mostow, J. (eds.) ITS 2010. LNCS, vol. 6094, pp. 115–124. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-13388-6_16
Sangwin, C.: Computer Aided Assessment of Mathematics. Oxford University Press, Oxford (2013)
Sklavakis, D., Refanidis, I.: An individualized web-based algebra tutor based on dynamic deep model tracing. In: Darzentas, J., Vouros, G.A., Vosinakis, S., Arnellos, A. (eds.) SETN 2008. LNCS (LNAI), vol. 5138, pp. 389–394. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-87881-0_38
Sklavakis, D., Refanidis, I.: Mathesis: an intelligent web-based algebra tutoring school. Int. J. Artif. Intell. Educ. 22(4), 191–218 (2013)
Song, J., Hahn, S., Tak, K., Kim, J.: An intelligent tutoring system for introductory C language course. Comput. Educ. 28(2), 93–102 (1997)
Suraweera, P., Mitrovic, A.: KERMIT: a constraint-based tutor for database modeling. In: Cerri, S.A., Gouardères, G., Paraguaçu, F. (eds.) ITS 2002. LNCS, vol. 2363, pp. 377–387. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-47987-2_41
Sykes, E.R.: Design, development and evaluation of the Java intelligent tutoring system. Technol. Instr. Cogn. Learn. 8(1), 25–65 (2010)
VanLehn, K.: The behavior of tutoring systems. Int. J. Artif. Intell. Educ. 16(3), 227–265 (2006)
VanLehn, K.: The relative effectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems. Educ. Psychol. 46(4), 197–221 (2011)
VanLehn, K., et al.: The Architecture of Why2-Atlas: A Coach for Qualitative Physics Essay Writing. In: Cerri, S.A., Gouardères, G., Paraguaçu, F. (eds.) ITS 2002. LNCS, vol. 2363, pp. 158–167. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-47987-2_20
VanLehn, K.: The andes physics tutoring system: lessons learned. Int. J. Artifi. Intell. Educ. 15(3), 147–204 (2005)
Weber, G., Brusilovsky, P.: Elm-art: an adaptive versatile system for web-based instruction. Int. J. Artif. Intell. Educ. (IJAIED) 12, 351–384 (2001)
Weragama, D., Reye, J.: Analysing student programs in the php intelligent tutoring system. Int. J. Artif. Intell. Educ. 24(2), 162–188 (2014)
Wetzel, J., et al.: The design and development of the dragoon intelligent tutoring system for model construction: lessons learned. Interact. Learn. Environ. 25(3), 361–381 (2017)
Zatarain-Cabada, R., Barrón-Estrada, M.L., Pérez, Y.H., Reyes-García, C.A.: Designing and implementing affective and intelligent tutoring systems in a learning social network. In: Batyrshin, I., Mendoza, M.G. (eds.) MICAI 2012. LNCS (LNAI), vol. 7630, pp. 444–455. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-37798-3_39
Zinn, C.: Algorithmic debugging to support cognitive diagnosis in tutoring systems. In: Bach, J., Edelkamp, S. (eds.) KI 2011. LNCS (LNAI), vol. 7006, pp. 357–368. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-24455-1_35
Acknowledgements
The authors would like to thank the anonymous reviewers and the members of the Utrecht reading club on educational technology for their helpful suggestions.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
van der Bent, R., Jeuring, J., Heeren, B. (2019). The Diagnosing Behaviour of Intelligent Tutoring Systems. In: Scheffel, M., Broisin, J., Pammer-Schindler, V., Ioannou, A., Schneider, J. (eds) Transforming Learning with Meaningful Technologies. EC-TEL 2019. Lecture Notes in Computer Science(), vol 11722. Springer, Cham. https://doi.org/10.1007/978-3-030-29736-7_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-29736-7_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-29735-0
Online ISBN: 978-3-030-29736-7
eBook Packages: Computer ScienceComputer Science (R0)