Abstract
Nowadays, the usability and the user experience, are important aspects for the success of any software product. In the educational domain, Learning Support Platforms are not the exception, and these quality attributes become essential to guarantee an appropriate teaching process. For this reason, in this study, we present a systematic literature review whose purpose was to identify the characteristics that these types of systems must meet to obtain usable products. In the same way, the features that contribute to the assurance of a satisfying user experience have been cataloged. The purpose of this study is to establish the basis towards the elaboration of a future framework to quantify the level of usability and user experience of learning support platforms. The systematic search retrieved a total of 105 studies, from which 23 were selected as relevant. A set of 15 sub-attributes with their corresponding guidelines were identified to serve as a guide to design graphical user interfaces in a correct way in this type of software systems.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Human-Computer Interaction
- Systematic review
- Usability evaluation
- User experience
- Guidelines and principles
- Education
1 Introduction
Over the years, technology has been evolving on a large scale, which has led many organizations to take advantage of this event and make efforts to implement learning management systems (LMS), which improves the quality and scope of learning processes [1]. The Education sector is one of the most interesting sectors to analyze regarding the use of this type of LMS tools [2]. In order for the LMS to fulfill its objective efficiently, certain aspects must be considered, such as motivation, interest, commitment, focus on the task, behavior and much more [3].
Sometimes, the efficiency of the LMS is affected by the lack of usability. This is because the user could invest more effort in trying to understand the software than in carrying out their activities. Therefore, there are two factors that must be considered for the success of the LMS: the quality of information and the way in which it is displayed to the user. According to Westfall, it was reported that 61% of people surveyed affirm that one of the inconveniences they had with LMS was learning how it works [4].
Therefore, usability is one of the key factors for the user to achieve satisfaction while using the LMS. This involves making the software easy to use and learn, with a user-friendly interface; displaying the content of a simple and orderly manner, thus preventing the user to focus on these issues [5]. Measuring usability is considered one of the great challenges, this is because there is no standard model which covers all the necessary usability attributes [6].
Heuristic evaluation is one of the most commonly used qualitative techniques to inspect software interfaces and find problems that affect usability [7]. On the other hand, quantitatively assessing the level of usability provides certain benefits, such as making objective comparisons between the software of the same type, which makes it easier for companies to make decisions to choose the best product [8]. With a numerical score, it favors the identification of the state of the level of usability of the software [9].
In this research, it is proposed to find those most important features about usability and electronic learning, as well as quantitative usability evaluations, taking into account heuristics and guidelines. To achieve this, a systematic literature review was carried out, following the Kitchenham methodology. Thus, the most important research on the subject was identified, which will serve as support to build a preliminary framework that allows quantifying the level of usability in those support tools for learning, in the near future.
2 A Systematic Literature Review
To identify those important key aspects that LMS must meet, a systematic review was conducted. The purpose of identifying those aspects was to analyze them, systematize them, and to propose in a future research, a framework to assess the usability and user experience based on the most important identified aspects.
The present systematic review was performed based on the parameters defined by Kitchenham and Charters [10]. In this case, the activities that were performed are: (1) definition of the research questions and search strategy, (2) selection of the primary studies, the extraction of the papers and (3) analysis of the results.
2.1 Research Questions
The main objective was to summarize some studies related to usability, user experience, learning management systems and usability evaluation methods. We used the PICOC table criteria in order to do this review according to the protocol established by Petticrew and Roberts [11]. In addition, we employed synonyms and related terms to find better results. These criteria are shown in Table 1.
Based on the concepts established using PICOC, we defined the following research questions oriented to identify relevant aspects in the design of graphical interfaces for learning management systems:
-
RQ1: Which are the most relevant aspects of usability and user experience?
-
RQ2: Which are the most important characteristics in LMS?
-
RQ3: Which are the activities/characteristics of the usability evaluation models?
-
RQ4: Which are the guidelines considered in other domains to evaluate usability?
2.2 Search Strategy
We defined our search strategy based on the general concepts. Some synonymous were selected to achieve a more comprehensive search. The search process was performed by using four recognized databases to search for primary studies: ACM Digital Library, IEEExplore, SCOPUS and SpringerLink. Grey literature was excluded since it is not peer reviewed.
In this phase, the search chains or queries that were used in the search engines of each database were formulated. It is important to mention that the syntax of the queries could vary according to the database that is employed. After grouping a series of concepts using the connectors AND/OR, the resulting search string was the following:
(“usability” OR “user experience” OR “UX” OR “HCI”) AND (“heuristic” OR “heuristic evaluation” OR “usability evaluation” OR “model evaluation usability”) AND (“quantify” OR “quantitative” OR “approach quantitative” OR “comparative” OR “comparative analysis” OR “methodology to evaluate usability” OR “quantify usability” OR “measure usability” OR “method to evaluate usability”) AND (“interface” OR “software” OR “web” OR “system” OR “satisfaction” OR “characteristic” OR “guideline” OR “design” OR “methodology” OR “case study”) AND (“elearning” OR “e-learning” OR “learning management system” OR “LMS” OR “education” OR “higher education” OR “university” OR “e-learning system” OR “web platform to teaching” OR “web platform” OR “web system to support teaching”).
2.3 Search Process and Data Extraction
In order to determine if an article must be considered as relevant, we defined the following inclusion criteria: the study should present a methodology, framework or study case in which the usability and user experience are evaluated. In the same way, we defined the exclusion criteria: in the study, the specialists do not apply a usability or user experience evaluation in an intangible product as a software. The automated search for our systematic mapping review was performed on October 20th, 2018. Table 2 shows the search results that were found. In addition, Table 3 shows the selected studies from the four databases used in this research. These studies were selected by discarding the studies that do not meet the inclusion criteria and present any of the exclusion criteria.
In addition, some secondary studies were identified applying the inclusion and exclusion criteria to the references established in the most relevant of the primary studies. These secondary papers are shown in Table 4.
Once these studies were obtained through the secondary review process, these were cataloged based on the relevant works. This information is shown in Table 5.
3 Results of the Review
3.1 Relevant Aspects of Usability and User Experience
There were 18 papers that answer RQ1 (SS1, SS4, SS6, SS14, SS24, SS26, SS27, SS29, SS30, SS35, SS36, SS37, SS38, SS44, SS47, SS51, SS53, SS54). The usability/UX aspects that found in the systematic review are shown in Table 6.
3.2 Important Characteristics in Learning Management Systems
In the same way, to answer RQ2, 14 papers were selected (SS5, SS9, SS13, SS16, SS19, SS21, SS28, SS31, SS32, SS39, SS42, SS46, SS49, SS50). According to the authors, there are relevant design aspects that contribute to satisfy the learning objectives that are frequently requested by the educational institutions when they search for learning tools. We detail some of those design aspects in Table 7.
3.3 Activities and Characteristics of the Usability Evaluation Models
There were 11 papers that answer RQ3 (SS3, SS12, SS15, SS17, SS20, SS22, SS23, SS28, SS32, SS45, SS52). Some important strategies related to the process that must be followed in a usability/UX evaluation were obtained from these studies. These approaches were compared to decide how to establish the assessment framework. We show some of those strategies in Table 8.
3.4 Guidelines to Evaluate the Usability Attribute
Finally, the RQ4 is related with the principles used in the heuristic evaluation models which allow to get the usability level - quantitatively or qualitatively - in the platforms from other domains, that can be used as well for learning support platforms. There were 18 papers that were used to answer this research question (SS2, SS3, SS7, SS8, SS10, SS11, SS18, SS23, SS25, SS28, SS32, SS33, SS34, SS40, SS41, SS43, SS48, SS55). Some the guidelines found are shown in Table 9.
4 Conclusions and Future Works
The usability and UX have become critical aspects to be considered in the development of software products. Nowadays, these quality attributes represent the main concerns of the software industries, since ensuring a high level of ease of use and UX in the applications, leads to establish an environment of appropriate use for the interaction with the system. The new paradigm in the software development is not only about providing the users with a tool to achieve their goals but also to ensure that the user experience is quality enough to generate satisfaction on the end user.
In this study, we performed a systematic literature review following a recognized and widely used methodology. According to this protocol, we identified 105 studies, from which 23 were selected. Furthermore, we considered 32 secondary studies, that provided more information to 4 research questions. This work allowed to identify the most relevant aspects in both usability as learning. Additionally, this review allowed to find characteristics that help to quantify the level usability and UX and some heuristics/guidelines related to domain of study.
The objective of this systematic literature review was to obtain relevant information from previous research to build a preliminary framework that allows quantifying the level of usability and UX in learning support platforms, through a checklist. Therefore, a deeper analysis should be carried out to generate the results to the objectives for the construction of the framework. Also, some of the results must be complemented and validated by interviews and expert judgment, respectively.
References
Penha, M., Correia, W., Soares, M., Campos, F., Barros, M.: Ergonomic evaluation of usability with users – application of the technique of cooperative evaluation. In: Marcus, A. (ed.) DUXU 2013. LNCS, vol. 8012, pp. 379–388. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-39229-0_41
Ilyas, M., Kadir, K.A., Adnan, Z.: Demystifying the learning management system (LMS): journey from e-learning to the strategic role. Eur. J. Bus. Manag. 9, 12–18 (2017)
Cavus, N., Alhih, M.S.: Learning management systems use in science education. Procedia - Soc. Behav. Sci. 143, 517–520 (2014). https://doi.org/10.1016/j.sbspro.2014.07.429
Westfall, B.: Learning Management System User Report - 2016. SoftwareAdvice.com (2016). https://www.softwareadvice.com/resources/lms-user-trends-2016/
Kumar, R., Hasteer, N.: Evaluating usability of a web application: a comparative analysis of open-source tools. In: Proceedings of the 2nd International Conference on Communication and Electronics Systems (ICCES 2017), Coimbatore, India, pp. 350–354 (2017). https://doi.org/10.1109/CESYS.2017.8321296
Hasan, L.A., Al-Sarayreh, K.T.: An integrated measurement model for evaluating usability attributes. In: Proceedings of the International Conference on Intelligent Information Processing, Security and Advanced Communication, pp. 1–6. ACM, Batna (2015). https://doi.org/10.1145/2816839.2816861
Granollers, T.: Usability evaluation with heuristics. new proposal from integrating two trusted sources. In: Marcus, A., Wang, W. (eds.) DUXU 2018. LNCS, vol. 10918, pp. 396–405. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-91797-9_28
Paz, F., Paz, F.A., Sánchez, M., Moquillaza, A., Collantes, L.: Quantifying the usability through a variant of the traditional heuristic evaluation process. In: Marcus, A., Wang, W. (eds.) DUXU 2018. LNCS, vol. 10918, pp. 496–508. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-91797-9_36
Granollers, T.: Validación experimental de un conjunto heurístico para evaluaciones de UX de sitios web de comercio-e. In: Proceedings of the 2016 IEEE 11th Colombian Computing Conference (CCC 2016), Popayan, Colombia, pp. 1–8 (2016). https://doi.org/10.1109/ColumbianCC.2016.7750783
Kitchenham, B., Pearl Brereton, O., Budgen, D., Turner, M., Bailey, J., Linkman, S.: Systematic literature reviews in software engineering – A systematic literature review. Inf. Softw. Technol. 51, 7–15 (2009). https://doi.org/10.1016/j.infsof.2008.09.009
Petticrew, M., Roberts, H.: Systematic Reviews in the Social Sciences: A Practical Guide. Blackwell Publishing, Hoboken (2005). ISBN 978-1-405-12110-1
Hedegaard, S., Simonsen, J.G.: Extracting usability and user experience information from online user reviews. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 2089–2098. ACM, Paris (2013). https://doi.org/10.1145/2470654.2481286
Al-Faries, A., Al-Khalifa, H.S., Al-Razgan, M.S., Al-Duwais, M.: Evaluating the accessibility and usability of top Saudi e-government services. In: Proceedings of the 7th International Conference on Theory and Practice of Electronic Governance, pp. 60–63. ACM, Seoul (2013). https://doi.org/10.1145/2591888.2591898
Dias, A.L., Fortes, R.P.d.M., Masiero, P.C.: HEUA: a heuristic evaluation with usability and accessibility requirements to assess web systems. In: Proceedings of the 11th Web for All Conference, pp. 1–4. ACM, Seoul (2014). https://doi.org/10.1145/2596695.2596706
Hovde, M.R.: Effective user experience in online technical communication courses: employing multiple methods within organizational contexts to assess usability. In: Proceedings of the 33rd Annual International Conference on the Design of Communication, pp. 1–5. ACM, Limerick (2015). https://doi.org/10.1145/2775441.2775453
Choma, J., Zaina, L.A.M., Silva, T.S.d.: Working beyond technical aspects: an approach for driving the usability inspection adding the perspective of user experience. In: Proceedings of the 34th ACM International Conference on the Design of Communication, pp. 1–10. ACM, Silver Spring (2016). https://doi.org/10.1145/2987592.2987607
Quiñones, D., Rusu, C., Rusu, V.: A methodology to develop usability/user experience heuristics. Comput. Stand. Interfaces 59, 109–129 (2018). https://doi.org/10.1016/j.csi.2018.03.002
Sagar, K., Saha, A.: Qualitative usability feature selection with ranking: a novel approach for ranking the identified usability problematic attributes for academic websites using data-mining techniques. Hum.-Centric Comput. Inf. Sci. 7, 1–24 (2017). https://doi.org/10.1186/s13673-017-0111-8
Hasan, L.: Usability problems on desktop and mobile interfaces of the Moodle learning management system (LMS). In: Proceedings of the 2018 International Conference on E-Business and Applications, pp. 69–73. ACM, Da Nang (2018). https://doi.org/10.1145/3194188.3194192
AlRoobaea, R.S., Al-Badi, A.H., Mayhew, P.J.: A framework for generating a domain specific inspection evaluation method: a comparative study on social networking websites. In: Proceedings of the 2013 Science and Information Conference (SAI 2013), London, UK, pp. 757–767 (2013). ISBN 978-0-9893193-0-0
Gordillo, A., Barra, E., Aguirre, S., Quemada, J.: The usefulness of usability and user experience evaluation methods on an e-Learning platform development from a developer’s perspective: a case study. In: Proceedings of the 2014 IEEE Frontiers in Education Conference (FIE 2014), Madrid, Spain, pp. 1–8 (2014). https://doi.org/10.1109/FIE.2014.7044340
Iman, M.R.H., Rasoolzadegan, A.: Quantitative evaluation of software usability with a fuzzy expert system. In: Proceedings of the 2015 5th International Conference on Computer and Knowledge Engineering (ICCKE 2015), Mashhad, Iran, pp. 325–330 (2015). https://doi.org/10.1109/ICCKE.2015.7365850
Mtebe, J.S., Kissaka, M.M.: Heuristics for evaluating usability of Learning Management Systems in Africa. In: Proceedings of the 2015 IST-Africa Conference, Lilongwe, Malawi, pp. 1–13 (2015). https://doi.org/10.1109/ISTAFRICA.2015.7190521
Kabir, M.A., Rehman, M.U., Majumdar, S.I.: An analytical and comparative study of software usability quality factors. In: Proceedings of the 2016 7th IEEE International Conference on Software Engineering and Service Science (ICSESS 2016), Beijing, China, pp. 800–803 (2016). https://doi.org/10.1109/ICSESS.2016.7883188
Ivanović, M., Putnik, Z., Komlenov, Ž.: Usability and privacy aspects of Moodle: students’ and teachers’ perspective. Informatica 37, 221–230 (2013)
Sabri, N.M., Mohamed, H., Soon, G.Y., Yusof, Y.m.H.: A quantitative approach in the usability evaluation of a courseware. J. Next Gener. Inf. Technol. 4, 29–38 (2013). https://doi.org/10.4156/jnit.vol4.issue2.4
Deraniyagala, R.L., Amdur, R.J., Boyer, A.L., Kaylor, S.: Usability study of the EduMod eLearning Program for contouring nodal stations of the head and neck. Pract. Radiat. Oncol. 5, 169–175 (2015). https://doi.org/10.1016/j.prro.2014.10.008
Junus, I.S., Santoso, H.B., Isal, Y.K., Utomo, A.Y.: Usability evaluation of the student centered e-Learning environment. Int. Rev. Res. Open Distance Learn. 16, 62–82 (2015). https://doi.org/10.19173/irrodl.v16i4.2175
Murillo, B., Vargas, S., Moquillaza, A., Fernández, L., Paz, F.: Usability testing as a complement of heuristic evaluation: a case study. In: Marcus, A., Wang, W. (eds.) DUXU 2017. LNCS, vol. 10288, pp. 434–444. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-58634-2_32
Emang, D.W.A.B., Lukman, R.N.I.R., Kamarulzaman, M.I.S., Zaaba, Z.F.: Usability studies on e-learning platforms: preliminary study in USM. In: Proceedings of the 2nd International Conference on Applied Science and Technology (ICAST 2017), Kedah, Malaysia, vol. 1891, pp. 1–8 (2017). https://doi.org/10.1063/1.5005373
Aparna, S.S., Baseer, K.K.: SIRIUS-WUEP: a heuristic-based framework for measuring and evaluating web usability in model-driven web development. In: Satapathy, S., Govardhan, A., Raju, K., Mandal, J. (eds.) Emerging ICT for Bridging the Future - Proceedings of the 49th Annual Convention of the Computer Society of India (CSI) Volume 1. Advances in Intelligent Systems and Computing, vol. 337, pp. 303–310. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-13728-5_34
Eason, K.D.: Towards the experimental study of usability. Behav. Inf. Technol. 3, 133–143 (1984). https://doi.org/10.1080/01449298408901744
Nielsen, J.: Usability Engineering. Morgan Kaufmann, Burlington (1993). ISBN 0125184050
ISO: Ergonomic requirements for office work with visual display terminals (VDTs) – Part 11: Guidance on usability - ISO 9241-11:1998. International Organization for Standardization (1998)
ISO: Software engineering – Product quality – Part 1: Quality model - ISO/IEC 9126:2001. International Organization for Standardization (2001)
Reeves, T.C., et al.: Usability and instructional design heuristics for E-Learning evaluation. In: Proceedings of the 2002 World Conference on Educational Multimedia, Hypermedia & Telecommunications (ED-MEDIA 2002), Denver, Colorado, USA, pp. 2–8 (2002). https://eric.ed.gov/?id=ED477084
Folmer, E., van Gurp, J., Bosch, J.: A framework for capturing the relationship between usability and software architecture. Softw. Process: Improv. Pract. 8, 67–87 (2003). https://doi.org/10.1002/spip.171
Abran, A., Khelifi, A., Suryn, W., Seffah, A.: Usability meanings and interpretations in ISO standards. Softw. Qual. J. 11, 325–338 (2003). https://doi.org/10.1023/A:1025869312943
Rosato, J., Dodds, C., Laughlin, S.: Usability of course management systems by students. Department of Computer Information Systems/Computer Science, College of Scholastica, Duluth (2006)
Mehlenbacher, B., et al.: Usable e-learning: a conceptual model for evaluation and design. In: 11th International Conference on Human-Computer Interaction (HCI International 2005), Las Vegas, NV, USA, pp. 1–10 (2005). http://citeseerx.ist.psu.edu/viewdoc/citations?doi=10.1.1.84.331
Dringus, L.P., Cohen, M.S.: An adaptable usability heuristic checklist for online courses. In: Proceedings Frontiers in Education 35th Annual Conference, Indianopolis, IN, USA (2005). https://doi.org/10.1109/FIE.2005.1611918
Ardito, C., et al.: An approach to usability evaluation of e-learning applications. Univ. Access Inf. Soc. 4, 270–283 (2006). https://doi.org/10.1007/s10209-005-0008-6
Seffah, A., Donyaee, M., Kline, R.B., Padda, H.K.: Usability measurement and metrics: a consolidated model. Softw. Qual. J. 14, 159–178 (2006). https://doi.org/10.1007/s11219-006-7600-8
Nielsen, J., Loranger, H.: Prioritizing Web Usability. New Riders Publishing, Thousand Oaks (2006). ISBN 0321350316
Bevan, N.: Classifying and selecting UX and usability measures. In: Proceedings of the Valid Useful User Experience Measurement (VUUM 2008), pp. 1–6. University of Iceland, Reykjavik (2008)
Ketola, P., Roto, V.: Exploring user experience measurement needs. In: Proceedings of the Valid Useful User Experience Measurement (VUUM 2008), pp. 23–26. University of Iceland, Reykjavik (2008). https://doi.org/10.4018/jthi.2009070104
Zaharias, P., Poylymenakou, A.: Developing a usability evaluation method for e-learning applications: beyond functional usability. Int. J. Hum.-Comput. Interact. 25, 75–98 (2009). https://doi.org/10.1080/10447310802546716
Giannakos, M.N.: A combinational evaluation method of computer applications. J. Basic Appl. Sci. 1, 240–242 (2009)
Al-Khalifa, H.S.: Heuristic evaluation of the usability of e-government websites: a case from Saudi Arabia. In: Proceedings of the 4th International Conference on Theory and Practice of Electronic Governance, pp. 238–242. ACM, Beijing (2010). http://doi.acm.org/10.1145/1930321.1930370
Al-sarrayrih, H.S., Knipping, L., Zorn, E.: Evaluation of a MOODLE based Learning Management System applied at Berlin Institute of Technology based on ISO-9126. In: 13th International Conference on Interactive Computer aided Learning (ICL2010), Hasselt, Belgium, pp. 1–8 (2010)
Giannakos, M.N.: The evaluation of an E-Learning Web-based platform. In: Proceedings of the 2nd International Conference on Computer Supported Education (CSEDU 2010), Valencia, Spain, pp. 433–438 (2010). https://doi.org/10.5220/0002799504330438
ISO: Systems and software engineering – Systems and software Quality Requirements and Evaluation (SQuaRE) – System and software quality models - ISO/IEC 25010:2011. International Organization for Standardization (2011)
Zaharias, P.: Heuristic evaluation of e-learning courses: a comparative analysis of two e-learning heuristic sets. Campus-Wide Inf. Syst. 29, 45–60 (2012). ISSN 1065-0741
Ghirardini, B.: E-Learning Methodologies: A Guide for Designing and Developing E-Learning Courses. Food and Agriculture Organization of the United Nations - FAO, USA (2012)
Kumar Dubey, S., Rana, A., Sharma, A.: Usability evaluation of object oriented software system using fuzzy logic approach. Int. J. Comput. Appl. 43, 1–6 (2012). https://doi.org/10.5120/6208-8778
Torrente, M.C.S., Prieto, A.B.M., Gutiérrez, D.A., de Sagastegui, M.E.A.: Sirius: a heuristic-based framework for measuring web usability adapted to the type of website. J. Syst. Softw. 86, 649–663 (2013). https://doi.org/10.1016/j.jss.2012.10.049
Şenol, L., Gecili, H., Durdu, P.O.: Usability evaluation of a Moodle based learning management system. In: Proceedings of the World Conference on Educational Media and Technology (EdMedia 2014), Tampere, Finland, pp. 850–858 (2014). ISBN 978-1-939797-08-7
Thuseethan, S., Achchuthan, S., Kuhanesan, S.: Usability evaluation of learning management systems in Sri Lankan Universities. Adv. Hum. Comput. Interact. 15, 1–13 (2015)
Gupta, D., Ahlawat, A., Sagar, K.: A critical analysis of a hierarchy based Usability Model. In: Proceedings of the 2014 International Conference on Contemporary Computing and Informatics (IC3I 2014), Mysore, India, pp. 255–260 (2014). https://doi.org/10.1109/IC3I.2014.7019810
Paz, F., Paz, F.A., Villanueva, D., Pow-Sang, J.A.: Heuristic evaluation as a complement to usability testing: a case study in web domain. In: Proceedings of the 12th International Conference on Information Technology - New Generations (ITNG 2015), Las Vegas, NV, USA, pp. 546–551 (2015). https://doi.org/10.1109/ITNG.2015.92
Issa, T., Isaias, P.: Sustainable Design: HCI, Usability and Environmental Concerns. Springer, Heidelberg (2015). https://doi.org/10.1007/978-1-4471-6753-2
Rogers, Y., Preece, J., Sharp, H.: Interaction Design: Beyond Human-Computer Interaction. Wiley, Hoboken (2015). ISBN 978-1-119-02075-2
Farmanesh, P., Samani, A.A.: Heuristic evaluation of the usability of Learning Management System (Moodle) at Eastern Mediterranean University, pp. 1–25 (2015). https://doi.org/10.5281/zenodo.59517
Acknowledgement
This study is highly supported by the Section of Informatics Engineering of the Pontifical Catholic University of Peru (PUCP) – Peru, and the “HCI, Design, User Experience, Accessibility & Innovation Technologies” Research Group (HCI-DUXAIT). HCI-DUXAIT is a research group of PUCP.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Salas, J. et al. (2019). Guidelines to Evaluate the Usability and User Experience of Learning Support Platforms: A Systematic Review. In: Ruiz, P., Agredo-Delgado, V. (eds) Human-Computer Interaction. HCI-COLLAB 2019. Communications in Computer and Information Science, vol 1114. Springer, Cham. https://doi.org/10.1007/978-3-030-37386-3_18
Download citation
DOI: https://doi.org/10.1007/978-3-030-37386-3_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-37385-6
Online ISBN: 978-3-030-37386-3
eBook Packages: Computer ScienceComputer Science (R0)