Keywords

1 Introduction

Human activity should not compromise the long-term balance between the economic, environmental and social pillars [1]. The evaluation of impact is a necessary part of all research and innovation (R&I) and attempts to improve the relationship between science and society. This follows a ‘perspective oriented to humane and social values’ [2] and recommendations from the European Union (EU) in relation to undertaking ‘responsible research and innovation (RRI)’ [3] and the involvement of societal actors in R&I [4]. New technologies have the potential to deliver benefits. However, such technologies are inherently uncertain. As stated by Capurro (2009), technology designers must examine the ethical implications of things which may not yet exist, or things which may have impacts we cannot predict [5]. In so doing, they must deal with uncertainty. This includes the ‘uncertainty of future products, uses and consequences, and associated ethical issues that will result from an emerging technology’ [6].

In asking what technology is and how it might be designed, we ask questions about who we are (identity) and what it means to be human [7]. As stated by Heidegger [1977], we examine the nature of existence and human autonomy [7]. Such ideas have led to the concept of ‘ontological design’ which addresses how the design of technology changes our human and social reality [8]. As such, we are designed by our designing and by that which we have designed [9].

Design/technology teams exercise choice in relation to what is valued and advancing technology that improves the human condition (and not worsens it). As researchers we need methods to assess and practice ethics, to ensure that new technologies positively contribute to human wellbeing and have positive impacts across the triple bottom line. In an ideal world, R&I teams are multi-disciplinary and include ethicists. Further, stakeholder evaluation underpins the generation of an evidence map and proposed solutions. However, this is not always the case. Methodologies are required to enable the active translation of ethical issues pertaining to the human and social dimensions of new technologies, in a manner that considers the diversity of practices across R&I and commercial research projects. To this end, this paper presents a new methodology for embedding ethics assessment in human machine interaction (HMI)/human factors (HF) design and evaluation activities.

2 Background

2.1 Underlying Concepts

Human Factors refers to ‘the practice of designing products, systems, or processes to take proper account of the interaction between them and the people who use them’ [10]. Ethics concerns the moral principles that govern a person’s behavior or how an activity is conducted. As researchers, we must distinguish research ethics (i.e. the normative aspects of engaging in scientific research) and the ethics of technological innovation and its impacts at different levels. ‘Digital ethics’ or information ethics deals with the impact of digital information and communication technologies (ICT) on society and the environment. Data ethics is defined as a branch of ethics that evaluates data practices with the potential to adversely impact on people and society [11].

2.2 The Practice of Ethics in R&I

A recent systematic review indicates that the practice of ethics in R&I is a relatively new topic [12]. While academic discussion on specific practices commenced in the 1990s, this research has gained considerable momentum in the last ten years [12]. According to Reijers et al. (2017), health technologies is the most represented in the literature, followed by the fields of information systems research and computer science [12].

Specific ethics approaches in R&I can take many forms. Reijers et al. (2017) categorize the different methods in relation to their application in the technology development lifecycle – distinguishing (1) ex ante methods, dealing with emerging technologies (2) intra methods, dealing with technology design and (3) ex post methods, dealing with ethical analysis of existing technologies [12]. Research evidence can include information from horizon scanning and participatory foresight activities, literature reviews, and field research with stakeholders [13]. Specific stakeholder evaluation research (i.e. empirical research) may take different formulations. Stakeholders may engage directly or indirectly with R&I teams. Researchers and stakeholders may engage with ethical challenges in a collaborative workshop. Or, research may be undertaken with stakeholders and later examined by research and design teams in a structured format. This format may follow specific conceptual frameworks and assessment approaches. Several key frameworks for ethical assessment have emerged. This includes (but is not limited to): ontological design [9], anticipatory technology ethics/foresight approach [13], value sensitive design [14], ethical impact assessment [15], the ETICA approach [16], and the techno-ethical scenarios approach [17]. Brey (2017) classifies five sets of ethical impact assessment approaches. This includes generic approaches, anticipatory/foresight approaches, risk assessment approaches, experimental approaches and participatory/deliberative ethics approaches [13]. Increasing, researchers are combining approaches. For example, Cotton (2014) combines participatory/deliberative ethics approaches and stakeholder approaches [18].

2.3 Ethics Canvases

Ethics canvases or visual tools which support the ethics assessment approach are not being used in commercial and research projects. In principle, these canvases allow non-ethicists such as Designers, Human Factors Researchers, Engineers, and Computer Scientists to engage in ethical issues pertaining to the emerging technology product. Examples of such canvases include the ‘Research Impacts Canvas’ (RIC) [19], The Ethical Matrix [20], The Digital Product Ethics Canvas and Impacts Canvas [21], The Humans & Machines Ethics Canvas’s [22], The Online Ethics Canvas [23], and the Data Ethics Canvas [24]. Some canvases focus on ethics and impact in a general sense, while others address specific themes. For example, the Online Ethics Canvas addresses the impact of new technology on human behavior and activity at individual and societal levels [23]. The Data Ethics Canvas considers ethical issues related to data privacy, data use and data quality [24].

3 Human Factors and Ethics Canvas

3.1 Rationale

Critically, human factors and ethical issues must be explored in an integrated way. Although valuable, the existing ethics canvases require further emphasis on framing the problem, specifying the psychosocial dimensions and impacts of new technologies and addressing specific stakeholder/end user requirements and impacts. Further, ethical issues need to be managed in terms of design decisions. These decisions need to be agreed and documented.

The ‘Human Factors & Ethics Canvas’ introduced by Cahill (2019) [25] reflects an integration of ethics and HF methods, particularly around the collection of evidence using stakeholder evaluation methods [26, 27] personae-based design [28], scenario-based design approaches [29]. Further, it makes use of ethical theories/perspectives that are used in relation to the analysis of technology innovation in relation to the analysis of benefit versus harm including Consequentialism, Deontology & Principlism [30].

3.2 Procedure

The HFEC can be used at any stage of the design process. As such, it spans the classification of methods proposed by Reijers et al. [12]. Overall, it combines anticipatory/foresight approaches and participatory/deliberative ethics approaches. In line with stakeholder evaluation approaches, the canvas can be evaluated using the ‘community of practice’ [27]. That is, using internal stakeholders (project team) and external stakeholders (relevant ends users/stakeholders and legitimate other parties who may be impacted by the technology). At a minimum, core internal stakeholders/core team members (including an ethicist {if available}, the HF lead, the design lead and the product owner/manager) are involved in completing the canvas. If the project team includes an ethicist, then they should take the role of the ‘HFEC’ coordinator - recording relevant information in the HFEC. Otherwise, this can be done by the HF lead or another designated member of the project team.

As indicated in Fig. 1, the HFEC is divided into seven stages or sections. For more, please see Appendix A. Stage 0 records project information. Stage 1 is all about framing the problem. Stage 2 involves understanding how the technology fits to the problem, defining stakeholder goals and needs and the specification of expected benefits for different stakeholders. This is followed by several more detailed examinations of core themes. These are: benefits, outcomes and impact (stage 3), personae and scenario (stage 4), data ethics (stage 5) and implementation (stage 6). The final stage (stage 7) presents the outcomes of the preceding analysis. An analysis of literature review data and information from team problem solving sessions can be used to populate the HFEC. However, it is best to complete Stage 3 and 4 either using stakeholder evaluation approaches (either direct engagement of stakeholders in ethical assessment or following the analysis of field research with stakeholders). In addition, Stage 6 can only be completed following implementation and evaluation of the proposed technologies. Ideally, this might occur in a field setting. However, information from simulation studies can also be used.

Fig. 1.
figure 1

Stages in Human Factors & Ethics Canvas

4 Discussion

As illustrated in the ethics canvas, there is much convergence between the analysis of new technology both from an ethics and human factors perspective (for example, addressing stakeholder need, expected benefits and outcomes, and impact [intended and unintended] – both at an individual and societal level). Ethical principles need to be both articulated and then embedded in the design concept. Personae/scenarios are useful in relation to considering and documenting the needs/perspectives of different stakeholders and adjudicating between conflicting goals/principles. Moreover, the translation of system objectives in relation to wellbeing and human benefit objectives (and associated metrics) ensures that wellbeing and human benefit are both a reference point and a design outcome.

As highlighted by Brey (2017), the ethics of emerging technologies ‘harbors the promise of early intervention when a technology is still malleable and there is still much room for choice in its development and social embedding’ [13]. However, researchers have a limited range of empirical data to use. As the technologies are not in use, there are ‘significant uncertainties regarding future developments and impacts’ (Brey 2017). Some theorists present philosophical objections to speculation about future impacts [31, 32]. For example, Nordmann (2007) contends that speculation about the future should be rejected as researchers cannot gain sufficient knowledge about the future to stipulate procedures for action or guidance in R&I processes [32]. Others argue that the available theories and methods do not provide adequate theoretical grounding in terms of how values might be embedded in design solutions [33]. In addition, VSD and related approaches must address the difference between designer’s intentions and user practice [34].

5 Conclusion

Assessing the ethical implications of things which may not yet exist, or things which may have impacts we cannot predict, is very difficult. However, this should not be barrier to posing important questions and ensuring that these questions are addressed as part of the design process. Thinking about both potential positive, negative consequences and unintended consequences enables designers to build in protections into the design concept. Overall, it is argued that the specification of an ethics canvas as part of a broader human factors design approach ensures that ethical issues are considered.