1 Introduction

Artificial intelligence (AI) has emerged as a transformative force in the e-commerce world, with the power to reshape how we shop and interact with technology [1, 2]. As technology continues to push the boundaries of e-tail, the online retail landscape, businesses find themselves in a race to keep pace with these rapid changes, adapting to new trends and innovations as they come [3]. E-tailing, or e-retailing, refers to the electronic commerce model enabling users to purchase goods or services directly from a seller via the Internet, facilitated through a web browser [4]. E-commerce, on the other hand, is a general term for any type of commercial activity conducted electronically [5]. According to recent reports, conversational AI can enhance e-tail users’ services by responding swiftly, promptly, and accurately to their enquiries; lowering customer service wait times; and boosting the number of users that customer support can manage [6,7,8]. Another report suggests that the global AI-powered e-commerce solution market will reach $16.8 billion by 2030, wherein AI will handle 80% of all customer interactions [9].

The advent of e-tail websites has revolutionized the way users shop online, and conversational AI agents have become an integral part of the online shopping experience [10, 11]. In recent years, conversational AI agents have emerged as a popular tool for providing personalized recommendations and information to users on e-tail platforms [12, 13]. Conversational artificial intelligence (CAI) agents are computer programs and systems designed to engage in natural language conversations by incorporating machine learning (ML), virtual reality (VR), and natural language processing (NLP) to understand and generate human-like text or speech responses [14, 15]. These agents provide recommendations based on users’ past behaviour, preferences, and contextual information [16, 17]. The integration of conversational AI agents into e-tail settings is a testament to the dynamic evolution of technology in reshaping human–computer interaction. Such agents aim to replicate not only the functionality but also the social dimensions of a personal shopping assistant [18, 19]. As the importance of these agents in e-tail continues to grow, it becomes paramount to explore the broader implications of their social presence in order to harness their benefits.

In the world of conversational AI agents, the term “social presence” carries a unique significance. It describes the degree to which users perceive these AI agents as social entities, almost as if they were engaging with another human being [20, 21]. These agents, with their increasingly human-like features, mannerisms, and conversational patterns, evoke the intricate dynamics of human-to-human interaction [22]. Yet, as AI blurs the lines between technology and humanity, it raises concerns about how users perceive and accept these agents. The concept of social presence introduces a new twist: technology is no longer just a tool for communication; it becomes a companion, actively shaping the very experiences it facilitates. Prior research has highlighted the transformative potential of conversational AI agents in the e-tail domain [23, 24]. Previous studies have indicated that AI agents improve user engagement and sales by providing a more natural, human-like shopping experience, enhanced by their ability to offer expert recommendations [25, 26].

While much of the focus has been on technical advancements, the human side of AI adoption is equally critical. Despite the transformative potential of conversational AI agents in enhancing e-tail user experiences, there exists a gap in understanding the nuanced facets of social presence, user perceptions, and trust within this context. While conversational AI agents aim to replicate human-like interactions, users may still harbor reservations and scepticism, leading to challenges in broader acceptance and adoption. This gap underscores the need to investigate the complex interplay between anthropomorphism, technology, and human behaviour to shape positive user perceptions and foster trust in conversational AI agents within e-tail settings [27, 28]. To foster broader user acceptance, it is imperative to understand how to shape positive user perceptions in conversational AI agents. Addressing this gap is vital for the effective integration of these agents in various online domains [29]. Our research aims towards understanding the psychological and behavioural impact of AI-driven agents on e-commerce users. Although prior studies have examined technical efficiency and operational benefits, little attention has been paid to how users perceive these agents and what drives their acceptance or rejection. By exploring these human factors, our study aims to shed light on the broader implications of conversational AI in online retail environments. Additionally, the research seeks to extend the framework of the Technology Acceptance Model (TAM) by incorporating key variables relevant to the adoption of conversational AI agents, contributing to the theoretical understanding of technology acceptance in the context of e-commerce. TAM was chosen as the theoretical framework for this research because of the framework’s long history and adaptability to encompass variables that are integral to the context of technology adoption [30,31,32,33]. Additionally, the significance of trust and social presence in technology adoption and online interactions further justifies their inclusion in the model. The subsequent sections provide an overview of the foundational concepts that underpin our study, outlining the methodology employed, presenting the resulting data, and discussing the implications of these findings.

2 Background and hypothesis development

2.1 Conversational AI agents and anthropomorphism

A conversational agent is a computer designed system that mimics human-like conversations using natural language, vocally or in-text [14, 34]. These agents may be observed in an array of applications, performing diverse and fascinating functions such as customer support [35, 36], education [37], e-tail [38], healthcare [39], fintech [40], and retail [41]. Conversational agents offer easy interfaces, are available 24 h a day, deliver rapid replies, are omnichannel, and can participate in human-like discussions [42, 43]. These agents have a significant influence on electronic businesses. Leading companies, including IBM with its Watson Assistant [44] and Soul Machines with its digital avatars like Roman (Fig. 1), are increasingly outsourcing conversational AI agents. This strategic approach enables e-tailers to harness specialized expertise and integrate advanced conversational agents, exemplifying a contemporary trend in leveraging external capabilities for enhanced technological solutions.

Fig. 1
figure 1

Roman, a digital avatar created by Soul Machines. Used with permission

Gestures (visual cues), voice (spoken cues), and natural written language can all be used to provide input to agents (linguistic cues) [45,46,47]. Previous studies on conversational AI agents have largely focused on their technical performance and operational efficiency. These works have documented the speed and accuracy with which AI agents can handle customer queries, as well as their ability to scale customer service operations. However, they have often overlooked the psychological and social dynamics that play a crucial role in user acceptance.

Anthropomorphism, on the other hand, is the attribution of human-like qualities to non-human objects [48,49,50]. In line with the Computers Are Social Actors (CASA) paradigm [51,52,53] when conversational agents have anthropomorphic clues (e.g., name, gender, and typing emulation), users react to them as if they were human beings [54].

2.2 Technology acceptance model (TAM)

The Technology Acceptance Model (TAM) has gained widespread recognition as the leading framework for understanding technology adoption, particularly when it comes to assessing the acceptance of new technologies [33, 55, 56]. TAM’s core premise is that technology users make rational decisions about how to engage with a given technology. These decisions hinge on two primary factors: perceived usefulness (PU) and perceived ease of use (PEOU), which together shape user attitudes and behaviours [57, 58]. Given TAM’s robust approach to analysing technology adoption, it became the central framework guiding this study.

The Technology Acceptance Model (TAM) has emerged as a cornerstone in understanding the factors that influence technological acceptability. Its strengths lie in its consistent measurement techniques, empirical reliability, and conceptual simplicity [24, 59, 60]. Moreover, TAM has demonstrated its versatility by explaining a significant portion of the variance in users’ intentions to use technology [61,62,63]. This wide applicability is further reinforced by its use across numerous studies, leading to a robust set of questions and metrics for evaluating technology adoption. However, while TAM serves as a reliable framework, it is important to consider additional factors unique to the technology being examined to capture the complete dynamics of its adoption. This comprehensive perspective helps ensure that the insights gained are both accurate and relevant to the context of the technology in question.

2.3 The effect of humanoid embodiment on social presence

Humanoid embodiment refers to the degree to which a conversational AI agent resembles a human in terms of appearance, voice, and behaviour [47, 64]. Previous research has suggested that humanoid embodiment can enhance the perceived social presence of conversational AI agents, which is the feeling of being in a social interaction with another intelligent being [65,66,67]. As conversational AI agents become more humanlike, they trigger more social desirability in user responses [54]. The physical embodiment of humanoid robots can generate a social effect on humans, such as enhancing their attention and memory, as well as reducing their anxiety and boredom [68, 69]. These studies provide evidence that humanoid embodiment positively influences perceived social presence in conversational AI agents. Hence, we hypothesize:

  • H1: Humanoid embodiment positively influences perceived social presence in conversational AI agents.

Conversational AI agents can increase users’ sense of social presence by providing a more personal and human-like interaction [70, 71]. Users are more likely to perceive the agent as social when the agent uses natural language, provides personalized recommendations, and has a more human-like personality [72, 73]. Conversational AI agents with a higher level of social presence gain more user acceptability and behavioural intent to adopt these agents [74].

Thus, it is hypothesized:

  • H2: Perceived social presence positively influences users’ attitudes toward conversational AI agents.

2.4 Effects of transparency and data privacy on trusting beliefs

Recent research underscores the pivotal role of transparency in user data collection as a catalyst for fostering trusting beliefs in conversational AI agents [75]. Users are more likely to trust recommendations when they have a clear understanding of how these recommendations are generated and personalized [76, 77]. Similarly, when AI agents provide users with insights into the decision-making processes behind recommendations, users are more inclined to trust the AI agent’s suggestions [78, 79]. These studies collectively highlight the significance of transparency as a key element in building trusting relationships between users and conversational AI agents, shedding light on its fundamental role in enhancing user experiences and acceptance of AI-driven recommendations.

Thus,

  • H3: Transparency in recommendations positively influences users’ trusting beliefs in conversational AI agents.

Data privacy is introduced as a key variable as it affects users’ sense of safety online [80, 81]. Data privacy risk is an inevitable step when examining hazards online, since technological advancement has raised security concerns about online identity theft and exploitation of user information [82, 83]. Furthermore, privacy risk is a barrier to the adoption of conversational AI agents since users do not have complete control over their information and are concerned about their personal information being sold to other parties. Users are more inclined to trust AI agents when there are strong data privacy safeguards in check [84, 85]. Users exhibit greater trust when they believe their data is secure and strict privacy measures are in place [86].

Hence, we hypothesize,

  • H4: Data privacy protection positively influences users’ trusting beliefs in conversational AI agents.

2.5 Trusting beliefs

From an e-tailer’s standpoint, trust can be defined as “the subjective probability by which the users expect that a website will perform a given transaction by their confident expectation” [87]. Users bear a certain risk while shopping online due to the nature of internet shopping [88]. When users are faced with unknown risks, trust can help them overcome this dilemma [89, 90]. In the absence of risk, trust becomes less crucial. Risk introduces a degree of unpredictability to the purchasing process, making trust an important tool for managing uncertainty and ensuring a smooth transaction [91, 92].

Researchers have investigated the significance of trust and its crucial function in online interactions and purchasing behaviour, emphasizing its capacity to yield favorable outcomes as anticipated [93, 94]. The TAM framework incorporates trust in numerous ways. Previous research has also found that trust is a factor of attitude as it influences users’ attitudes [95, 96].

Hence, it is hypothesized:

  • H5: Users’ trusting beliefs positively influence users’ attitudes toward conversational AI agents.

2.6 The interplay of perceived usefulness and perceived ease of use

Perceived usefulness is defined as “the extent to which individuals believe that using the new technology will enhance their task performance” [97, 98]. The usefulness of a technology is critical to its acceptability. This notion was expanded onto the adoption, or willingness to adopt, conversational AI agents. According to research, usefulness has a significant beneficial effect on behavioural intention to use AI agents [99, 100]. Empirical evidence from prior studies indicates that, within the technical domain, perceived usability plays a crucial role in promoting the adoption and use of a specific technology [101, 102]. Thus, it is reasonable to investigate the influence of perceived utility on user attitudes toward conversational AI agents. Conversational agents’ value must be recognized since they may provide personalization, social presence, expert suggestions, adaptability, and convenience. As a result, it is hypothesized:

  • H6: Perceived usefulness positively influences users’ attitudes towards conversational AI agents.

The perceived ease of use of a technology is defined as “the degree to which an individual believes that using a particular technology will be free of mental effort” [33]. The ease of use of technology is a sign of its acceptability. Users readily embrace technology that is easy to understand and use.

Studies indicate that technology users often have preconceived expectations regarding the ease or difficulty of utilizing a certain technology [103,104,105]. To understand users’ expectations, researchers need to explore their perceptions of ease of use. Studies indicate that ease of use plays a pivotal role in shaping users’ views on technological inventions, highlighting its importance as a factor to consider in research on user behaviour [106,107,108].

Perceived ease of use has a direct influence on perceived usefulness in the Technology Acceptance Model. This direct effect represents the immediate influence that users’ perceptions of a technology’s ease of use exert on their assessment of its usefulness. When users deem a technology as simple to use, their sense of its usefulness improves. Hence, we hypothesize:

  • H7a: Perceived ease of use positively influences perceived usefulness of conversational AI agents.

  • H7b: Perceived ease of use positively influences users’ attitudes towards conversational AI agents.

2.7 Users’ attitudes and intentions

Behaviour towards technology is a direct result of components such as behavioural intention, which is generated by a user’s attitude [109]. Attitude is referred to as a user’s favorable or unfavorable feelings around the usage of AI agents on the e-tail platform [110]. The user’s perception of the consequences of their conduct has a substantial impact on their willingness to behave in that manner [111]. As a result, it is pertinent to link user attitude and intention to use, which is hypothesized as follows:

  • H8: Users’ attitudes toward the agent positively influence their intention to adopt conversational AI agents.

The background review identified a gap in understanding human interaction with AI agents. This gap includes the social presence of these agents, user perceptions of their reliability and trustworthiness, and the broader behavioural trends in e-tail platforms. By targeting these areas, our research offers a fresh perspective on AI’s role in e-commerce and aims to fill a critical void in the existing literature.

3 Research framework

The research hypotheses are depicted in the conceptual framework given in Fig. 2. This study investigates the direct impact of social presence, trusting beliefs, perceived utility, and perceived ease of use on users’ attitudes about conversational AI agents and behavioural intents to adopt them, as indicated in the conceptual framework. Furthermore, we investigate the significance of transparency and data privacy in establishing trusting beliefs in conversational AI agents, as well as the function of humanoid embodiment in replicating the agent’s social presence.

Fig. 2
figure 2

Conceptual model and hypotheses

4 Research methodology

4.1 Measurements

The present study employs a comprehensive measurement approach featuring a scale comprised of 37 items designed to assess nine distinct variables. The questionnaire utilized for data collection drew upon well-established and pre-validated scales, with careful adaptations to tailor their relevance to the context of our study. Participants in the research were requested to express their perspectives on these variables utilizing a “five-point Likert scale.” This Likert scale, spanning from one to five, assigned the numerical values one for a position of “strongly disagree” and five for a standpoint of “strongly agree.” The items for the construct of humanoid embodiment were adopted from Keeling et al. [112], social presence from Gefen and Straub [95], transparency from Matsui and Yamada [113], data privacy from Zhang et al. [82], trusting beliefs from Qiu and Benbasat [114], perceived usefulness and perceived ease of use from Davis [115], attitude towards technology from Qiu and Benbasat [65], and behavioural intentions from Kim et al. [55].

4.2 Procedure

An online survey instrument was constructed using Google Forms, accompanied by a comprehensive message explaining the study’s objectives and rationale. The study employed a detailed research procedure to examine the effectiveness and user perceptions of conversational AI agents in the e-tail context. An integral component of the study involved the integration of a video sourced from Soul Machines, an online platform specializing in the development of conversational AI agents for e-tail applications [116]. This video featured a simulated conversational interaction between an AI agent and a human user, showcasing the practical implementation of these agents in providing products and services.

Subsequently, participants were systematically presented with the aforementioned video and then prompted to respond to a series of inquiries crafted to elicit their perceptions and reactions to the demonstrated interaction. The questions encompassed diverse facets, including user experience, perceived usefulness, social presence, and trust in the AI agent.

4.3 Sample and data collection

This study focused on a specific target population comprising individuals who possess familiarity with the concept of conversational AI agents in the context of e-tail platforms. To gather data from this selective population, a purposive sampling technique was employed. Notably, the demographic distribution of the respondents revealed a predominant representation within the age bracket of 25 to 34 years. The data collection process garnered a total of 180 survey responses, from which a judiciously selected subset of 158 responses was deemed suitable for subsequent analysis. This process of purposive sampling and data collection ensured a focus on respondents with a relevant knowledge base, thereby enhancing the study’s relevance and comprehensiveness.

The hypotheses were tested using the partial least squares structural equation modeling (PLS-SEM) tool. PLS-SEM was employed for the analysis due to its adaptability to data needs, model complexity, and relationship specifications.

5 Findings

5.1 Measurement model assessment

At first, the measurement model is tested for item factor loadings, construct’s convergent reliability, and discriminant validity followed by path analysis for the structural model [117]. Table 1 displays the factor loadings, Cronbach alpha (for construct reliability), and average variance explained (AVE) values to ensure the convergent reliability of the constructs. The value for factor loadings should be above 0.70 to be accepted [118]. The Cronbach alpha values should be between 0.70 and 0.90 to be reliable and the AVE values should be above 0.50 [117]. The factor loadings in our model are above the threshold value, and the Cronbach alpha and AVE are also above the threshold value and thus are accepted. This validates the reliability of our model.

Table 1 Standardized item loadings, AVE, and Cronbach alpha values

The discriminant validity of the constructs is assessed using Heterotrait-Monotrait ratio. The values for the HTMT ratio should be less than 0.85 [119]. Table 2 shows the HTMT values which are less than 0.85 and thus, our model has discriminant validity.

Table 2 The Heterotrait-Monotrait ratio for discriminant validity

Through the convergent and discriminant validity analysis, we can say that our measurement model is reliable and all the constructs are distinct from each other.

5.2 Hypothesis testing

We further performed bootstrapping using 5000 samples to test our hypotheses. Results are tabulated in Tables 3 and 4. A visual representation of the model in PLS-SEM is presented in Fig. 3.

Table 3 Path coefficients and their significance
Table 4 Hypothesis testing and results
Fig. 3
figure 3

Structural equation model

5.3 Structural model assessment

5.3.1 Evaluation for explanation power

The model was tested for explanation power (R2), as depicted in Table 5. The explanation power of 0.25, 0.50, and 0.70 implies weak, moderate, and strong relations respectively. For our analysis, a moderate relationship was established. The f2 value helps in assessing the impact of the independent variable on the dependent variables. For f2 values, 0.02, 0.15, and 0.35 denote small, medium, and large effects [118]. Our study shows a large effect size for the intention. Through Q2, the predictability of the model is assessed. Any value above zero is acceptable to suggest accuracy; for our model, the values were 0.561 and 0.556 for intention and attitude respectively. Hence, the accuracy of the model was suggested.

Table 5 Structural model results

5.3.2 Goodness of fit index

Standardized root means square values (SRMR) are used to check the goodness of fit of the model [118]. A value less than 0.08 shows a good fit, and this model suggests a good fit with an SRMR value of 0.060.

6 Discussion

As the adoption of conversational AI by users has led to a paradigm shift across various domains, the interaction between individuals and technology has been revamped. Despite the challenges posed by these technological advancements, users are facilitating the adoption of these “untact” services, which allow them to connect with websites without face-to-face interaction [120]. This study delves into the key AI and technology-related variables which impact users’ behavioural intentions to adopt AI agents. Our findings explored the interplay between AI-related factors, technology-related factors, and human-related factors which can shape a positive attitude towards these AI agents and also further facilitate the adoption by e-tailers. Our research showcased the growing acceptance of these conversational AI agents by users across the e-tail domain.

From our study, we found that users are engaging with AI agents for several tasks ranging from product selection to cart management during their e-tail journey. First, transparency (β = 0.435, p = 0.013) was found to be a key determinant of trusting beliefs and thus, more transparent AI systems can help in building the trust of the users, whereas persistent concerns surrounding data privacy underscore the need for enhanced protective measures to gain users’ trust and acceptance. Second, humanoid embodiment (β = 0.660, p = 0.001) was a significant predictor of social presence, which suggests that when conversational AI agents exhibit human-like characteristics or appearance, users are more likely to perceive them as social entities. This finding highlights the importance of designing AI agents with anthropomorphic features to enhance users’ sense of social presence during interactions. It implies that the degree to which AI agents resemble humans plays a crucial role in shaping users’ perceptions and experiences, ultimately influencing their engagement and acceptance of the technology.

Third, the paper confirmed the impact of trusting beliefs (β = 0.201, p = 0.013), social presence (β = 0.268, p = 0.001), perceived usefulness (β = 0.394, p = 0.001), and perceived ease of use (β = 0.140, p = 0.001) on attitude. The level of trust in these AI agents significantly impacts the attitude of users; thus, the e-tail websites should focus on building more secure and transparent AI agents which can enhance users’ level of trust and facilitate a positive attitude. As social presence was another determinant of attitude, it can be inferred that an agent with a high social presence or a sense of social interaction can lead to a more favorable user attitude than one with a low social presence. Thus, e-tail websites should aim to develop more human-like agents. The study also found that when users perceive the agents as easy to use, they are more likely to find them useful. This result underscores the importance of a seamless user experience. Users’ perception of AI agents as useful has a positive impact on their overall attitude toward them. This finding aligns with TAM which suggests that perceived usefulness is a key driver of technology acceptance. A positive attitude (β = 0.789, p = 0.001) of users can strongly impact users’ intentions towards these AI agents. Overall, these findings have a range of theoretical contributions and practical implications, which we explore in the following sections.

6.1 Theoretical contributions

From a theoretical perspective, this study contributes to the growing body of literature on conversational AI agents in e-commerce and e-tail platforms, focusing on user perceptions and social presence. It underscores the influence of anthropomorphism on user behaviour, shedding light on the underlying psychological mechanisms. According to the present research, anthropomorphizing conversational AI agents offer a more personalised shopping experience to e-tail users. Our findings suggest a strong correlation between anthropomorphized agents and a higher user acceptance, adding to the theory of anthropomorphism. Secondly, this study also significantly contributes to the theory of social presence by exploring the intricate dynamics of user interaction with conversational AI agents on e-tail platforms. This study also contributes to academia by extending the TAM framework to incorporate key variables relevant to the study of conversational AI agents. By enhancing the TAM framework, this research provides a robust model that offers a comprehensive understanding of user behaviour and acceptance of AI-driven conversational agents in various contexts. Academicians can utilize this refined model to guide future research endeavors in conversational AI technology.

6.2 Managerial implications

The deployment of conversational AI agents on e-tail platforms carries profound managerial implications. Firstly, these agents can significantly enhance customer support by providing uninterrupted service, particularly during non-working hours, thus reducing dependency on human agents. Secondly, the ability of AI agents to deliver personalized product recommendations fosters user satisfaction and cultivates brand loyalty. Furthermore, these agents serve as invaluable sources of real-time data, facilitating research and development initiatives and informing robust marketing strategies. The multilingual capabilities of AI agents contribute to global reach, broadening the customer base but the issue of user data privacy prevails. E-tailers ought to adhere to data privacy regulations and prioritize the trust and integrity of users.

The study faces some challenges, including addressing user reservations and scepticism towards AI agents and navigating the complexities of social presence in technology-mediated interactions. Future researchers can be directed towards these issues as overcoming these challenges is essential for advancing our understanding of user behaviour and technology adoption in the rapidly evolving e-tail landscape.