Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Exploring consumers' response to text-based chatbots in e-commerce: the moderating role of task complexity and chatbot disclosure

Exploring consumers' response to text-based chatbots in e-commerce: the moderating role of task... Purpose – Artificial intelligence (AI)-based chatbots have brought unprecedented business potential. This study aims to explore consumers’ trust and response to a text-based chatbot in e-commerce, involving the moderating effects of task complexity and chatbot identity disclosure. Design/methodology/approach – A survey method with 299 useable responses was conducted in this research. This study adopted the ordinary least squares regression to test the hypotheses. Findings – First, the consumers’ perception of both the empathy and friendliness of the chatbot positively impacts their trust in it. Second, task complexity negatively moderates the relationship between friendliness and consumers’ trust. Third, disclosure of the text-based chatbot negatively moderates the relationship between empathy and consumers’ trust, while it positively moderates the relationship between friendliness and consumers’ trust. Fourth, consumers’ trust in the chatbot increases their reliance on the chatbot and decreases their resistance to the chatbot in future interactions. Research limitations/implications – Adopting the stimulus–organism–response (SOR) framework, this study provides important insights on consumers’ perception and response to the text-based chatbot. The findings of this research also make suggestions that can increase consumers’ positive responses to text-based chatbots. Originality/value – Extant studies have investigated the effects of automated bots’ attributes on consumers’ perceptions. However, the boundary conditions of these effects are largely ignored. This research is one of the first attempts to provide a deep understanding of consumers’ responses to a chatbot. Keywords Text-based chatbot, Trust, Consumers’ response, Task complexity, Identity disclosure Paper type Research paper © Xusen Cheng, Ying Bao, Alex Zarifis, Wankun Gong and Jian Mou. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) licence. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and Internet Research non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this licence may be seen at http://creativecommons.org/licences/by/4.0/legalcode pp. 496-517 Emerald Publishing Limited This work is supported by the Fundamental Research Funds for the Central Universities, and the 1066-2243 DOI 10.1108/INTR-08-2020-0460 Research Funds of Renmin University of China (Grant No.21XNO002). 1. Introduction Consumers’ Disembodied conversational agents (DCAs) have now become quite common (Araujo, 2018; response to Luo et al., 2019). Galvanized by artificial intelligence (AI) and machine learning, DCAs in the text-based form of chatbots can simulate human interactions through text chats or voice commands in chatbots various fields and have brought unprecedented business potential (Luo et al., 2019). A chatbot is defined as “a machine conversation system which interacts with human users via natural conversational language” (Shawar and Atwell, 2005, p. 489). Specifically, the market size of chatbots is expanding from $250m in 2017 to over $1.34bn by 2024 (Pise, 2018). Many famous brands and major platforms, such as eBay, Facebook, WeChat, Amazon and Apple’s Siri, have rolled out chatbots to take orders, recommend products or provide other customer services (Thompson, 2018; Go and Sundar, 2019). Gartner predicts that by 2021, almost 15% of the customer services will be completely handled by AI (Mitchell, 2018). AI-enabled chatbots have provided several unique business benefits in the e-commerce setting. First, the chatbots are able to understand consumers’ requests and automate the customer services, addressing their requests effectively (Daugherty et al., 2019). Moreover, compared to a human, the AI-enabled chatbot will not experience negative emotions or work exhaustion and can always interact with consumers in a friendly manner (Luo et al., 2019). Additionally, the chatbots can easily handle a large number of customer communications at the same time and increase the customer-service efficiency. While the adoption of AI in e-commerce can bring several business benefits for the companies or platforms from the perspective of the supply side, a key challenge for the application of chatbots is the customers’ perception from the perspective of the demand side (Froehlich, 2018). Although the chatbots can interact with consumers in a more friendly way compared with the customer service of employees, attributes of the technology have created customer pushback. For example, consumers may find it uncomfortable talking with a technology-based chatbot for their specific personal needs or purchasing decisions. Moreover, consumers may have a prejudice that the computer programs lack empathy and personal feelings, perceiving the chatbots as less trustworthy than humans. This negative perception of AI-enabled chatbots will even have negative effects on their perception of the brands or companies. Therefore, firms should apply the chatbot wisely to provide better customer service. For instance, firms can increase effectiveness by introducing chatbots to complete easy customer requests, but the technology might not be applied for complex tasks when customers have very demanding requests, which need employees to intervene (e.g. service recovery, price negotiation). As a result, there is a call for a deeper understanding of consumers’ perceptions and responses to a text-based chatbot in the e-commerce setting, involving the effects of the chatbots’ specific attributes. Extensive research exists on AI-enabled systems in business practice, including social robots (de Kervenoael et al., 2020), machine “teammates” (Seeber et al., 2020) and chatbots (Luo et al., 2019). Findings indicate that attributes of the AI-enabled computer programs, such as anthropomorphism (Castelo et al., 2019; Desideri et al., 2019; Yuan and Dennis, 2019), humanness (Wirtz et al., 2018), empathy (de Kervenoael et al., 2020) and adaptivity (Heerink et al., 2010) will significantly impact consumers’ perception, decision-making (Duan et al., 2019), purchase intention (Mende et al., 2019) and use intention (Heerink et al., 2010; de Kervenoael et al., 2020). Although existing studies have drawn attention to consumers’ positive perception of the application of text-based chatbots in the e-commerce setting, current empirical evidence on consumers’ pushback is still limited. Moreover, consumers’ perceptions or responses may differ in different conditions. While the boundary conditions of consumers’ perception and response behavior have been largely ignored in existing studies. Against this backdrop, we address both the practical and theoretical aspects with the following research questions: RQ1. What is the impact of the unique attributes of a text-based chatbot on consumer INTR responses to the text-based chatbot in the e-commerce context? 32,2 RQ2. How does the task complexity and chatbot identity disclosure moderate the relationship between these attributes and consumer responses? RQ3. How does consumers’ trust impact their responses to the text-based chatbot in e- commerce? In order to answer these research questions, we draw upon the stimulus–organism–response (SOR) framework (Mehrabian and Russell, 1974) and test a model of consumers’ trust and subsequent responses to a text-based chatbot in the e-commerce setting. We also integrate the moderating role of task complexity and consumers’ perception of the disclosure of the chatbot to test the boundary conditions of different consumer responses. The SOR framework provides an ideal theoretical framework to examine the links between stimulus, organism and response systematically (Kim et al., 2020). Therefore, it is suitable to address our research questions. This paper is structured as follows. First, we introduce the theoretical background and present the hypotheses development. Next, we present the research methodology and data analysis results. Finally, we conclude with a brief summary of the findings, the theoretical contributions and the practical implications. Limitations and future research are also provided. 2. Theoretical background and hypotheses development 2.1 Stimulus–organism–response model The “stimulus-organism-response framework” was initially proposed by Mehrabian and Russell (1974) and then modified by Jacoby (2002). The SOR framework suggests that individuals will react to the environmental factors and have certain behaviors (Kamboj et al., 2018). Specifically, individuals’ reaction behaviors consist of two typical forms: approach behaviors with positive actions, such as the preference to affiliate, explore and averting behaviors with the opposite actions, such as the desire to react negatively (Kim et al., 2020). Existing studies have extended the SOR framework to many areas, including the online shopping experience (Eroglu et al.,2003), website experience (Mollen and Wilson, 2010), consumer behavior in the retail industry (Rose et al., 2012) and tourism industry (Kim et al.,2020). Among the abovementioned contexts, e- commerce is one of the most representative. The SOR theory has been adopted in the e-commerce context as a psychology theory to investigate consumer behavior (Fiore and Kim, 2007; Chang et al., 2011; Wu and Li, 2018). For example, Eroglu et al. (2003) argued that the environmental or atmospheric cues of the online platform (S-stimuli) can affect consumers’ cognition or emotion (O-organism), which will then influence their behavioral outcomes (R-response). Similarly, in this research, we adopted the SOR framework to explain consumers’ organism process and behavior outcomes toward the text-based chatbot in the e-commerce setting. The SOR framework consists of three elements: stimulus, organism and response (Kamboj et al., 2018). Similar to the “input-response-output framework”, the “stimulus,”“organism” and “response” correspond to the “input,”“response” and “output.” More specifically, “stimulus” is defined as the environmental factors encountered by individuals (Jacoby, 2002), which will arouse them (Eroglu et al., 2001). In this research setting, “stimulus” is defined as the unique attributes of the text-based chatbot that will arouse the consumers. The second element “organism” can be defined as individuals’ affective and cognitive condition and their process of subconsciously intervening between the stimuli and responses (Jacoby, 2002). For example, in the commercial setting, the “organism” includes the individuals’ cognitive and affective perceptions of their experience as a consumer (Rose et al., 2012). The third element “response” has been described as individuals’ behavioral consequences or outcomes to the environment (Donovan and Rossiter, 1982). As aforementioned, individuals’ response to the environment includes both positive actions and negative actions. For example, in the Consumers’ commercial setting, a positive response may include consumers’ willingness to stay, explore response to or affiliate (Bitner, 1992), while a negative response may refer to individuals’ willingness to text-based react negatively. In this research setting, we take both consumers’ positive and negative chatbots reactions toward the text-based chatbot into consideration. 2.2 Stimulus: unique attributes of a text-based chatbot Turing proposed the Turing test to determine AI or machine intelligence in 1950 (Turing and Haugeland, 1950). A machine that passed the Turing test can be identified as more realistic and human-like, which makes it more compelling for consumers (Gilbert and Forney, 2015). Thus, the realistic and human-like characteristics of the chatbots will have a great impact on the usability of the chatbot. In the e-commerce service, AI has been applied in the form of text-based chatbots that can respond to consumers’ requests automatically (Kaplan and Haenlein, 2019). The AI-enabled text-based chatbot has been widely used in the online purchasing process. On the one hand, the service provided by a text-based chatbot can be more customer-friendly when compared with some traditional human customer service. According to social response theory (Nass et al., 1994) and anthropomorphism theory (Blut et al., 2021), customers will perceive higher satisfaction during the interaction with human-like machines. While on the other hand, some consumers perceive that these chatbots always give rigid or standard answers and are not empathetic enough. As a result, we mainly focus on the effects of these two attributes (empathy and friendliness) of a text- based chatbot on consumers’ outcomes. 2.2.1 Empathy of the chatbot. Empathy can be defined as “ability to identify, understand, and react to others’ thoughts, feelings, behavior, and experiences” (Murray et al., 2019, p. 1366). It is a multidimensional construct that covers both cognitive and affective perceptions (de Kervenoael et al., 2020; Powell and Roberts, 2017). In extant research on service management and information systems, empathy can be defined as the required skills for successful interactions between consumers and robots (Birnbaum et al., 2016). According to Czaplewski et al. (2002)’s RATER model, empathy is also categorized as one of the five dimensions (reliability, assurance, tangibles, empathy and responsiveness) of service quality (Czaplewski et al., 2002). In the traditional service setting, employees have been trained to always take the consumers’ point of view. Even if many consumers may ask similar questions every day, professional employees are not supposed to be impatient and they are trained to always be empathetic and offer alternatives to address consumers’ problems. In the online service setting, some text-based chatbots have been set to automatically reply to consumers when there is a need. The chatbot with a high level of empathy can understand the consumers’ needs better and provide alternatives to address their issues. As a result, consumers will be more likely to have a positive perception of the services provided by the chatbot. 2.2.2 Friendliness of the chatbot. In a traditional business setting, employees are required to be friendly to consumers and convey positive emotions to them (Tsai and Huang, 2002). Friendly and helpful service providers can please consumers and increase their positive perception of the service quality (Chen et al., 2012). Following Pugh (2001), when employees convey more positive emotions, consumers will be more satisfied with the service, resulting in their positive response to the store, such as repurchase and positive word of mouth. In previous research on consumers’ acceptance of robots, users gave high ratings on the effects of friendliness and trust (Fridin and Belokopytov, 2014). When interacting with the text- based chatbots, rigid or indifferent responses from the chatbot will result in consumer pushback. On the other hand, a friendly manner will positively impact consumers’ moods and perceptions of the store and brands (Tsai and Huang, 2002). 2.3 Organism: trust toward the text-based chatbot INTR As aforementioned, “organism” refers to individuals’ cognitive or affective condition. The 32,2 formation of trust also involves individuals’ emotions and feelings. For example, individuals may prefer not to trust others simply because of their perceived bad feelings (Hoff and Bashir, 2015). Therefore, the process of consumers’ trust formation on the text-based chatbot can be regarded as the “organism” in this research setting. Specifically, trust can be defined as the individual or group’s willingness to be vulnerable to the other party (Hoy and Tschannen-Moran, 1999; Mayer et al., 1995). Generally, trust has been widely discussed in many research fields in recent years, such as psychology, economics, sociology and information systems (Breuer et al., 2016; Ert et al., 2016). In extant research, trust has been investigated mostly in the context of interpersonal interaction. Nowadays, with the growing interaction between humans and robots, trust can also be applied to explain the way people interact with technology (Hoff and Bashir, 2015). In this research, we mainly focus on consumers’ trust toward the text-based chatbot during the online purchasing process. Extant research has investigated the antecedents of trust from several lenses. The most widely accepted antecedents of trust are ability, integrity and benevolence (Jarvenpaa et al., 1997). Considering trust between humans and autonomous machines, anthropomorphic characteristics will impact users’ trust perception (Złotowski et al., 2017). According to social response theory (Nass et al., 1994), the interactions between human and computer are fundamentally social (Adam et al., 2020). Even if the consumers know that the computer does not hold feelings or emotions as humans do, they still tend to perceive the computers or machines as social actors. The theory has also provided evidence on studies investigating how humans adopt social rules to the anthropomorphically designed machines (Adam et al., 2020). In other words, humans will react similarly to those machines with human characteristics as they react to humans (Ischen et al., 2019). Moreover, the anthropomorphism theory also argues that individuals’ human-like perceptions of the robots will facilitate the human–robot interaction, thus increasing customers’ positive perception of the interactions (Blut et al., 2021). As aforementioned, we mainly focus on empathy and friendliness as the anthropomorphic attributes of the chatbots. Based on social response theory and anthropomorphism theory, we can infer that consumers may perceive the text-based chatbots as more trustworthy if they are empathetic and customer-friendly during the interaction with the chatbots. Hence, we postulate the following hypotheses: H1. Empathy of the text-based chatbot is positively related to consumers’ trust toward the chatbot. H2. Friendliness of the text-based chatbot is positively related to consumers’ trust toward the chatbot. 2.4 Moderators: task complexity and disclosure of the text-based chatbot 2.4.1 Task complexity. The customer service in an e-commerce platform is supposed to help consumers to address their problems during the whole customer journey. The nature of customer service is to meet customers’ needs across the stages of prepurchase, purchase and postpurchase (Lemon and Verhoef, 2016; Zhang et al., 2016). In the prepurchase stage, consumers may need to search for information, recognize their needs and consider whether the product will meet their needs. During the interaction process, consumers may ask about the details, functions or discount strategies of the product itself. Thus, the task of the chatbot in this stage is to provide specific information about the products and assist the consumers with decision-making. In the purchase stage, consumers’ behavior includes mainly choice, ordering and payment. During this process, consumers’ interaction with the chatbot may be less than in the other two stages. In the postpurchase stage, consumers’ behaviors mainly Consumers’ contain postpurchase engagement or service requests. According to previous studies, response to interaction between the store and consumer in this stage covers service recovery (Kelley and text-based Davis, 1994), repurchase decision (Bolton, 1998), products returning (Wood, 2001) or other chatbots forms of engagement (van Doorn et al., 2010). Tasks that the chatbot needs to address in this process may be more complex than the other two stages. According to the task–technology fit theory (Goodhue and Thompson, 1995), the technology should meet the customer’s needs or requirements for specific tasks (Blut et al., 2021). Once the technology can meet the customer’s specific needs, the interaction will be more satisfying. Therefore, in addition to the attributes of the chatbots in H1 and H2, more emphasis should be attached to the feature of the tasks that the chatbots address. Based on the task–technology fit theory, the attributes of the task can moderate the relationship between chatbots’ attributes and customers’ intention. Therefore, we focus on the effects of task complexity on customers’ behavioral outcomes. The effects of task complexity have been discussed in previous studies from several facets, such as task performance or team performance (Bjørn and Ngwenyama, 2009; Dayan and Di Benedetto, 2010). As aforementioned, tasks in different customer journey stages are different. Therefore, we applied task complexity in this research setting to better understand consumers’ trust toward the chatbot in different contexts. When the problem is complex to address, consumers may turn to customer service for help. In this circumstance, they may pay more attention to the customer service providers’ level of knowledge and professionalism. Thus, only giving rigid and standardized responses with friendliness and empathy cannot improve consumers’ perceptions. In contrast, they may even doubt the chatbot’s ability to address the problems. Consequently, we propose the following hypotheses: H3a. Task complexity moderates the relationship between empathy and consumers’ trust toward the chatbot, such that the positive relationship is weaker when the task is complex. H3b. Task complexity moderates the relationship between friendliness and consumers’ trust toward the chatbot, such that the positive relationship is weaker when the task is complex. 2.4.2 Identity disclosure of the chatbot. According to Davenport (2019), one way to increase trust in automation is to increase the disclosure of the system or its adoption as much as possible (Davenport, 2019). Consumers also have the right to be informed of the identity of the chatbot (Luo et al., 2019). Disclosure in the e-commerce setting refers to the state when the text-based chatbots reveal their machine identity before interacting with consumers (Luo et al., 2019). For example, in some online stores, the customer service will disclose its identity using the following sentences “Dear consumer, welcome to ABC store and I am the AI chatbot of this store” at the beginning of the interaction with them. Some stores also choose to state the AI identity beside the profile photo of the customer service staff. While a text-based chatbot can provide several business benefits for the company, some consumers may always feel uncomfortable talking with chatbots and perceive them as less empathetic than traditional customer service employees. While they are informed that they are interacting with a chatbot instead of a human, the positive effects of chatbots’ empathy and friendliness will be increased due to the empathetic and friendly service. Accordingly, we propose the following hypotheses: H4a. Disclosure of the chatbot moderates the relationship between empathy and consumers’ trust toward the chatbot, such that the positive relationship is stronger when the identity of the chatbot is disclosed. H4b. Disclosure of the chatbot moderates the relationship between friendliness and INTR consumers’ trust toward the chatbot, such that the positive relationship is stronger 32,2 when the identity of the chatbot is disclosed. 2.5 Response: avoidance/approach behavior toward the text-based chatbot According to the SOR framework, consumers’ response behavior consists of both averting behaviors and approach behaviors (Kim et al., 2020). Extant studies mainly focus on consumers’ approach behaviors or intention toward the automated bots. To provide a holistic view of consumers’ response to the chatbots, we take both consumers’ averting response (resistance) and approach response (reliance) to the text-based chatbot into consideration. 2.5.1 AI reliance. Individualstendtorelymoreonahumanratherthananalgorithm for forecasting stock prices (Onkal et al.,2009). Unlike humans, automated algorithms cannot provide reasons or explanations for their decision, thus increasing consumers’ distrust. Although previous research has investigated consumers’ future intention with the bots (Heerink et al., 2010; de Kervenoael et al., 2020), there has been little investigation and validation of practical interventions for the customer service provider to increase consumers’ reliance on the automated machines instead of a human, especially in the e-commerce setting. Therefore, we tend to provide empirical evidence on consumers’ reliance on the AI-based chatbot. As discussed in the literature on trust, cognitive trust involves individuals’ confidence in others’ performance and reliability. Therefore, we suggest that if consumers perceive the chatbot as trustworthy during the interaction process, they will be more likely to rely on the chatbot to address their problems or provide decisions. Thus, we hypothesize that: H5. Consumers’ trust toward the chatbot is positively related to their reliance on the chatbot. 2.5.2 AI resistance. Extant research on consumers’ response to AI mainly focuses on their acceptance or future use intention of the AI-based bots (Ciechanowski et al., 2019; Fridin and Belokopytov, 2014; Heerink et al., 2010; Piçarra and Giger, 2018; Wirtz et al., 2018). Although consumer pushback has been proven to be a major concern, only a few studies have investigated consumers’ resistance to the AI-based bots (Longoni et al., 2019). Findings in the healthcare context indicate that AI service providers are less likely to account for consumers’ personal requests or unique characteristics. This can lead consumers to perceive the AI service provider as less trustworthy and they may exhibit resistant behavior toward them. In the e-commerce setting, when consumers find the text-based chatbots are not trustworthy or not able to meet their needs, they will avoid interacting with them and switch to the traditional human customer service (Longoni et al., 2019). Therefore, the following hypothesis is proposed: H6. Consumers’ trust toward the chatbot is negatively related to their resistance to the chatbot. The conceptual model is presented in Figure 1. 3. Research method 3.1 Sample and data collection We measured the variables using existing scales adapted from previous research (see the Appendix). A pretest was conducted with 30 students in our research group to ensure the quality of the survey. Suggestions were also provided by these participants. After Moderators Consumers’ response to Unique attributes of Task complexity text-based Customer outcomes text-based chatbot chatbots H3b H3a Reliance on Empathy H1 the chatbot H5 Trust towards the chatbot H6 H2 Resistance to Friendliness H4b H4a the chatbot Figure 1. A conceptual model of customers’ response to Disclosure of the the text-based chatbot chatbot rearranging the survey, the formal version was then distributed mainly through a professional survey website in China via a social network. Several representative chatbots in the e-commerce setting were provided at the beginning of the survey. Those without any chatbot experience during the purchasing process on e-commerce sites were filtered out. The participants were told that their information was anonymous. Finally, a total of 441 respondents completed the survey including the pretest. After discarding incomplete and rushed responses, there were a total of 299 effective responses, which were adequate for this research. Our target participants were consumers who had experienced the text-based chatbot in an e-commerce setting before. Table 1 presents the demographics of these participants. As is shown in the table, the participants consist of 28.8% males and 71.2% females. More than 82.9% of them are below 35 years old. The age range of the participants is consistent with this research setting because young people nowadays interact with text-based chatbots more frequently in the online purchasing process. Additionally, most of the participants are shown to frequently interact with the text-based chatbots in the purchasing process. Therefore, the participants are suitable for our research setting. Constructs and measurement items Category Frequency (N 5 299) % Gender Males 86 28.8 Females 213 71.2 Age <25 147 49.1 26–35 101 33.8 36–45 28 9.4 >45 23 7.7 Experience with the chatbot in e-commerce Seldom 38 12.7 Less 95 31.8 General 106 35.5 Table 1. More 50 16.7 Demographics of A lot 10 3.3 participants (N 5 299) 3.2 Measurements INTR The data was collected in China, so we conducted the forward–backward translation method 32,2 to ensure the validity of translation (Lin et al., 2019). We used multi-item measures with a seven-point Likert-style scale (1 5 strongly disagree, 7 5 strongly agree). The survey included seven reflective constructs, including the unique attributes of text-based chatbots (e.g. empathy, friendliness), moderators (e.g. task complexity, disclosure of the text-based chatbot), organism of the consumers’ perception (e.g. trust toward the text-based chatbot) and responses of the consumers (e.g. resistance and reliance on the text-based chatbot). Control variables such as age, gender and consumers’ experience with a text-based chatbot when purchasing online were also considered. Specifically, the empathy measure was developed following the guidance of previous research (Cheng et al., 2018; Czaplewski et al., 2002; de Kervenoael et al., 2020). The friendliness measure was measured according to Tsai and Huang (2002). The task complexity measure was developed by adopting the measurements by Morgeson and Humphrey (2006). The disclosure measure was evaluated according to the information disclosure measure in extant research (Jensen and Yetgin, 2017; Luo et al., 2019; Starzyk et al., 2006). The trust measure was evaluated according to prior research on trust (Cyr et al., 2009; Everard and Galletta, 2005). The resistance measure was developed following previous research on the resistance to information technology implementation (Kim and Kankanhalli, 2009; Lapointe and Rivard, 2005). The reliance measure was developed according to previous self-reliance measures (Edwards et al., 2003). 3.3 Data analysis We conducted the data analysis with SmartPLS 3.2.8 and SPSS 22.0. Specifically, SmartPLS was used to analyze the reliability, validity and common method variance (CMV) (Shiau et al., 2019), while SPSS was used to test the hypotheses. The data analysis process was conducted in three stages: reliability and validity analysis, CMV and hypotheses testing. 3.3.1 Reliability and validity analysis. We first performed several statistical tests to ensure the reliability and validity of all the measures. As is shown in Table 2, we employed the value of Cronbach’s α to test the internal consistency reliability. The Cronbach’s α of all the measures exceeds the recommended threshold, indicating acceptable reliability. Moreover, the variance inflation factor (VIF) values of the subconstructs are all below the recommended threshold of 10, suggesting acceptable reliability (Hua et al., 2020). The results of VIF also suggest that multicollinearity is not a significant concern in this research (Thatcher et al., 2018). Validity of the construct measures was tested with two procedures. Firstly, each construct’s average variance extracted (AVE) is above 0.5 and most of the constructs’ factor loadings are above 0.7, suggesting satisfactory convergent validity. Discriminant validity was measured with the Fornell–Larcker criterion and cross-loading evaluation. As is shown in Table 3, the square root of AVE for each construct is greater than their correlations with other constructs (Liang et al., 2019). As is shown in Table 4, each indicator’s factor loading is greater than its cross-loadings with other constructs, suggesting satisfactory discriminant validity. 3.3.2 Common method variance. To address the concerns of CMV due to the self-reported surveys, we conducted several procedural and statistical remedies. In terms of procedural remedies, we tried to assure the respondents that they can answer these questions as honestly as possible to reduce the evaluation apprehension and protect their anonymity (Cheng et al., 2020). In addition, we tried to improve the scale items by avoiding vague concepts and keeping the questions simple and specific (Podsakoff et al., 2003). In terms of statistical remedies, we adopted the two most popular tests for the CMV, including Harman’s one-factor test (Cram et al., 2019; Saldanha et al., 2017) and a marker variable technique (Griffith and Lusch, 2007; Hua et al., 2020; Krishnan et al., 2006; Lindell and Brandt, 2000) Specifically, Consumers’ VIF Factor loadings Cronbach’s α Average variance response to Constructs Items (<10) (>0.7) (>0.7) extracted (AVE) (>0.5) text-based Empathy (EMP) EMP1 2.191 0.885*** 0.863 0.785 chatbots EMP2 2.471 0.902*** EMP3 2.081 0.871*** Friendliness (FRI) FRI1 3.469 0.934*** 0.926 0.870 FRI2 4.014 0.937*** FRI3 3.358 0.927*** Task complexity COM1 2.515 0.893*** 0.881 0.807 (COM) COM2 3.511 0.918*** COM3 2.272 0.883*** Disclosure of the DIS1 1.454 0.710*** 0.717 0.733 chatbot (DIS) DIS2 1.454 0.981*** Trust toward the TRU1 1.337 0.671*** 0.782 0.697 chatbot (TRU) TRU2 2.415 0.919*** TRU3 2.161 0.893*** Reliance on the REL1 4.270 0.946*** 0.951 0.910 chatbot (REL) REL2 5.675 0.957*** REL3 5.981 0.960*** Resistance to the RES1 1.767 0.792*** 0.773 0.691 chatbot (RES) RES2 2.442 0.918*** RES3 1.574 0.776*** Table 2. Note(s): VIF – Variance Inflation Factor. The recommended threshold of each indicator is in the Psychometric parentheses; ***p < 0.001 properties of measures Harman’s single-factor test indicated no single factor. Then we conducted the “marker variable” test. Following Lindell and Brandt (2000), a marker variable is defined as a variable that is minimally correlated with the study variables. In this study, we chose age as a marker variable (Griffith and Lusch, 2007). The results suggest that the marker variable is not significantly related to all the variables in this model. Furthermore, the structural model paths of interest maintain statistical significance after including the marker variable. Therefore, CMV is not a serious problem in this study. 3.3.3 Hypotheses testing. We conducted the ordinary least squares regression to empirically estimate our proposed hypotheses. The results of the hypotheses tests are presented in Table 5 and Figure 2. As is shown in Table 5, Model 1 only contains the control variables. To answer the first research question (RQ1)(What is the impact of the unique attributes of a text-based chatbot on consumer responses to the text-based chatbot in the Construct M SD EMP FRI COM DIS TRU REL RES EMP 4.129 1.088 0.886 FRI 5.183 0.989 0.236 0.933 COM 2.860 1.043 0.182 0.298 0.898 DIS 4.489 0.943 0.142 0.202 0.175 0.856 TRU 4.266 1.000 0.753 0.262 0.227 0.194 0.835 Table 3. REL 4.457 1.330 0.587 0.012 0.057 0.111 0.625 0.954 Mean, SD and RES 5.034 0.996 0.302 0.097 0.012 0.106 0.402 0.337 0.831 discriminant validity Note(s): Diagonal elements (italic) are the square root of AVE for each construct. Off-diagonal elements are the evaluation based on correlations between constructs; EMP: empathy, FRI: friendliness, COM: task complexity, DIS: disclosure of the Fornell–Larcker chatbot, TRU: trust toward the chatbot, REL: reliance on the chatbot, RES: resistance to the chatbot criterion INTR Indicator EMP FRI COM DIS TRU REL RES 32,2 EMP1 0.885 0.216 0.220 0.165 0.679 0.522 0.299 EMP2 0.902 0.156 0.118 0.106 0.674 0.565 0.222 EMP3 0.871 0.256 0.145 0.106 0.648 0.473 0.281 FRI1 0.247 0.934 0.289 0.212 0.260 0.053 0.110 FRI2 0.197 0.937 0.304 0.192 0.224 0.046 0.078 FRI3 0.211 0.927 0.243 0.161 0.246 0.018 0.081 COM1 0.168 0.273 0.893 0.187 0.222 0.062 0.001 COM2 0.151 0.281 0.918 0.111 0.155 0.123 0.056 COM3 0.167 0.252 0.883 0.160 0.219 0.009 0.012 DIS1 0.009 0.169 0.132 0.710 0.059 0.034 0.116 DIS2 0.165 0.191 0.170 0.981 0.213 0.121 0.093 TRU1 0.481 0.426 0.286 0.326 0.671 0.248 0.120 TRU2 0.701 0.174 0.188 0.138 0.919 0.603 0.402 TRU3 0.677 0.156 0.147 0.102 0.893 0.628 0.412 REL1 0.549 0.050 0.043 0.069 0.612 0.946 0.334 REL2 0.560 0.024 0.066 0.117 0.591 0.957 0.337 REL3 0.572 0.006 0.056 0.133 0.586 0.960 0.294 RES1 0.264 0.286 0.163 0.110 0.317 0.421 0.792 RES2 0.283 0.045 0.006 0.106 0.367 0.323 0.918 RES3 0.203 0.077 0.134 0.048 0.315 0.097 0.776 Table 4. Note(s): All italic values are significant with p < 0.001; EMP: empathy, FRI: friendliness, COM: task Discriminant validity based on cross-loading complexity, DIS: disclosure of the chatbot, TRU: trust toward the chatbot, REL: reliance on the chatbot, RES: evaluation resistance to the chatbot e-commerce context?), we contained the main effects in Model 2 and tested the impacts of empathy and friendliness on consumers’ trust toward the text-based chatbot. The results of Model 2 indicate that empathy and friendliness are both positively related to consumers’ trust (Model 2: β 5 0.652, p < 0.001; β 5 0.093, p < 0.05). Thus, hypotheses 1 and hypothesis 2 are supported. To answer the second research question (RQ2)(How does the task complexity and chatbot identity disclosure moderate the relationship between these attributes and consumer responses?), we adopted a stepwise procedure and tested the moderation effects in H3a and H3b. Model 3 in Table 5 reports the moderation test results including the main effects and interaction effects. We mean-centered the construct scores and created the interaction terms of empathy, friendliness, task complexity and disclosure of the chatbot. After adding the interaction terms to the model, the results show that task complexity significantly moderates the relationship between friendliness and consumers’ trust toward the chatbot (Model 3: β 5 0.073, p < 0.05), thus supporting H3b. Using the same methods and procedures, we also tested for the moderation effects of disclosure of the chatbot on the relationship between chatbot attributes and consumers’ trust. The results indicate that disclosure of the chatbot significantly moderates the relationship between empathy and consumers’ trust (Model 3: β 5 0.154, p < 0.05). In addition, disclosure of the chatbot also significantly moderates the relationship between friendliness and consumers’ trust (Model 3: β 5 0.184, p < 0.01), thus supporting H4b, while H4a is not supported. To answer the third research question (RQ3)(How does consumers’ trust impact their responses to the text-based chatbot in e-commerce?), we tested the relationship between trust and consumers’ response (reliance and resistance) in Model 4, Model 5, Model 6 and Model 7. As for the relationship between consumers’ trust and reliance on the chatbot, Model 4 only contains the control variables, while the results in Model 5 indicate the positive relationship between consumers’ trust and reliance on the chatbot (Model 5: β 5 0.783, p < 0.001), thus Consumers’ response to text-based chatbots Table 5. Regression results Trust toward the chatbot Reliance on the chatbot Resistance to the chatbot Variables Model 1 Model 2 Model 3 Model 4 Model 5 Model 6 Model 7 Main effects Empathy 0.652*** (0.037) 0.527*** (0.065) Friendliness 0.093* (0.042) 0.275*** (0.07) Task complexity 0.079* (0.039) 0.084* (0.04) Disclosure of the 0.084* (0.041) 0.008 (0.048) chatbot Trust toward the 0.783*** (0.061) 0.382*** (0.054) chatbot Interaction effects Empathy * Task 0.024 (0.038) complexity Friendliness * 0.073* (0.032) Task complexity Empathy * 0.154* (0.067) Disclosure of the chatbot Friendliness * 0.184** (0.066) Disclosure of the chatbot Control variables Age 0.097 (0.065) 0.05 (0.043) 0.051 (0.042) 0.23** (0.085) 0.154* (0.069) 0.006 (0.065) 0.043 (0.06) Gender 0.032 (0.130) 0.034 (0.085) 0.021 (0.084) 0.097 (0.171) 0.072 (0.137) 0.067 (0.13) 0.055 (0.12) Experience with 0.021 (0.058) 0.007 (0.038) 0.005 (0.038) 0.077 (0.076) 0.093 (0.061) 0.023 (0.058) 0.031 (0.053) the chatbot Constant 4.204*** (0.321) 0.936* (0.384) 0.776 (0.461) 3.014*** (0.421) 0.277 (0.425) 4.971*** (0.32) 6.58*** (0.372) Observations 299 299 299 299 299 299 299 R-squared 0.009 0.584 0.605 0.03 0.376 0.009 0.137 Note(s): Standard errors are in the parentheses; *p < 0.05, **p < 0.01, ***p < 0.001 Moderators INTR 32,2 Unique attributes of Task complexity Customer outcomes text-based chatbot H3b H3a Reliance on 0.024 – 0.073* H1 the chatbot Empathy H5 2 R =37.6% Trust towards the Chatbot H6 R = 60.5% Resistance to H2 the chatbot Friendliness H4b H4a R =13.7% 0.184** – 0.154* Disclosure of the Figure 2. chatbot Theoretical model with results Note(s): *p < 0.05, **p < 0.01, ***p < 0.001 supporting H5. Similarly, for the relationship between consumers’ trust and resistance to the chatbot, Model 6 contains the control variables of the research model. Finally, the results in Model 7 show that consumers’ trust is negatively related to their resistance to the text-based chatbot (Model 7: β 5 0.382, p < 0.001), thus supporting H6. 4. Discussions and conclusions 4.1 Summary of findings Following Mehrabian and Russell’s SOR framework (1974), the current study tests a model of consumers’ trust and response to the text-based chatbot in the e-commerce setting, involving the moderating role of the different complexity of a task and the disclosure of the use of a chatbot. The specific findings of this study are summarized as follows: Considering the first research question (RQ1), this study comes to the following findings. We mainly target the impacts of two specific attributes (empathy and friendliness) on consumers’ responses in this study. Consumers’ perceived empathy and friendliness of the text-based chatbot are both positively connected to their trust toward the chatbot. Specifically, the positive effect of empathy on trust is larger than that of friendliness. Our explanation for this finding is that, compared with a friendly manner, consumers prefer the chatbot to be able to appreciate their perspective and meet their specific needs. Therefore, empathy of the customer service provider is essential for successful interactions. Considering the second research question (RQ2), this study provides empirical evidence on the boundary conditions of consumers’ perception and response. We investigate the moderating effects of task complexity and text-based chatbot disclosure on the relationship between the chatbots’ attributes and trust. The findings indicate that task complexity negatively moderates the relationship between friendliness and consumers’ trust toward the chatbot, such that when the consumers’ task is complex, the positive effect of friendliness on consumers’ trust is weaker. Conversely, task complexity does not significantly impact the relationship between empathy and consumer’ trust toward the chatbot. Our explanation for this finding is that, when the task is complex to address, consumers may be more concerned about the chatbots’ professional level and ability to solve problems, instead of their attitudes or service manner. Considering the identity disclosure of the text-based chatbot, the findings Consumers’ indicate that the disclosure of the chatbot negatively moderates the relationship between response to empathy and consumers’ trust toward the chatbot, while it positively moderates the text-based relationship between friendliness and consumers’ trust toward the chatbot. chatbots Considering the third research question (RQ3), we take both consumers’ approach response (reliance on the chatbot) and averting response (resistance to the chatbot) into consideration. Findings indicate that consumers’ trust toward the chatbot is positively related to their reliance on the chatbot and negatively related to consumers’ resistance to the chatbot. More specifically, the positive effect between consumers’ trust and reliance is greater than the negative effect between consumers’ trust and resistance. Thus, the effects of consumers’ trust on their approach response are larger than that on their averting response. 4.2 Theoretical contributions This study makes several theoretical contributions. First, this study contributes to the SOR framework by extending it to the automated chatbot in the e-commerce context. Previously, the SOR framework has mainly targeted traditional contexts (Kamboj et al., 2018; Kim et al., 2020). This extension contributes to the SOR framework by investigating specific attributes of automated text-based chatbots and how such attributes form consumers’ trust and response in the e-commerce setting. The theoretical model in this research also contributes to the information systems literature. Second, this study also considers the distinct characteristics of customer service in e-commerce by incorporating task complexity (Bjørn and Ngwenyama, 2009; Dayan and Di Benedetto, 2010) and chatbot identity disclosure (Luo et al., 2019) as the boundary conditions of consumers’ trust in this context. Prior research on consumers’ response to the automated chatbot mainly discussed several attributes of the automated bots, such as anthropomorphism (Castelo et al., 2019; Desideri et al., 2019; Yuan and Dennis, 2019), humanness (Wirtz et al., 2018) and adaptivity (Heerink et al., 2010), the boundary conditions of these effects are largely ignored. The findings of this research show that when the tasks are different and the disclosure level is different, consumers’ responses will also be different. Therefore, identifying the boundary conditions can also provide possible ways to reconcile the mixed findings in relevant research, thus opening an avenue for future study. Third, this study provides a deeper understanding of consumers’ responses to the automated AI-based chatbot, involving both approach response and averting response. As a new emerging technology, AI-based chatbots are providing the company and consumer with an expanding variety of benefits, therefore the effective application of automated bots is a big concern, especially in shaping the future of e-commerce. The extant literature on consumers’ response to the AI-based bots has examined consumers’ future intention with the automated machine (Ciechanowski et al., 2019; Fridin and Belokopytov, 2014; Heerink et al., 2010; Piçarra and Giger, 2018; Wirtz et al., 2018). Therefore, there is a call for a systematic understanding of consumers’ responses to a chatbot. This research adds to the understanding of both consumers’ reliance and resistance to a text-based chatbot. Boundary conditions of consumers’ responses are also identified, thus filling a gap in previous research that only focused on consumers’ positive responses to the automated bots. 4.3 Practical implications Our study also provides several practical implications for both customer service providers and the technology developers. First, as the AI-based chatbot is quite ubiquitous in the e-commerce setting, this research contributes to the limited understanding of consumers’ response to the chatbot in the business context. Although the automated bots can arouse users’ interest during the interactions, the online store should also be aware of consumers’ intention to resist. The findings of this study can arouse the attention of the online stores that INTR want to utilize chatbots more effectively than the human alternative but also more effective 32,2 than their competitors. Second, the findings of this study provide suggestions for the technology developer that creates and customizes the chatbot. This study provides empirical evidence for the effects of task complexity and identity disclosure. Specifically, consumers’ perceptions of the chatbot will be different when their tasks are different. Interestingly, when the task is complex, the positive effects of chatbots’ friendliness on consumers’ trust will be weaker. Therefore, we suggest that it is more important for the chatbot to increase their professional level and address consumers’ problems instead of merely replying to the consumers with standard answers in a friendly manner. Moreover, the disclosure of the machine identity should also be considered. Consumers’ perceived disclosure of the automated chatbot may have different effects on the relationship between empathy/friendliness and trust. Therefore, the technology developer that customizes the chatbot should balance its characteristics to cater to consumers’ different needs. Given the current capabilities of AI, this will require additional time and resources to further train and customize the chatbot, butitisworth it. 4.4 Limitations and future research There are also limitations in this research. First, apart from empathy and friendliness, there are also other attributes of the chatbot that will affect consumers’ responses, such as the intelligence level of the chatbot and passing the “Turing test” (Turing and Haugeland, 1950; Castelo et al., 2019; Desideri et al., 2019; Heerink et al., 2010; Wirtz et al., 2018; Yuan and Dennis, 2019). More specific characteristics of the chatbots will be taken into consideration in future research. Second, as there is a distinction between humans and machines, especially in the customer service institution, future researchshouldalsoextendthe understanding of the difference between human and machine customer service providers. Third, this study only investigated the chatbot in the e-commerce setting. While chatbots developed for the other domains (e.g. healthcare, sharing economy) may behave differently when compared to chatbots developed for the e-commerce setting. Specific issues need to be detected in other domains in future studies. Moreover, the characteristics of the customers may also have an impact on their acceptance or perception of the chatbot during purchasing. Therefore, attributes of customers should be taken into consideration in future research. References Adam, M., Wessel, M. and Benlian, A. (2020), “AI-based chatbots in customer service and their effects on user compliance”, Electronic Markets, pp. 1-19, doi: 10.1007/s12525-020-00414-7. Araujo, T. (2018), “Living up to the chatbot hype: the influence of anthropomorphic design cues and communicative agency framing on conversational agent and company perceptions”, Computers in Human Behavior, Vol. 85, pp. 183-189. Birnbaum, G.E., Mizrahi, M., Hoffman, G., Reis, H.T., Finkel, E.J. and Sass, O. (2016), “What robots can teach us about intimacy: the reassuring effects of robot responsiveness to human disclosure”, Computers in Human Behavior, Vol. 63, pp. 416-423. Bitner, M.J. (1992), “Servicescapes: the impact of physical surroundings on customers and employees”, Journal of Marketing, Vol. 56 No. 2, pp. 57-71. Bjørn, P. and Ngwenyama, O. (2009), “Virtual team collaboration: building shared meaning, resolving breakdowns and creating translucence”, Information Systems Journal, Vol. 19 No. 3, pp. 227-253. € Blut, M., Wang, C., Wunderlich, N.V. and Brock, C. (2021), “Understanding anthropomorphism in Consumers’ service provision: a meta-analysis of physical robots, chatbots, and other AI”, Journal of the response to Academy of Marketing Science, pp. 1-27, doi: 10.1007/s11747-020-00762-y. text-based Bolton, R.N. (1998), “A dynamic model of the duration of the customer’s relationship with a continuous chatbots service provider: the role of satisfaction”, Marketing Science, Vol. 17 No. 1, pp. 45-65. Breuer, C., Huffmeier, € J. and Hertel, G. (2016), “Does trust matter more in virtual teams? A meta- analysis of trust and team effectiveness considering virtuality and documentation as moderators”, Journal of Applied Psychology, Vol. 101 No. 8, pp. 1151-1177. Castelo, N., Bos, M.W. and Lehmann, D.R. (2019), “Task-dependent algorithm aversion”, Journal of Marketing Research, Vol. 56 No. 5, pp. 809-825. Chang, H.J., Eckman, M. and Yan, R.N. (2011), “Application of the stimulus-organism-response model to the retail environment: the role of hedonic motivation in impulse buying behavior”, International Review of Retail, Distribution and Consumer Research, Vol. 21 No. 3, pp. 233-249. Chen, C.C., Lin, S.Y., Cheng, C.H. and Tsai, C.C. (2012), “Service quality and corporate social responsibility, influence on post-purchase intentions of sheltered employment institutions”, Research in Developmental Disabilities, Vol. 33 No. 6, pp. 1832-1840. Cheng, X., Fu, S. and de Vreede, G.J. (2018), “A mixed method investigation of sharing economy driven car-hailing services: online and offline perspectives”, International Journal of Information Management, Vol. 41 No. 10, pp. 57-64. Cheng, X., Bao, Y. and Zarifis, A. (2020), “Investigating the impact of IT-mediated information interruption on emotional exhaustion in the workplace”, Information Processing and Management, Vol. 57 No. 6, p. 102281. Ciechanowski, L., Przegalinska, A., Magnuski, M. and Gloor, P. (2019), “In the shades of the uncanny valley: an experimental study of human–chatbot interaction”, Future Generation Computer Systems, Vol. 92, pp. 539-548. Cram, W.A., D’arcy, J. and Proudfoot, J.G. (2019), “Seeing the forest and the trees: a meta-analysis of the antecedents to information security policy compliance”, MIS Quarterly, Vol. 43 No. 2, pp. 525-554. Cyr, D., Head, M., Larios, H. and Pan, B. (2009), “Exploring human images in website design: a multi- method approach”, MIS Quarterly, Vol. 33 No. 3, pp. 539-566. Czaplewski, A.J., Olson, E.M. and Slater, S.F. (2002), “Applying the RATER model for service success”, Marketing Management, Vol. 11 No. 1, pp. 14-17. Daugherty, P.R., James Wilson, H. and Michelman, P. (2019), “Revisiting the jobs artificial intelligence will create”, MIT Sloan Management Review, Vol. 60 No. 4, pp. 1-8. Davenport, T.H. (2019), “Can we solve AI’s ’trust problem’?”, MIT Sloan Management Review, Vol. 60 No. 2, pp. 18-19. Dayan, M. and Di Benedetto, C.A. (2010), “The impact of structural and contextual factors on trust formation in product development teams”, Industrial Marketing Management, Vol. 39 No. 4, pp. 691-703. de Kervenoael, R., Hasan, R., Schwob, A. and Goh, E. (2020), “Leveraging human-robot interaction in hospitality services: incorporating the role of perceived value, empathy, and information sharing into visitors’ intentions to use social robots”, Tourism Management,Vol. 78, p. 104042. Desideri, L., Ottaviani, C., Malavasi, M., di Marzio, R. and Bonifacci, P. (2019), “Emotional processes in human-robot interaction during brief cognitive testing”, Computers in Human Behavior, Vol. 90, pp. 331-342. Donovan, R. and Rossiter, J. (1982), “Store atmosphere: an environmental psychology approach”, Journal of Retailing, Vol. 58 No. 1, pp. 34-57. Duan, Y., Edwards, J.S. and Dwivedi, Y.K. (2019), “Artificial intelligence for decision making in the era INTR of Big Data – evolution, challenges and research agenda”, International Journal of Information 32,2 Management, Vol. 48, pp. 63-71. Edwards, J.C., Rust, K.G., McKinley, W. and Moon, G. (2003), “Business ideologies and perceived breach of contract during downsizing: the role of the ideology of employee self-reliance”, Journal of Organizational Behavior, Vol. 24 No. 1, pp. 1-23. Eroglu, S.A., Machleit, K.A. and Davis, L.M. (2001), “Atmospheric qualities of online retailing: a conceptual model and implications”, Journal of Business Research, Vol. 54 No. 2, pp. 177-184. Eroglu, S.A., Machleit, K.A. and Davis, L.M. (2003), “Empirical testing of a model of online store atmospherics and shopper responses”, Psychology and Marketing, Vol. 20 No. 2, pp. 139-150. Ert, E., Fleischer, A. and Magen, N. (2016), “Trust and reputation in the sharing economy: the role of personal photos in Airbnb”, Tourism Management, Vol. 55, pp. 62-73. Everard, A. and Galletta, D.F. (2005), “How presentation flaws affect perceived site quality, trust, and intention to purchase from an online store”, Journal of Management Information Systems, Vol. 22 No. 3, pp. 56-95. Fiore, A.M. and Kim, J. (2007), “An integrative framework capturing experiential and utilitarian shopping experience”, International Journal of Retail and Distribution Management, Vol. 35 No. 6, pp. 421-442. Fridin, M. and Belokopytov, M. (2014), “Acceptance of socially assistive humanoid robot by preschool and elementary school teachers”, Computers in Human Behavior, Vol. 33, pp. 23-31. Froehlich, A. (2018), “Pros and cons of chatbots in the IT helpdesk”, available at: https://www. informationweek.com/strategic-cio/it-strategy/pros-and-cons -of-chatbots-in-the-it-helpdesk/a/d- id/1332942 (accessed 10 February 2018). Gilbert, R.L. and Forney, A. (2015), “Can avatars pass the Turing test? Intelligent agent perception in a 3D virtual environment”, International Journal of Human-Computer Studies, Vol. 73, pp. 30-36. Go, E. and Sundar, S.S. (2019), “Humanizing chatbots: the effects of visual, identity and conversational cues on humanness perceptions”, Computers in Human Behavior, Vol. 97, pp. 304-316. Goodhue, D.L. and Thompson, R.L. (1995), “Task-technology fit and individual performance”, MIS Quarterly, Vol. 19 No. 2, pp. 213-236. Griffith, D.A. and Lusch, R.F. (2007), “Getting marketers to invest in firm-specific capital”, Journal of Marketing, Vol. 71 No. 1, pp. 129-145. Heerink, M., Kr€ose, B., Evers, V. and Wielinga, B. (2010), “Assessing acceptance of assistive social agent technology by older adults: the Almere model”, International Journal of Social Robotics, Vol. 2 No. 4, pp. 361-375. Hoff, K.A. and Bashir, M. (2015), “Trust in automation: integrating empirical evidence on factors that influence trust”, Human Factors, Vol. 57 No. 3, pp. 407-434. Hoy, W.K. and Tschannen-Moran, M. (1999), “Five faces of trust: an empirical confirmation in urban elementary schools”, Journal of School Leadership, Vol. 9 No. 3, pp. 184-208. Hua, Y., Cheng, X., Hou, T. and Luo, R. (2020), “Monetary rewards, intrinsic motivators, and work engagement in the IT-enabled sharing economy: a mixed-methods investigation of internet taxi drivers”, Decision Sciences, Vol. 51 No. 3, pp. 755-785. Ischen, C., Araujo, T., Voorveld, H., van Noort, G. and Smit, E. (2019), “Privacy concerns in chatbot interactions”, in Følstad, A., Araujo, T., Papadopoulos, S., Law, E.L.-C., Granmo, O.-C., Luger, E. and Brandtzaeg, P.B. (Eds), Conversations 2019: Chatbot Research and Design, Springer, Cham, pp. 34-48. Jacoby, J. (2002), “Stimulus-organism-response reconsidered: an evolutionary step in modeling (consumer) behavior”, Journal of Consumer Psychology, Vol. 12 No. 1, pp. 51-57. Jarvenpaa, S.L., Knoll, K. and Leidner, D.E. (1997), “Is anybody out there? Antecedents of trust in global virtual teams”, Journal of Management Information Systems,Vol.14No. 4, pp. 29-64. Jensen, M.L. and Yetgin, E. (2017), “Prominence and interpretation of online conflict of interest Consumers’ disclosures”, MIS Quarterly, Vol. 41 No. 2, pp. 629-643. response to Kamboj, S., Sarmah, B., Gupta, S. and Dwivedi, Y. (2018), “Examining branding co-creation in brand text-based communities on social media: applying the paradigm of Stimulus-Organism-Response”, chatbots International Journal of Information Management, Vol. 39, pp. 169-185. Kaplan, A. and Haenlein, M. (2019), “Siri, Siri, in my hand: who’s the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence”, Business Horizons, Vol. 62 No. 1, pp. 15-25. Kelley, S.W. and Davis, M.A. (1994), “Antecedents to customer expectations for service recovery”, Journal of the Academy of Marketing Science, Vol. 22 No. 1, pp. 52-61. Kim, H.W. and Kankanhalli, A. (2009), “Investigating user resistance to information systems implementation: a status quo bias perspective”, MIS Quarterly, Vol. 33 No. 3, pp. 567-582. Kim, M.J., Lee, C.K. and Jung, T. (2020), “Exploring consumer behavior in virtual reality tourism using an extended stimulus-organism-response model”, Journal of Travel Research, Vol. 59 No. 1, pp. 69-89. Krishnan, R., Martin, X. and Noorderhaven, N.G. (2006), “When does trust matter to alliance performance?”, Academy of Management Journal, Vol. 49 No. 5, pp. 894-917. Lapointe, L. and Rivard, S. (2005), “A multilevel model of resistance to information technology implementation”, MIS Quarterly, Vol. 29 No. 3, pp. 461-491. Lemon, K.N. and Verhoef, P.C. (2016), “Understanding customer experience throughout the customer journey”, Journal of Marketing, Vol. 80 No. 6, pp. 69-96. Liang, H., Xue, Y., Pinsonneault, A. and Wu, Y. (2019), “What users do besides problem-focused coping when facing it security threats: an emotion-focused coping perspective”, MIS Quarterly, Vol. 43 No. 2, pp. 373-394. Lin, J., Luo, Z., Cheng, X. and Li, L. (2019), “Understanding the interplay of social commerce affordances and swift guanxi: an empirical study”, Information and Management, Vol. 56 No. 2, pp. 213-224. Lindell, M.K. and Brandt, C.J. (2000), “Climate quality and climate consensus as mediators of the relationship between organizational antecedents and outcomes”, Journal of Applied Psychology, Vol. 85 No. 3, pp. 331-348. Longoni, C., Bonezzi, A. and Morewedge, C.K. (2019), “Resistance to medical artificial intelligence”, Journal of Consumer Research, Vol. 46 No. 4, pp. 629-650. Luo, X., Tong, S., Fang, Z. and Qu, Z. (2019), “Frontiers: machines vs humans: the impact of artificial intelligence chatbot disclosure on customer purchases”, Marketing Science, Vol. 38 No. 6, pp. 937-947. Mayer, R.C., Davis, J.H. and Schoorman, F.D. (1995), “An integrative model of organizational trust”, Academy of Management Review, Vol. 20 No. 3, pp. 709-734. Mehrabian, A. and Russell, J.A. (1974), An Approach to Environmental Psychology, The MIT Press, Cambridge, MA. Mende, M., Scott, M.L., van Doorn, J., Grewal, D. and Shanks, I. (2019), “Service robots rising: how humanoid robots influence service experiences and elicit compensatory consumer responses”, Journal of Marketing Research, Vol. 56 No. 4, pp. 535-556. Mitchell, V. (2018), “Gartner: why humans will still be at the core of great CX”, available at: https://www. cmo.com.au/article/642649/gartner-why-humans-will-still-core-great-cx/ (accessed 20 June 2018). Mollen, A. and Wilson, H. (2010), “Engagement, telepresence and interactivity in online consumer experience: reconciling scholastic and managerial perspectives”, Journal of Business Research, Vol. 63 Nos 9–10, pp. 919-925. Morgeson, F.P. and Humphrey, S.E. (2006), “The Work Design Questionnaire (WDQ): developing and validating a comprehensive measure for assessing job design and the nature of work”, Journal of Applied Psychology, Vol. 91 No. 6, pp. 1321-1339. Murray, J., Elms, J. and Curran, M. (2019), “Examining empathy and responsiveness in a high-service INTR context”, International Journal of Retail and Distribution Management, Vol. 47 No. 12, 32,2 pp. 1364-1378. Nass, C., Steuer, J. and Tauber, E.R. (1994), “Computers are social actors”, Proceedings of the SIGCHI Conference on Human Factors in Computing Systems: Celebrating Interdependence, ACM, Boston, MA, pp. 72-78. € € Onkal, D., Goodwin, P., Thomson, M., Gonul, S. and Pollock, A. (2009), “The relative influence of advice from human experts and statistical methods on forecast adjustments”, Journal of Behavioral Decision Making, Vol. 22 No. 4, pp. 390-409. Pise, R. (2018), “Chatbot market size is set to exceed USD 134 billion by 2024 - ClickZ”, available at: https://www.clickz.com/chatbot-market-size-is-set-to-exceed-usd-1-34-billion-by-2024/215518/ (accessed 6 July 2018). Piçarra, N. and Giger, J.C. (2018), “Predicting intention to work with social robots at anticipation stage: assessing the role of behavioral desire and anticipated emotions”, Computers in Human Behavior, Vol. 86, pp. 129-146. Podsakoff, P.M., MacKenzie, S.B., Lee, J.Y. and Podsakoff, N.P. (2003), “Common method biases in behavioral research: a critical review of the literature and recommended remedies”, Journal of Applied Psychology, Vol. 88 No. 5, pp. 879-903. Powell, P.A. and Roberts, J. (2017), “Situational determinants of cognitive, affective, and compassionate empathy in naturalistic digital interactions”, Computers in Human Behavior, Vol. 68, pp. 137-148. Pugh, S.D. (2001), “Service with a smile: emotional contagion in the service encounter”, Academy of Management Journal, Vol. 44 No. 5, pp. 1018-1027. Rose, S., Clark, M., Samouel, P. and Hair, N. (2012), “Online customer experience in e-retailing: an empirical model of antecedents and outcomes”, Journal of Retailing, Vol. 88 No. 2, pp. 308-322. Saldanha, T.J.V., Mithas, S. and Krishnan, M.S. (2017), “Leveraging customer involvement for fueling innovation: the role of relational and analytical information processing capabilities”, MIS Quarterly, Vol. 41 No. 1, pp. 267-286. Seeber, I., Bittner, E., Briggs, R.O., de Vreede, T., de Vreede, G.J., Elkins, A., Maier, R., Merz, A.B., Oeste-Reiß, S., Randrup, N., Schwabe, G. and S€ollner, M. (2020), “Machines as teammates: a research agenda on AI in team collaboration”, Information and Management, Vol. 57 No. 2, p. 103174. Shawar, B.A. and Atwell, E.S. (2005), “Using corpora in machine-learning chatbot systems”, International Journal of Corpus Linguistics, Vol. 10 No. 4, pp. 489-516. Shiau, W.L., Sarstedt, M. and Hair, J.F. (2019), “Internet research using partial least squares structural equation modeling (PLS-SEM)”, Internet Research, Vol. 29 No. 3, pp. 398-406. Starzyk, K.B., Holden, R.R., Fabrigar, L.R. and MacDonald, T.K. (2006), “The personal acquaintance measure: a tool for appraising one’s acquaintance with any person”, Journal of Personality and Social Psychology, Vol. 90 No. 5, pp. 833-847. Thatcher, J.B., Wright, R.T., Sun, H., Zagenczyk, T.J. and Klein, R. (2018), “Mindfulness in information technology use: definitions, distinctions, and a new measure”, MIS Quarterly, Vol. 42 No. 3, pp. 831-847. Thompson, C. (2018), “May AI help you?”, available at: https://www.nytimes.com/interactive/2018/11/ 14/magazine/ tech-design-ai-chatbot.html (accessed 14 November 2018). Tsai, W.C. and Huang, Y.M. (2002), “Mechanisms linking employee affective delivery and customer behavioral intentions”, Journal of Applied Psychology, Vol. 87 No. 5, pp. 1001-1008. Turing, A. and Haugeland, J. (1950), “Mind design II: philosophy, psychology, and artificial intelligence”, in Haugeland, J. (Ed.), Computing Machinery and Intelligence, MIT Press, Cambridge, MA, pp. 29-56. van Doorn, J., Lemon, K.N., Mittal, V., Nass, S., Pick, D., Pirner, P. and Verhoef, P.C. (2010), “Customer Consumers’ engagement behavior: theoretical foundations and research directions”, Journal of Service response to Research, Vol. 13 No. 3, pp. 253-266. text-based Wirtz, J., Patterson, P.G., Kunz, W.H., Gruber, T., Lu, V.N., Paluch, S. and Martins, A. (2018), “Brave chatbots new world: service robots in the frontline”, Journal of Service Management, Vol. 29 No. 5, pp. 907-931. Wood, S.L. (2001), “Remote purchase environments: the influence of return policy leniency on two- stage decision processes”, Journal of Marketing Research, Vol. 38 No. 2, pp. 157-169. Wu, Y.L. and Li, E.Y. (2018), “Marketing mix, customer value, and customer loyalty in social commerce: a stimulus-organism-response perspective”, Internet Research, Vol. 28 No. 1, pp. 74-104. Yuan, L. and Dennis, A.R. (2019), “Acting like humans? Anthropomorphism and consumer’s willingness to pay in electronic commerce”, Journal of Management Information Systems, Vol. 36 No. 2, pp. 450-477. Zhang, M., Jin, B., Wang, G.A., Goh, T.N. and He, Z. (2016), “A study of key success factors of service enterprises in China”, Journal of Business Ethics, Vol. 134 No. 1, pp. 1-14. Złotowski, J., Yogeeswaran, K. and Bartneck, C. (2017), “Can we control it? Autonomous robots threaten human identity, uniqueness, safety, and resources”, International Journal of Human Computer Studies, Vol. 100, pp. 48-54. Appendix Constructs and measurement items References Empathy (1) Text-based chatbots in e-commerce usually understand Cheng et al. (2018), Czaplewski et al. (2002), the specific needs of the customers de Kervenoael et al. (2020) (2) Text-based chatbots in e-commerce usually give customers individual attention (3) Text-based chatbots in e-commerce are available whenever it’s convenient for customers (4) If a customer requires help, the text-based chatbots would do their best to help (dropped) Friendliness (1) The text-based chatbot had a kind service during our Tsai and Huang (2002) interaction (2) The text-based chatbot provides the service in a friendly manner (3) The text-based chatbot treats me nicely Task complexity (1) The task I communicate with the customer service are Morgeson and Humphrey (2006) simple and uncomplicated (2) The task I communicate with the customer service comprises relatively uncomplicated tasks (3) The task I communicate with the customer service involves performing relatively simple tasks Table A1. Construct measurements (continued) INTR Constructs and measurement items References 32,2 Disclosure of the chatbot (1) The identity of the text-based chatbot is disclosed before Jensen and Yetgin (2017), Luo et al. (2019), conversation Starzyk et al. (2006) (2) The e-commerce platform hides the identity of text-based chatbot from me 516 (3) I am informed with the identity of the text-based chatbot before conversation (dropped) Trust toward the chatbot (1) The text-based chatbot is honest and truthful Cyr et al. (2009), Everard and Galletta (2005) (2) The text-based chatbot is capable of addressing my issues (3) The text-based chatbot’s behavior and response can meet my expectations (4) I trust the suggestions and decisions provided by the text- based chatbot (dropped) Reliance on the chatbot (1) I depend on the text-based chatbot for decision making Edwards et al. (2003) (2) I depend on the text-based chatbot for service in e- commerce (3) I depend on the text-based chatbot for suggestions in purchasing process Resistance to the chatbot (1) I prefer to use the human customer service rather than use Kim and Kankanhalli (2009), Lapointe and the text-based chatbot Rivard (2005) (2) I will try to avoid using the service provided by text-based chatbot in e-commerce (3) I do not agree with the suggestions provided by the text- based chatbot in e-commerce Table A1. Note(s): Items with (dropped) were removed in the data analysis About the authors Xusen Cheng is a professor of Information Systems in the School of Information, Renmin University of China, Beijing, China. He has obtained his PhD degree in Manchester Business School in the University of Manchester in the UK. His research focuses on information system and management, e-commerce and sharing economy, AI and behavior. His research paper has been accepted/appeared in journals such as MIS Quarterly, Journal of Management Information Systems, European Journal of Information Systems, Decision Sciences, Tourism Management, Information & Management, Information Processing and Management, Group Decision and Negotiation, Information Technology and People, British Journal of Educational Technology, Computers in Human Behavior, Electronic Commerce Research, Internet Research, among others. He serves as senior editor for Information Technology and People, associate editor for Information Technology for Development, Editorial Board for Electronic Commerce Research and Editorial Review Board for Internet Research, guest editor for several journals such as British Journal of Educational Technology, Electronic Markets, Journal of Global Information Technology Management and serves as track cochair and minitrack cochair for many conferences. His papers have also been presented in the leading conferences such as International Conference of Information System (ICIS), Hawaii International Conference of System Science (HICSS) and American Conference of Information Systems (AMCIS). Ying Bao is a PhD candidate in the School of Information Technology and Management in the University of International Business and Economics, Beijing, China. Her interests focus on user behaviors and trust in the sharing economy and corporate social responsibility. Her research papers have appeared in Information Processing and Management, Internet Research, Group Decision and Negotiation. Her papers have also been presented in the leading conferences such as Hawaii International Conference of System Science (HICSS) and American Conference on Information Consumers’ System (AMCIS). response to Alex Zarifis is a research associate in the School of Business and Economics at the University of text-based Loughborough, UK. His first degree was a BSc in Management with Information Systems from the University of Leeds, followed by an MSc in Business Information Technology and a PhD in Business chatbots Administration both from the University of Manchester. His research focuses on the user’s perspective in the areas of e-commerce, trust, privacy, artificial intelligence, blockchain, fintech, leadership, online collaboration and online education. His research has featured in journals such as Information Processing and Management, Computers in Human Behavior, Information Technology and People, International Journal of Electronic Business and Electronic Commerce Research. He served as the guest editor of British Journal of Education Technology and Electronic Markets. His research has been presented at conferences such as the European Conference on Information Systems (ECIS) and Hawaii International Conference on System Sciences (HICSS). Wankun Gong is a postgraduate student in College of Humanities and Law in Beijing University of Chemical Technology, China. His research interests include e-government, e-service and e-commerce, human behavior and management, public and personal behavior analysis, trust and technology-enabled innovation. Jian Mou is an associate professor of School of Business in Pusan National University, South Korea. His research interests include e-commerce, human–computer interaction, trust and risk issues in e-service and the dynamic nature in systems use. His research has been published in the Journal of the Association for Information Systems, Internet Research, Information and Management, International Journal of Information Management, Computers in Human Behavior, Electronic Commerce Research, Information Processing and Management, Information Development, Behaviour and Information Technology, Journal of Retailing and Consumer Services, Electronic Commerce Research and Applications, Industrial Management & Data Systems, Journal of Global Information Management, IT & People, International Journal of Human–Computer Interaction and Information Technology and Tourism, as well as in the proceedings such as ICIS, ECIS, PACIS and AMCIS. He serves as senior editor for Information Technology and People, associate editor for Internet Research, and Behaviour and Information Technology, Editorial Board for Electronic Commerce Research and guest editor for several journals such as Information Technology for Development, Electronic Markets and Journal of Global Information Technology Management. Jian Mou is the corresponding author and can be contacted at: jian.mou@pusan.ac.kr For instructions on how to order reprints of this article, please visit our website: www.emeraldgrouppublishing.com/licensing/reprints.htm Or contact us for further details: permissions@emeraldinsight.com http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Internet Research Emerald Publishing

Exploring consumers' response to text-based chatbots in e-commerce: the moderating role of task complexity and chatbot disclosure

Internet Research , Volume 32 (2): 22 – Mar 15, 2022

Loading next page...
 
/lp/emerald-publishing/exploring-consumers-response-to-text-based-chatbots-in-e-commerce-the-4O9Qrsplha

References (94)

Publisher
Emerald Publishing
Copyright
© Xusen Cheng, Ying Bao, Alex Zarifis, Wankun Gong and Jian Mou
ISSN
1066-2243
DOI
10.1108/intr-08-2020-0460
Publisher site
See Article on Publisher Site

Abstract

Purpose – Artificial intelligence (AI)-based chatbots have brought unprecedented business potential. This study aims to explore consumers’ trust and response to a text-based chatbot in e-commerce, involving the moderating effects of task complexity and chatbot identity disclosure. Design/methodology/approach – A survey method with 299 useable responses was conducted in this research. This study adopted the ordinary least squares regression to test the hypotheses. Findings – First, the consumers’ perception of both the empathy and friendliness of the chatbot positively impacts their trust in it. Second, task complexity negatively moderates the relationship between friendliness and consumers’ trust. Third, disclosure of the text-based chatbot negatively moderates the relationship between empathy and consumers’ trust, while it positively moderates the relationship between friendliness and consumers’ trust. Fourth, consumers’ trust in the chatbot increases their reliance on the chatbot and decreases their resistance to the chatbot in future interactions. Research limitations/implications – Adopting the stimulus–organism–response (SOR) framework, this study provides important insights on consumers’ perception and response to the text-based chatbot. The findings of this research also make suggestions that can increase consumers’ positive responses to text-based chatbots. Originality/value – Extant studies have investigated the effects of automated bots’ attributes on consumers’ perceptions. However, the boundary conditions of these effects are largely ignored. This research is one of the first attempts to provide a deep understanding of consumers’ responses to a chatbot. Keywords Text-based chatbot, Trust, Consumers’ response, Task complexity, Identity disclosure Paper type Research paper © Xusen Cheng, Ying Bao, Alex Zarifis, Wankun Gong and Jian Mou. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) licence. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and Internet Research non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this licence may be seen at http://creativecommons.org/licences/by/4.0/legalcode pp. 496-517 Emerald Publishing Limited This work is supported by the Fundamental Research Funds for the Central Universities, and the 1066-2243 DOI 10.1108/INTR-08-2020-0460 Research Funds of Renmin University of China (Grant No.21XNO002). 1. Introduction Consumers’ Disembodied conversational agents (DCAs) have now become quite common (Araujo, 2018; response to Luo et al., 2019). Galvanized by artificial intelligence (AI) and machine learning, DCAs in the text-based form of chatbots can simulate human interactions through text chats or voice commands in chatbots various fields and have brought unprecedented business potential (Luo et al., 2019). A chatbot is defined as “a machine conversation system which interacts with human users via natural conversational language” (Shawar and Atwell, 2005, p. 489). Specifically, the market size of chatbots is expanding from $250m in 2017 to over $1.34bn by 2024 (Pise, 2018). Many famous brands and major platforms, such as eBay, Facebook, WeChat, Amazon and Apple’s Siri, have rolled out chatbots to take orders, recommend products or provide other customer services (Thompson, 2018; Go and Sundar, 2019). Gartner predicts that by 2021, almost 15% of the customer services will be completely handled by AI (Mitchell, 2018). AI-enabled chatbots have provided several unique business benefits in the e-commerce setting. First, the chatbots are able to understand consumers’ requests and automate the customer services, addressing their requests effectively (Daugherty et al., 2019). Moreover, compared to a human, the AI-enabled chatbot will not experience negative emotions or work exhaustion and can always interact with consumers in a friendly manner (Luo et al., 2019). Additionally, the chatbots can easily handle a large number of customer communications at the same time and increase the customer-service efficiency. While the adoption of AI in e-commerce can bring several business benefits for the companies or platforms from the perspective of the supply side, a key challenge for the application of chatbots is the customers’ perception from the perspective of the demand side (Froehlich, 2018). Although the chatbots can interact with consumers in a more friendly way compared with the customer service of employees, attributes of the technology have created customer pushback. For example, consumers may find it uncomfortable talking with a technology-based chatbot for their specific personal needs or purchasing decisions. Moreover, consumers may have a prejudice that the computer programs lack empathy and personal feelings, perceiving the chatbots as less trustworthy than humans. This negative perception of AI-enabled chatbots will even have negative effects on their perception of the brands or companies. Therefore, firms should apply the chatbot wisely to provide better customer service. For instance, firms can increase effectiveness by introducing chatbots to complete easy customer requests, but the technology might not be applied for complex tasks when customers have very demanding requests, which need employees to intervene (e.g. service recovery, price negotiation). As a result, there is a call for a deeper understanding of consumers’ perceptions and responses to a text-based chatbot in the e-commerce setting, involving the effects of the chatbots’ specific attributes. Extensive research exists on AI-enabled systems in business practice, including social robots (de Kervenoael et al., 2020), machine “teammates” (Seeber et al., 2020) and chatbots (Luo et al., 2019). Findings indicate that attributes of the AI-enabled computer programs, such as anthropomorphism (Castelo et al., 2019; Desideri et al., 2019; Yuan and Dennis, 2019), humanness (Wirtz et al., 2018), empathy (de Kervenoael et al., 2020) and adaptivity (Heerink et al., 2010) will significantly impact consumers’ perception, decision-making (Duan et al., 2019), purchase intention (Mende et al., 2019) and use intention (Heerink et al., 2010; de Kervenoael et al., 2020). Although existing studies have drawn attention to consumers’ positive perception of the application of text-based chatbots in the e-commerce setting, current empirical evidence on consumers’ pushback is still limited. Moreover, consumers’ perceptions or responses may differ in different conditions. While the boundary conditions of consumers’ perception and response behavior have been largely ignored in existing studies. Against this backdrop, we address both the practical and theoretical aspects with the following research questions: RQ1. What is the impact of the unique attributes of a text-based chatbot on consumer INTR responses to the text-based chatbot in the e-commerce context? 32,2 RQ2. How does the task complexity and chatbot identity disclosure moderate the relationship between these attributes and consumer responses? RQ3. How does consumers’ trust impact their responses to the text-based chatbot in e- commerce? In order to answer these research questions, we draw upon the stimulus–organism–response (SOR) framework (Mehrabian and Russell, 1974) and test a model of consumers’ trust and subsequent responses to a text-based chatbot in the e-commerce setting. We also integrate the moderating role of task complexity and consumers’ perception of the disclosure of the chatbot to test the boundary conditions of different consumer responses. The SOR framework provides an ideal theoretical framework to examine the links between stimulus, organism and response systematically (Kim et al., 2020). Therefore, it is suitable to address our research questions. This paper is structured as follows. First, we introduce the theoretical background and present the hypotheses development. Next, we present the research methodology and data analysis results. Finally, we conclude with a brief summary of the findings, the theoretical contributions and the practical implications. Limitations and future research are also provided. 2. Theoretical background and hypotheses development 2.1 Stimulus–organism–response model The “stimulus-organism-response framework” was initially proposed by Mehrabian and Russell (1974) and then modified by Jacoby (2002). The SOR framework suggests that individuals will react to the environmental factors and have certain behaviors (Kamboj et al., 2018). Specifically, individuals’ reaction behaviors consist of two typical forms: approach behaviors with positive actions, such as the preference to affiliate, explore and averting behaviors with the opposite actions, such as the desire to react negatively (Kim et al., 2020). Existing studies have extended the SOR framework to many areas, including the online shopping experience (Eroglu et al.,2003), website experience (Mollen and Wilson, 2010), consumer behavior in the retail industry (Rose et al., 2012) and tourism industry (Kim et al.,2020). Among the abovementioned contexts, e- commerce is one of the most representative. The SOR theory has been adopted in the e-commerce context as a psychology theory to investigate consumer behavior (Fiore and Kim, 2007; Chang et al., 2011; Wu and Li, 2018). For example, Eroglu et al. (2003) argued that the environmental or atmospheric cues of the online platform (S-stimuli) can affect consumers’ cognition or emotion (O-organism), which will then influence their behavioral outcomes (R-response). Similarly, in this research, we adopted the SOR framework to explain consumers’ organism process and behavior outcomes toward the text-based chatbot in the e-commerce setting. The SOR framework consists of three elements: stimulus, organism and response (Kamboj et al., 2018). Similar to the “input-response-output framework”, the “stimulus,”“organism” and “response” correspond to the “input,”“response” and “output.” More specifically, “stimulus” is defined as the environmental factors encountered by individuals (Jacoby, 2002), which will arouse them (Eroglu et al., 2001). In this research setting, “stimulus” is defined as the unique attributes of the text-based chatbot that will arouse the consumers. The second element “organism” can be defined as individuals’ affective and cognitive condition and their process of subconsciously intervening between the stimuli and responses (Jacoby, 2002). For example, in the commercial setting, the “organism” includes the individuals’ cognitive and affective perceptions of their experience as a consumer (Rose et al., 2012). The third element “response” has been described as individuals’ behavioral consequences or outcomes to the environment (Donovan and Rossiter, 1982). As aforementioned, individuals’ response to the environment includes both positive actions and negative actions. For example, in the Consumers’ commercial setting, a positive response may include consumers’ willingness to stay, explore response to or affiliate (Bitner, 1992), while a negative response may refer to individuals’ willingness to text-based react negatively. In this research setting, we take both consumers’ positive and negative chatbots reactions toward the text-based chatbot into consideration. 2.2 Stimulus: unique attributes of a text-based chatbot Turing proposed the Turing test to determine AI or machine intelligence in 1950 (Turing and Haugeland, 1950). A machine that passed the Turing test can be identified as more realistic and human-like, which makes it more compelling for consumers (Gilbert and Forney, 2015). Thus, the realistic and human-like characteristics of the chatbots will have a great impact on the usability of the chatbot. In the e-commerce service, AI has been applied in the form of text-based chatbots that can respond to consumers’ requests automatically (Kaplan and Haenlein, 2019). The AI-enabled text-based chatbot has been widely used in the online purchasing process. On the one hand, the service provided by a text-based chatbot can be more customer-friendly when compared with some traditional human customer service. According to social response theory (Nass et al., 1994) and anthropomorphism theory (Blut et al., 2021), customers will perceive higher satisfaction during the interaction with human-like machines. While on the other hand, some consumers perceive that these chatbots always give rigid or standard answers and are not empathetic enough. As a result, we mainly focus on the effects of these two attributes (empathy and friendliness) of a text- based chatbot on consumers’ outcomes. 2.2.1 Empathy of the chatbot. Empathy can be defined as “ability to identify, understand, and react to others’ thoughts, feelings, behavior, and experiences” (Murray et al., 2019, p. 1366). It is a multidimensional construct that covers both cognitive and affective perceptions (de Kervenoael et al., 2020; Powell and Roberts, 2017). In extant research on service management and information systems, empathy can be defined as the required skills for successful interactions between consumers and robots (Birnbaum et al., 2016). According to Czaplewski et al. (2002)’s RATER model, empathy is also categorized as one of the five dimensions (reliability, assurance, tangibles, empathy and responsiveness) of service quality (Czaplewski et al., 2002). In the traditional service setting, employees have been trained to always take the consumers’ point of view. Even if many consumers may ask similar questions every day, professional employees are not supposed to be impatient and they are trained to always be empathetic and offer alternatives to address consumers’ problems. In the online service setting, some text-based chatbots have been set to automatically reply to consumers when there is a need. The chatbot with a high level of empathy can understand the consumers’ needs better and provide alternatives to address their issues. As a result, consumers will be more likely to have a positive perception of the services provided by the chatbot. 2.2.2 Friendliness of the chatbot. In a traditional business setting, employees are required to be friendly to consumers and convey positive emotions to them (Tsai and Huang, 2002). Friendly and helpful service providers can please consumers and increase their positive perception of the service quality (Chen et al., 2012). Following Pugh (2001), when employees convey more positive emotions, consumers will be more satisfied with the service, resulting in their positive response to the store, such as repurchase and positive word of mouth. In previous research on consumers’ acceptance of robots, users gave high ratings on the effects of friendliness and trust (Fridin and Belokopytov, 2014). When interacting with the text- based chatbots, rigid or indifferent responses from the chatbot will result in consumer pushback. On the other hand, a friendly manner will positively impact consumers’ moods and perceptions of the store and brands (Tsai and Huang, 2002). 2.3 Organism: trust toward the text-based chatbot INTR As aforementioned, “organism” refers to individuals’ cognitive or affective condition. The 32,2 formation of trust also involves individuals’ emotions and feelings. For example, individuals may prefer not to trust others simply because of their perceived bad feelings (Hoff and Bashir, 2015). Therefore, the process of consumers’ trust formation on the text-based chatbot can be regarded as the “organism” in this research setting. Specifically, trust can be defined as the individual or group’s willingness to be vulnerable to the other party (Hoy and Tschannen-Moran, 1999; Mayer et al., 1995). Generally, trust has been widely discussed in many research fields in recent years, such as psychology, economics, sociology and information systems (Breuer et al., 2016; Ert et al., 2016). In extant research, trust has been investigated mostly in the context of interpersonal interaction. Nowadays, with the growing interaction between humans and robots, trust can also be applied to explain the way people interact with technology (Hoff and Bashir, 2015). In this research, we mainly focus on consumers’ trust toward the text-based chatbot during the online purchasing process. Extant research has investigated the antecedents of trust from several lenses. The most widely accepted antecedents of trust are ability, integrity and benevolence (Jarvenpaa et al., 1997). Considering trust between humans and autonomous machines, anthropomorphic characteristics will impact users’ trust perception (Złotowski et al., 2017). According to social response theory (Nass et al., 1994), the interactions between human and computer are fundamentally social (Adam et al., 2020). Even if the consumers know that the computer does not hold feelings or emotions as humans do, they still tend to perceive the computers or machines as social actors. The theory has also provided evidence on studies investigating how humans adopt social rules to the anthropomorphically designed machines (Adam et al., 2020). In other words, humans will react similarly to those machines with human characteristics as they react to humans (Ischen et al., 2019). Moreover, the anthropomorphism theory also argues that individuals’ human-like perceptions of the robots will facilitate the human–robot interaction, thus increasing customers’ positive perception of the interactions (Blut et al., 2021). As aforementioned, we mainly focus on empathy and friendliness as the anthropomorphic attributes of the chatbots. Based on social response theory and anthropomorphism theory, we can infer that consumers may perceive the text-based chatbots as more trustworthy if they are empathetic and customer-friendly during the interaction with the chatbots. Hence, we postulate the following hypotheses: H1. Empathy of the text-based chatbot is positively related to consumers’ trust toward the chatbot. H2. Friendliness of the text-based chatbot is positively related to consumers’ trust toward the chatbot. 2.4 Moderators: task complexity and disclosure of the text-based chatbot 2.4.1 Task complexity. The customer service in an e-commerce platform is supposed to help consumers to address their problems during the whole customer journey. The nature of customer service is to meet customers’ needs across the stages of prepurchase, purchase and postpurchase (Lemon and Verhoef, 2016; Zhang et al., 2016). In the prepurchase stage, consumers may need to search for information, recognize their needs and consider whether the product will meet their needs. During the interaction process, consumers may ask about the details, functions or discount strategies of the product itself. Thus, the task of the chatbot in this stage is to provide specific information about the products and assist the consumers with decision-making. In the purchase stage, consumers’ behavior includes mainly choice, ordering and payment. During this process, consumers’ interaction with the chatbot may be less than in the other two stages. In the postpurchase stage, consumers’ behaviors mainly Consumers’ contain postpurchase engagement or service requests. According to previous studies, response to interaction between the store and consumer in this stage covers service recovery (Kelley and text-based Davis, 1994), repurchase decision (Bolton, 1998), products returning (Wood, 2001) or other chatbots forms of engagement (van Doorn et al., 2010). Tasks that the chatbot needs to address in this process may be more complex than the other two stages. According to the task–technology fit theory (Goodhue and Thompson, 1995), the technology should meet the customer’s needs or requirements for specific tasks (Blut et al., 2021). Once the technology can meet the customer’s specific needs, the interaction will be more satisfying. Therefore, in addition to the attributes of the chatbots in H1 and H2, more emphasis should be attached to the feature of the tasks that the chatbots address. Based on the task–technology fit theory, the attributes of the task can moderate the relationship between chatbots’ attributes and customers’ intention. Therefore, we focus on the effects of task complexity on customers’ behavioral outcomes. The effects of task complexity have been discussed in previous studies from several facets, such as task performance or team performance (Bjørn and Ngwenyama, 2009; Dayan and Di Benedetto, 2010). As aforementioned, tasks in different customer journey stages are different. Therefore, we applied task complexity in this research setting to better understand consumers’ trust toward the chatbot in different contexts. When the problem is complex to address, consumers may turn to customer service for help. In this circumstance, they may pay more attention to the customer service providers’ level of knowledge and professionalism. Thus, only giving rigid and standardized responses with friendliness and empathy cannot improve consumers’ perceptions. In contrast, they may even doubt the chatbot’s ability to address the problems. Consequently, we propose the following hypotheses: H3a. Task complexity moderates the relationship between empathy and consumers’ trust toward the chatbot, such that the positive relationship is weaker when the task is complex. H3b. Task complexity moderates the relationship between friendliness and consumers’ trust toward the chatbot, such that the positive relationship is weaker when the task is complex. 2.4.2 Identity disclosure of the chatbot. According to Davenport (2019), one way to increase trust in automation is to increase the disclosure of the system or its adoption as much as possible (Davenport, 2019). Consumers also have the right to be informed of the identity of the chatbot (Luo et al., 2019). Disclosure in the e-commerce setting refers to the state when the text-based chatbots reveal their machine identity before interacting with consumers (Luo et al., 2019). For example, in some online stores, the customer service will disclose its identity using the following sentences “Dear consumer, welcome to ABC store and I am the AI chatbot of this store” at the beginning of the interaction with them. Some stores also choose to state the AI identity beside the profile photo of the customer service staff. While a text-based chatbot can provide several business benefits for the company, some consumers may always feel uncomfortable talking with chatbots and perceive them as less empathetic than traditional customer service employees. While they are informed that they are interacting with a chatbot instead of a human, the positive effects of chatbots’ empathy and friendliness will be increased due to the empathetic and friendly service. Accordingly, we propose the following hypotheses: H4a. Disclosure of the chatbot moderates the relationship between empathy and consumers’ trust toward the chatbot, such that the positive relationship is stronger when the identity of the chatbot is disclosed. H4b. Disclosure of the chatbot moderates the relationship between friendliness and INTR consumers’ trust toward the chatbot, such that the positive relationship is stronger 32,2 when the identity of the chatbot is disclosed. 2.5 Response: avoidance/approach behavior toward the text-based chatbot According to the SOR framework, consumers’ response behavior consists of both averting behaviors and approach behaviors (Kim et al., 2020). Extant studies mainly focus on consumers’ approach behaviors or intention toward the automated bots. To provide a holistic view of consumers’ response to the chatbots, we take both consumers’ averting response (resistance) and approach response (reliance) to the text-based chatbot into consideration. 2.5.1 AI reliance. Individualstendtorelymoreonahumanratherthananalgorithm for forecasting stock prices (Onkal et al.,2009). Unlike humans, automated algorithms cannot provide reasons or explanations for their decision, thus increasing consumers’ distrust. Although previous research has investigated consumers’ future intention with the bots (Heerink et al., 2010; de Kervenoael et al., 2020), there has been little investigation and validation of practical interventions for the customer service provider to increase consumers’ reliance on the automated machines instead of a human, especially in the e-commerce setting. Therefore, we tend to provide empirical evidence on consumers’ reliance on the AI-based chatbot. As discussed in the literature on trust, cognitive trust involves individuals’ confidence in others’ performance and reliability. Therefore, we suggest that if consumers perceive the chatbot as trustworthy during the interaction process, they will be more likely to rely on the chatbot to address their problems or provide decisions. Thus, we hypothesize that: H5. Consumers’ trust toward the chatbot is positively related to their reliance on the chatbot. 2.5.2 AI resistance. Extant research on consumers’ response to AI mainly focuses on their acceptance or future use intention of the AI-based bots (Ciechanowski et al., 2019; Fridin and Belokopytov, 2014; Heerink et al., 2010; Piçarra and Giger, 2018; Wirtz et al., 2018). Although consumer pushback has been proven to be a major concern, only a few studies have investigated consumers’ resistance to the AI-based bots (Longoni et al., 2019). Findings in the healthcare context indicate that AI service providers are less likely to account for consumers’ personal requests or unique characteristics. This can lead consumers to perceive the AI service provider as less trustworthy and they may exhibit resistant behavior toward them. In the e-commerce setting, when consumers find the text-based chatbots are not trustworthy or not able to meet their needs, they will avoid interacting with them and switch to the traditional human customer service (Longoni et al., 2019). Therefore, the following hypothesis is proposed: H6. Consumers’ trust toward the chatbot is negatively related to their resistance to the chatbot. The conceptual model is presented in Figure 1. 3. Research method 3.1 Sample and data collection We measured the variables using existing scales adapted from previous research (see the Appendix). A pretest was conducted with 30 students in our research group to ensure the quality of the survey. Suggestions were also provided by these participants. After Moderators Consumers’ response to Unique attributes of Task complexity text-based Customer outcomes text-based chatbot chatbots H3b H3a Reliance on Empathy H1 the chatbot H5 Trust towards the chatbot H6 H2 Resistance to Friendliness H4b H4a the chatbot Figure 1. A conceptual model of customers’ response to Disclosure of the the text-based chatbot chatbot rearranging the survey, the formal version was then distributed mainly through a professional survey website in China via a social network. Several representative chatbots in the e-commerce setting were provided at the beginning of the survey. Those without any chatbot experience during the purchasing process on e-commerce sites were filtered out. The participants were told that their information was anonymous. Finally, a total of 441 respondents completed the survey including the pretest. After discarding incomplete and rushed responses, there were a total of 299 effective responses, which were adequate for this research. Our target participants were consumers who had experienced the text-based chatbot in an e-commerce setting before. Table 1 presents the demographics of these participants. As is shown in the table, the participants consist of 28.8% males and 71.2% females. More than 82.9% of them are below 35 years old. The age range of the participants is consistent with this research setting because young people nowadays interact with text-based chatbots more frequently in the online purchasing process. Additionally, most of the participants are shown to frequently interact with the text-based chatbots in the purchasing process. Therefore, the participants are suitable for our research setting. Constructs and measurement items Category Frequency (N 5 299) % Gender Males 86 28.8 Females 213 71.2 Age <25 147 49.1 26–35 101 33.8 36–45 28 9.4 >45 23 7.7 Experience with the chatbot in e-commerce Seldom 38 12.7 Less 95 31.8 General 106 35.5 Table 1. More 50 16.7 Demographics of A lot 10 3.3 participants (N 5 299) 3.2 Measurements INTR The data was collected in China, so we conducted the forward–backward translation method 32,2 to ensure the validity of translation (Lin et al., 2019). We used multi-item measures with a seven-point Likert-style scale (1 5 strongly disagree, 7 5 strongly agree). The survey included seven reflective constructs, including the unique attributes of text-based chatbots (e.g. empathy, friendliness), moderators (e.g. task complexity, disclosure of the text-based chatbot), organism of the consumers’ perception (e.g. trust toward the text-based chatbot) and responses of the consumers (e.g. resistance and reliance on the text-based chatbot). Control variables such as age, gender and consumers’ experience with a text-based chatbot when purchasing online were also considered. Specifically, the empathy measure was developed following the guidance of previous research (Cheng et al., 2018; Czaplewski et al., 2002; de Kervenoael et al., 2020). The friendliness measure was measured according to Tsai and Huang (2002). The task complexity measure was developed by adopting the measurements by Morgeson and Humphrey (2006). The disclosure measure was evaluated according to the information disclosure measure in extant research (Jensen and Yetgin, 2017; Luo et al., 2019; Starzyk et al., 2006). The trust measure was evaluated according to prior research on trust (Cyr et al., 2009; Everard and Galletta, 2005). The resistance measure was developed following previous research on the resistance to information technology implementation (Kim and Kankanhalli, 2009; Lapointe and Rivard, 2005). The reliance measure was developed according to previous self-reliance measures (Edwards et al., 2003). 3.3 Data analysis We conducted the data analysis with SmartPLS 3.2.8 and SPSS 22.0. Specifically, SmartPLS was used to analyze the reliability, validity and common method variance (CMV) (Shiau et al., 2019), while SPSS was used to test the hypotheses. The data analysis process was conducted in three stages: reliability and validity analysis, CMV and hypotheses testing. 3.3.1 Reliability and validity analysis. We first performed several statistical tests to ensure the reliability and validity of all the measures. As is shown in Table 2, we employed the value of Cronbach’s α to test the internal consistency reliability. The Cronbach’s α of all the measures exceeds the recommended threshold, indicating acceptable reliability. Moreover, the variance inflation factor (VIF) values of the subconstructs are all below the recommended threshold of 10, suggesting acceptable reliability (Hua et al., 2020). The results of VIF also suggest that multicollinearity is not a significant concern in this research (Thatcher et al., 2018). Validity of the construct measures was tested with two procedures. Firstly, each construct’s average variance extracted (AVE) is above 0.5 and most of the constructs’ factor loadings are above 0.7, suggesting satisfactory convergent validity. Discriminant validity was measured with the Fornell–Larcker criterion and cross-loading evaluation. As is shown in Table 3, the square root of AVE for each construct is greater than their correlations with other constructs (Liang et al., 2019). As is shown in Table 4, each indicator’s factor loading is greater than its cross-loadings with other constructs, suggesting satisfactory discriminant validity. 3.3.2 Common method variance. To address the concerns of CMV due to the self-reported surveys, we conducted several procedural and statistical remedies. In terms of procedural remedies, we tried to assure the respondents that they can answer these questions as honestly as possible to reduce the evaluation apprehension and protect their anonymity (Cheng et al., 2020). In addition, we tried to improve the scale items by avoiding vague concepts and keeping the questions simple and specific (Podsakoff et al., 2003). In terms of statistical remedies, we adopted the two most popular tests for the CMV, including Harman’s one-factor test (Cram et al., 2019; Saldanha et al., 2017) and a marker variable technique (Griffith and Lusch, 2007; Hua et al., 2020; Krishnan et al., 2006; Lindell and Brandt, 2000) Specifically, Consumers’ VIF Factor loadings Cronbach’s α Average variance response to Constructs Items (<10) (>0.7) (>0.7) extracted (AVE) (>0.5) text-based Empathy (EMP) EMP1 2.191 0.885*** 0.863 0.785 chatbots EMP2 2.471 0.902*** EMP3 2.081 0.871*** Friendliness (FRI) FRI1 3.469 0.934*** 0.926 0.870 FRI2 4.014 0.937*** FRI3 3.358 0.927*** Task complexity COM1 2.515 0.893*** 0.881 0.807 (COM) COM2 3.511 0.918*** COM3 2.272 0.883*** Disclosure of the DIS1 1.454 0.710*** 0.717 0.733 chatbot (DIS) DIS2 1.454 0.981*** Trust toward the TRU1 1.337 0.671*** 0.782 0.697 chatbot (TRU) TRU2 2.415 0.919*** TRU3 2.161 0.893*** Reliance on the REL1 4.270 0.946*** 0.951 0.910 chatbot (REL) REL2 5.675 0.957*** REL3 5.981 0.960*** Resistance to the RES1 1.767 0.792*** 0.773 0.691 chatbot (RES) RES2 2.442 0.918*** RES3 1.574 0.776*** Table 2. Note(s): VIF – Variance Inflation Factor. The recommended threshold of each indicator is in the Psychometric parentheses; ***p < 0.001 properties of measures Harman’s single-factor test indicated no single factor. Then we conducted the “marker variable” test. Following Lindell and Brandt (2000), a marker variable is defined as a variable that is minimally correlated with the study variables. In this study, we chose age as a marker variable (Griffith and Lusch, 2007). The results suggest that the marker variable is not significantly related to all the variables in this model. Furthermore, the structural model paths of interest maintain statistical significance after including the marker variable. Therefore, CMV is not a serious problem in this study. 3.3.3 Hypotheses testing. We conducted the ordinary least squares regression to empirically estimate our proposed hypotheses. The results of the hypotheses tests are presented in Table 5 and Figure 2. As is shown in Table 5, Model 1 only contains the control variables. To answer the first research question (RQ1)(What is the impact of the unique attributes of a text-based chatbot on consumer responses to the text-based chatbot in the Construct M SD EMP FRI COM DIS TRU REL RES EMP 4.129 1.088 0.886 FRI 5.183 0.989 0.236 0.933 COM 2.860 1.043 0.182 0.298 0.898 DIS 4.489 0.943 0.142 0.202 0.175 0.856 TRU 4.266 1.000 0.753 0.262 0.227 0.194 0.835 Table 3. REL 4.457 1.330 0.587 0.012 0.057 0.111 0.625 0.954 Mean, SD and RES 5.034 0.996 0.302 0.097 0.012 0.106 0.402 0.337 0.831 discriminant validity Note(s): Diagonal elements (italic) are the square root of AVE for each construct. Off-diagonal elements are the evaluation based on correlations between constructs; EMP: empathy, FRI: friendliness, COM: task complexity, DIS: disclosure of the Fornell–Larcker chatbot, TRU: trust toward the chatbot, REL: reliance on the chatbot, RES: resistance to the chatbot criterion INTR Indicator EMP FRI COM DIS TRU REL RES 32,2 EMP1 0.885 0.216 0.220 0.165 0.679 0.522 0.299 EMP2 0.902 0.156 0.118 0.106 0.674 0.565 0.222 EMP3 0.871 0.256 0.145 0.106 0.648 0.473 0.281 FRI1 0.247 0.934 0.289 0.212 0.260 0.053 0.110 FRI2 0.197 0.937 0.304 0.192 0.224 0.046 0.078 FRI3 0.211 0.927 0.243 0.161 0.246 0.018 0.081 COM1 0.168 0.273 0.893 0.187 0.222 0.062 0.001 COM2 0.151 0.281 0.918 0.111 0.155 0.123 0.056 COM3 0.167 0.252 0.883 0.160 0.219 0.009 0.012 DIS1 0.009 0.169 0.132 0.710 0.059 0.034 0.116 DIS2 0.165 0.191 0.170 0.981 0.213 0.121 0.093 TRU1 0.481 0.426 0.286 0.326 0.671 0.248 0.120 TRU2 0.701 0.174 0.188 0.138 0.919 0.603 0.402 TRU3 0.677 0.156 0.147 0.102 0.893 0.628 0.412 REL1 0.549 0.050 0.043 0.069 0.612 0.946 0.334 REL2 0.560 0.024 0.066 0.117 0.591 0.957 0.337 REL3 0.572 0.006 0.056 0.133 0.586 0.960 0.294 RES1 0.264 0.286 0.163 0.110 0.317 0.421 0.792 RES2 0.283 0.045 0.006 0.106 0.367 0.323 0.918 RES3 0.203 0.077 0.134 0.048 0.315 0.097 0.776 Table 4. Note(s): All italic values are significant with p < 0.001; EMP: empathy, FRI: friendliness, COM: task Discriminant validity based on cross-loading complexity, DIS: disclosure of the chatbot, TRU: trust toward the chatbot, REL: reliance on the chatbot, RES: evaluation resistance to the chatbot e-commerce context?), we contained the main effects in Model 2 and tested the impacts of empathy and friendliness on consumers’ trust toward the text-based chatbot. The results of Model 2 indicate that empathy and friendliness are both positively related to consumers’ trust (Model 2: β 5 0.652, p < 0.001; β 5 0.093, p < 0.05). Thus, hypotheses 1 and hypothesis 2 are supported. To answer the second research question (RQ2)(How does the task complexity and chatbot identity disclosure moderate the relationship between these attributes and consumer responses?), we adopted a stepwise procedure and tested the moderation effects in H3a and H3b. Model 3 in Table 5 reports the moderation test results including the main effects and interaction effects. We mean-centered the construct scores and created the interaction terms of empathy, friendliness, task complexity and disclosure of the chatbot. After adding the interaction terms to the model, the results show that task complexity significantly moderates the relationship between friendliness and consumers’ trust toward the chatbot (Model 3: β 5 0.073, p < 0.05), thus supporting H3b. Using the same methods and procedures, we also tested for the moderation effects of disclosure of the chatbot on the relationship between chatbot attributes and consumers’ trust. The results indicate that disclosure of the chatbot significantly moderates the relationship between empathy and consumers’ trust (Model 3: β 5 0.154, p < 0.05). In addition, disclosure of the chatbot also significantly moderates the relationship between friendliness and consumers’ trust (Model 3: β 5 0.184, p < 0.01), thus supporting H4b, while H4a is not supported. To answer the third research question (RQ3)(How does consumers’ trust impact their responses to the text-based chatbot in e-commerce?), we tested the relationship between trust and consumers’ response (reliance and resistance) in Model 4, Model 5, Model 6 and Model 7. As for the relationship between consumers’ trust and reliance on the chatbot, Model 4 only contains the control variables, while the results in Model 5 indicate the positive relationship between consumers’ trust and reliance on the chatbot (Model 5: β 5 0.783, p < 0.001), thus Consumers’ response to text-based chatbots Table 5. Regression results Trust toward the chatbot Reliance on the chatbot Resistance to the chatbot Variables Model 1 Model 2 Model 3 Model 4 Model 5 Model 6 Model 7 Main effects Empathy 0.652*** (0.037) 0.527*** (0.065) Friendliness 0.093* (0.042) 0.275*** (0.07) Task complexity 0.079* (0.039) 0.084* (0.04) Disclosure of the 0.084* (0.041) 0.008 (0.048) chatbot Trust toward the 0.783*** (0.061) 0.382*** (0.054) chatbot Interaction effects Empathy * Task 0.024 (0.038) complexity Friendliness * 0.073* (0.032) Task complexity Empathy * 0.154* (0.067) Disclosure of the chatbot Friendliness * 0.184** (0.066) Disclosure of the chatbot Control variables Age 0.097 (0.065) 0.05 (0.043) 0.051 (0.042) 0.23** (0.085) 0.154* (0.069) 0.006 (0.065) 0.043 (0.06) Gender 0.032 (0.130) 0.034 (0.085) 0.021 (0.084) 0.097 (0.171) 0.072 (0.137) 0.067 (0.13) 0.055 (0.12) Experience with 0.021 (0.058) 0.007 (0.038) 0.005 (0.038) 0.077 (0.076) 0.093 (0.061) 0.023 (0.058) 0.031 (0.053) the chatbot Constant 4.204*** (0.321) 0.936* (0.384) 0.776 (0.461) 3.014*** (0.421) 0.277 (0.425) 4.971*** (0.32) 6.58*** (0.372) Observations 299 299 299 299 299 299 299 R-squared 0.009 0.584 0.605 0.03 0.376 0.009 0.137 Note(s): Standard errors are in the parentheses; *p < 0.05, **p < 0.01, ***p < 0.001 Moderators INTR 32,2 Unique attributes of Task complexity Customer outcomes text-based chatbot H3b H3a Reliance on 0.024 – 0.073* H1 the chatbot Empathy H5 2 R =37.6% Trust towards the Chatbot H6 R = 60.5% Resistance to H2 the chatbot Friendliness H4b H4a R =13.7% 0.184** – 0.154* Disclosure of the Figure 2. chatbot Theoretical model with results Note(s): *p < 0.05, **p < 0.01, ***p < 0.001 supporting H5. Similarly, for the relationship between consumers’ trust and resistance to the chatbot, Model 6 contains the control variables of the research model. Finally, the results in Model 7 show that consumers’ trust is negatively related to their resistance to the text-based chatbot (Model 7: β 5 0.382, p < 0.001), thus supporting H6. 4. Discussions and conclusions 4.1 Summary of findings Following Mehrabian and Russell’s SOR framework (1974), the current study tests a model of consumers’ trust and response to the text-based chatbot in the e-commerce setting, involving the moderating role of the different complexity of a task and the disclosure of the use of a chatbot. The specific findings of this study are summarized as follows: Considering the first research question (RQ1), this study comes to the following findings. We mainly target the impacts of two specific attributes (empathy and friendliness) on consumers’ responses in this study. Consumers’ perceived empathy and friendliness of the text-based chatbot are both positively connected to their trust toward the chatbot. Specifically, the positive effect of empathy on trust is larger than that of friendliness. Our explanation for this finding is that, compared with a friendly manner, consumers prefer the chatbot to be able to appreciate their perspective and meet their specific needs. Therefore, empathy of the customer service provider is essential for successful interactions. Considering the second research question (RQ2), this study provides empirical evidence on the boundary conditions of consumers’ perception and response. We investigate the moderating effects of task complexity and text-based chatbot disclosure on the relationship between the chatbots’ attributes and trust. The findings indicate that task complexity negatively moderates the relationship between friendliness and consumers’ trust toward the chatbot, such that when the consumers’ task is complex, the positive effect of friendliness on consumers’ trust is weaker. Conversely, task complexity does not significantly impact the relationship between empathy and consumer’ trust toward the chatbot. Our explanation for this finding is that, when the task is complex to address, consumers may be more concerned about the chatbots’ professional level and ability to solve problems, instead of their attitudes or service manner. Considering the identity disclosure of the text-based chatbot, the findings Consumers’ indicate that the disclosure of the chatbot negatively moderates the relationship between response to empathy and consumers’ trust toward the chatbot, while it positively moderates the text-based relationship between friendliness and consumers’ trust toward the chatbot. chatbots Considering the third research question (RQ3), we take both consumers’ approach response (reliance on the chatbot) and averting response (resistance to the chatbot) into consideration. Findings indicate that consumers’ trust toward the chatbot is positively related to their reliance on the chatbot and negatively related to consumers’ resistance to the chatbot. More specifically, the positive effect between consumers’ trust and reliance is greater than the negative effect between consumers’ trust and resistance. Thus, the effects of consumers’ trust on their approach response are larger than that on their averting response. 4.2 Theoretical contributions This study makes several theoretical contributions. First, this study contributes to the SOR framework by extending it to the automated chatbot in the e-commerce context. Previously, the SOR framework has mainly targeted traditional contexts (Kamboj et al., 2018; Kim et al., 2020). This extension contributes to the SOR framework by investigating specific attributes of automated text-based chatbots and how such attributes form consumers’ trust and response in the e-commerce setting. The theoretical model in this research also contributes to the information systems literature. Second, this study also considers the distinct characteristics of customer service in e-commerce by incorporating task complexity (Bjørn and Ngwenyama, 2009; Dayan and Di Benedetto, 2010) and chatbot identity disclosure (Luo et al., 2019) as the boundary conditions of consumers’ trust in this context. Prior research on consumers’ response to the automated chatbot mainly discussed several attributes of the automated bots, such as anthropomorphism (Castelo et al., 2019; Desideri et al., 2019; Yuan and Dennis, 2019), humanness (Wirtz et al., 2018) and adaptivity (Heerink et al., 2010), the boundary conditions of these effects are largely ignored. The findings of this research show that when the tasks are different and the disclosure level is different, consumers’ responses will also be different. Therefore, identifying the boundary conditions can also provide possible ways to reconcile the mixed findings in relevant research, thus opening an avenue for future study. Third, this study provides a deeper understanding of consumers’ responses to the automated AI-based chatbot, involving both approach response and averting response. As a new emerging technology, AI-based chatbots are providing the company and consumer with an expanding variety of benefits, therefore the effective application of automated bots is a big concern, especially in shaping the future of e-commerce. The extant literature on consumers’ response to the AI-based bots has examined consumers’ future intention with the automated machine (Ciechanowski et al., 2019; Fridin and Belokopytov, 2014; Heerink et al., 2010; Piçarra and Giger, 2018; Wirtz et al., 2018). Therefore, there is a call for a systematic understanding of consumers’ responses to a chatbot. This research adds to the understanding of both consumers’ reliance and resistance to a text-based chatbot. Boundary conditions of consumers’ responses are also identified, thus filling a gap in previous research that only focused on consumers’ positive responses to the automated bots. 4.3 Practical implications Our study also provides several practical implications for both customer service providers and the technology developers. First, as the AI-based chatbot is quite ubiquitous in the e-commerce setting, this research contributes to the limited understanding of consumers’ response to the chatbot in the business context. Although the automated bots can arouse users’ interest during the interactions, the online store should also be aware of consumers’ intention to resist. The findings of this study can arouse the attention of the online stores that INTR want to utilize chatbots more effectively than the human alternative but also more effective 32,2 than their competitors. Second, the findings of this study provide suggestions for the technology developer that creates and customizes the chatbot. This study provides empirical evidence for the effects of task complexity and identity disclosure. Specifically, consumers’ perceptions of the chatbot will be different when their tasks are different. Interestingly, when the task is complex, the positive effects of chatbots’ friendliness on consumers’ trust will be weaker. Therefore, we suggest that it is more important for the chatbot to increase their professional level and address consumers’ problems instead of merely replying to the consumers with standard answers in a friendly manner. Moreover, the disclosure of the machine identity should also be considered. Consumers’ perceived disclosure of the automated chatbot may have different effects on the relationship between empathy/friendliness and trust. Therefore, the technology developer that customizes the chatbot should balance its characteristics to cater to consumers’ different needs. Given the current capabilities of AI, this will require additional time and resources to further train and customize the chatbot, butitisworth it. 4.4 Limitations and future research There are also limitations in this research. First, apart from empathy and friendliness, there are also other attributes of the chatbot that will affect consumers’ responses, such as the intelligence level of the chatbot and passing the “Turing test” (Turing and Haugeland, 1950; Castelo et al., 2019; Desideri et al., 2019; Heerink et al., 2010; Wirtz et al., 2018; Yuan and Dennis, 2019). More specific characteristics of the chatbots will be taken into consideration in future research. Second, as there is a distinction between humans and machines, especially in the customer service institution, future researchshouldalsoextendthe understanding of the difference between human and machine customer service providers. Third, this study only investigated the chatbot in the e-commerce setting. While chatbots developed for the other domains (e.g. healthcare, sharing economy) may behave differently when compared to chatbots developed for the e-commerce setting. Specific issues need to be detected in other domains in future studies. Moreover, the characteristics of the customers may also have an impact on their acceptance or perception of the chatbot during purchasing. Therefore, attributes of customers should be taken into consideration in future research. References Adam, M., Wessel, M. and Benlian, A. (2020), “AI-based chatbots in customer service and their effects on user compliance”, Electronic Markets, pp. 1-19, doi: 10.1007/s12525-020-00414-7. Araujo, T. (2018), “Living up to the chatbot hype: the influence of anthropomorphic design cues and communicative agency framing on conversational agent and company perceptions”, Computers in Human Behavior, Vol. 85, pp. 183-189. Birnbaum, G.E., Mizrahi, M., Hoffman, G., Reis, H.T., Finkel, E.J. and Sass, O. (2016), “What robots can teach us about intimacy: the reassuring effects of robot responsiveness to human disclosure”, Computers in Human Behavior, Vol. 63, pp. 416-423. Bitner, M.J. (1992), “Servicescapes: the impact of physical surroundings on customers and employees”, Journal of Marketing, Vol. 56 No. 2, pp. 57-71. Bjørn, P. and Ngwenyama, O. (2009), “Virtual team collaboration: building shared meaning, resolving breakdowns and creating translucence”, Information Systems Journal, Vol. 19 No. 3, pp. 227-253. € Blut, M., Wang, C., Wunderlich, N.V. and Brock, C. (2021), “Understanding anthropomorphism in Consumers’ service provision: a meta-analysis of physical robots, chatbots, and other AI”, Journal of the response to Academy of Marketing Science, pp. 1-27, doi: 10.1007/s11747-020-00762-y. text-based Bolton, R.N. (1998), “A dynamic model of the duration of the customer’s relationship with a continuous chatbots service provider: the role of satisfaction”, Marketing Science, Vol. 17 No. 1, pp. 45-65. Breuer, C., Huffmeier, € J. and Hertel, G. (2016), “Does trust matter more in virtual teams? A meta- analysis of trust and team effectiveness considering virtuality and documentation as moderators”, Journal of Applied Psychology, Vol. 101 No. 8, pp. 1151-1177. Castelo, N., Bos, M.W. and Lehmann, D.R. (2019), “Task-dependent algorithm aversion”, Journal of Marketing Research, Vol. 56 No. 5, pp. 809-825. Chang, H.J., Eckman, M. and Yan, R.N. (2011), “Application of the stimulus-organism-response model to the retail environment: the role of hedonic motivation in impulse buying behavior”, International Review of Retail, Distribution and Consumer Research, Vol. 21 No. 3, pp. 233-249. Chen, C.C., Lin, S.Y., Cheng, C.H. and Tsai, C.C. (2012), “Service quality and corporate social responsibility, influence on post-purchase intentions of sheltered employment institutions”, Research in Developmental Disabilities, Vol. 33 No. 6, pp. 1832-1840. Cheng, X., Fu, S. and de Vreede, G.J. (2018), “A mixed method investigation of sharing economy driven car-hailing services: online and offline perspectives”, International Journal of Information Management, Vol. 41 No. 10, pp. 57-64. Cheng, X., Bao, Y. and Zarifis, A. (2020), “Investigating the impact of IT-mediated information interruption on emotional exhaustion in the workplace”, Information Processing and Management, Vol. 57 No. 6, p. 102281. Ciechanowski, L., Przegalinska, A., Magnuski, M. and Gloor, P. (2019), “In the shades of the uncanny valley: an experimental study of human–chatbot interaction”, Future Generation Computer Systems, Vol. 92, pp. 539-548. Cram, W.A., D’arcy, J. and Proudfoot, J.G. (2019), “Seeing the forest and the trees: a meta-analysis of the antecedents to information security policy compliance”, MIS Quarterly, Vol. 43 No. 2, pp. 525-554. Cyr, D., Head, M., Larios, H. and Pan, B. (2009), “Exploring human images in website design: a multi- method approach”, MIS Quarterly, Vol. 33 No. 3, pp. 539-566. Czaplewski, A.J., Olson, E.M. and Slater, S.F. (2002), “Applying the RATER model for service success”, Marketing Management, Vol. 11 No. 1, pp. 14-17. Daugherty, P.R., James Wilson, H. and Michelman, P. (2019), “Revisiting the jobs artificial intelligence will create”, MIT Sloan Management Review, Vol. 60 No. 4, pp. 1-8. Davenport, T.H. (2019), “Can we solve AI’s ’trust problem’?”, MIT Sloan Management Review, Vol. 60 No. 2, pp. 18-19. Dayan, M. and Di Benedetto, C.A. (2010), “The impact of structural and contextual factors on trust formation in product development teams”, Industrial Marketing Management, Vol. 39 No. 4, pp. 691-703. de Kervenoael, R., Hasan, R., Schwob, A. and Goh, E. (2020), “Leveraging human-robot interaction in hospitality services: incorporating the role of perceived value, empathy, and information sharing into visitors’ intentions to use social robots”, Tourism Management,Vol. 78, p. 104042. Desideri, L., Ottaviani, C., Malavasi, M., di Marzio, R. and Bonifacci, P. (2019), “Emotional processes in human-robot interaction during brief cognitive testing”, Computers in Human Behavior, Vol. 90, pp. 331-342. Donovan, R. and Rossiter, J. (1982), “Store atmosphere: an environmental psychology approach”, Journal of Retailing, Vol. 58 No. 1, pp. 34-57. Duan, Y., Edwards, J.S. and Dwivedi, Y.K. (2019), “Artificial intelligence for decision making in the era INTR of Big Data – evolution, challenges and research agenda”, International Journal of Information 32,2 Management, Vol. 48, pp. 63-71. Edwards, J.C., Rust, K.G., McKinley, W. and Moon, G. (2003), “Business ideologies and perceived breach of contract during downsizing: the role of the ideology of employee self-reliance”, Journal of Organizational Behavior, Vol. 24 No. 1, pp. 1-23. Eroglu, S.A., Machleit, K.A. and Davis, L.M. (2001), “Atmospheric qualities of online retailing: a conceptual model and implications”, Journal of Business Research, Vol. 54 No. 2, pp. 177-184. Eroglu, S.A., Machleit, K.A. and Davis, L.M. (2003), “Empirical testing of a model of online store atmospherics and shopper responses”, Psychology and Marketing, Vol. 20 No. 2, pp. 139-150. Ert, E., Fleischer, A. and Magen, N. (2016), “Trust and reputation in the sharing economy: the role of personal photos in Airbnb”, Tourism Management, Vol. 55, pp. 62-73. Everard, A. and Galletta, D.F. (2005), “How presentation flaws affect perceived site quality, trust, and intention to purchase from an online store”, Journal of Management Information Systems, Vol. 22 No. 3, pp. 56-95. Fiore, A.M. and Kim, J. (2007), “An integrative framework capturing experiential and utilitarian shopping experience”, International Journal of Retail and Distribution Management, Vol. 35 No. 6, pp. 421-442. Fridin, M. and Belokopytov, M. (2014), “Acceptance of socially assistive humanoid robot by preschool and elementary school teachers”, Computers in Human Behavior, Vol. 33, pp. 23-31. Froehlich, A. (2018), “Pros and cons of chatbots in the IT helpdesk”, available at: https://www. informationweek.com/strategic-cio/it-strategy/pros-and-cons -of-chatbots-in-the-it-helpdesk/a/d- id/1332942 (accessed 10 February 2018). Gilbert, R.L. and Forney, A. (2015), “Can avatars pass the Turing test? Intelligent agent perception in a 3D virtual environment”, International Journal of Human-Computer Studies, Vol. 73, pp. 30-36. Go, E. and Sundar, S.S. (2019), “Humanizing chatbots: the effects of visual, identity and conversational cues on humanness perceptions”, Computers in Human Behavior, Vol. 97, pp. 304-316. Goodhue, D.L. and Thompson, R.L. (1995), “Task-technology fit and individual performance”, MIS Quarterly, Vol. 19 No. 2, pp. 213-236. Griffith, D.A. and Lusch, R.F. (2007), “Getting marketers to invest in firm-specific capital”, Journal of Marketing, Vol. 71 No. 1, pp. 129-145. Heerink, M., Kr€ose, B., Evers, V. and Wielinga, B. (2010), “Assessing acceptance of assistive social agent technology by older adults: the Almere model”, International Journal of Social Robotics, Vol. 2 No. 4, pp. 361-375. Hoff, K.A. and Bashir, M. (2015), “Trust in automation: integrating empirical evidence on factors that influence trust”, Human Factors, Vol. 57 No. 3, pp. 407-434. Hoy, W.K. and Tschannen-Moran, M. (1999), “Five faces of trust: an empirical confirmation in urban elementary schools”, Journal of School Leadership, Vol. 9 No. 3, pp. 184-208. Hua, Y., Cheng, X., Hou, T. and Luo, R. (2020), “Monetary rewards, intrinsic motivators, and work engagement in the IT-enabled sharing economy: a mixed-methods investigation of internet taxi drivers”, Decision Sciences, Vol. 51 No. 3, pp. 755-785. Ischen, C., Araujo, T., Voorveld, H., van Noort, G. and Smit, E. (2019), “Privacy concerns in chatbot interactions”, in Følstad, A., Araujo, T., Papadopoulos, S., Law, E.L.-C., Granmo, O.-C., Luger, E. and Brandtzaeg, P.B. (Eds), Conversations 2019: Chatbot Research and Design, Springer, Cham, pp. 34-48. Jacoby, J. (2002), “Stimulus-organism-response reconsidered: an evolutionary step in modeling (consumer) behavior”, Journal of Consumer Psychology, Vol. 12 No. 1, pp. 51-57. Jarvenpaa, S.L., Knoll, K. and Leidner, D.E. (1997), “Is anybody out there? Antecedents of trust in global virtual teams”, Journal of Management Information Systems,Vol.14No. 4, pp. 29-64. Jensen, M.L. and Yetgin, E. (2017), “Prominence and interpretation of online conflict of interest Consumers’ disclosures”, MIS Quarterly, Vol. 41 No. 2, pp. 629-643. response to Kamboj, S., Sarmah, B., Gupta, S. and Dwivedi, Y. (2018), “Examining branding co-creation in brand text-based communities on social media: applying the paradigm of Stimulus-Organism-Response”, chatbots International Journal of Information Management, Vol. 39, pp. 169-185. Kaplan, A. and Haenlein, M. (2019), “Siri, Siri, in my hand: who’s the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence”, Business Horizons, Vol. 62 No. 1, pp. 15-25. Kelley, S.W. and Davis, M.A. (1994), “Antecedents to customer expectations for service recovery”, Journal of the Academy of Marketing Science, Vol. 22 No. 1, pp. 52-61. Kim, H.W. and Kankanhalli, A. (2009), “Investigating user resistance to information systems implementation: a status quo bias perspective”, MIS Quarterly, Vol. 33 No. 3, pp. 567-582. Kim, M.J., Lee, C.K. and Jung, T. (2020), “Exploring consumer behavior in virtual reality tourism using an extended stimulus-organism-response model”, Journal of Travel Research, Vol. 59 No. 1, pp. 69-89. Krishnan, R., Martin, X. and Noorderhaven, N.G. (2006), “When does trust matter to alliance performance?”, Academy of Management Journal, Vol. 49 No. 5, pp. 894-917. Lapointe, L. and Rivard, S. (2005), “A multilevel model of resistance to information technology implementation”, MIS Quarterly, Vol. 29 No. 3, pp. 461-491. Lemon, K.N. and Verhoef, P.C. (2016), “Understanding customer experience throughout the customer journey”, Journal of Marketing, Vol. 80 No. 6, pp. 69-96. Liang, H., Xue, Y., Pinsonneault, A. and Wu, Y. (2019), “What users do besides problem-focused coping when facing it security threats: an emotion-focused coping perspective”, MIS Quarterly, Vol. 43 No. 2, pp. 373-394. Lin, J., Luo, Z., Cheng, X. and Li, L. (2019), “Understanding the interplay of social commerce affordances and swift guanxi: an empirical study”, Information and Management, Vol. 56 No. 2, pp. 213-224. Lindell, M.K. and Brandt, C.J. (2000), “Climate quality and climate consensus as mediators of the relationship between organizational antecedents and outcomes”, Journal of Applied Psychology, Vol. 85 No. 3, pp. 331-348. Longoni, C., Bonezzi, A. and Morewedge, C.K. (2019), “Resistance to medical artificial intelligence”, Journal of Consumer Research, Vol. 46 No. 4, pp. 629-650. Luo, X., Tong, S., Fang, Z. and Qu, Z. (2019), “Frontiers: machines vs humans: the impact of artificial intelligence chatbot disclosure on customer purchases”, Marketing Science, Vol. 38 No. 6, pp. 937-947. Mayer, R.C., Davis, J.H. and Schoorman, F.D. (1995), “An integrative model of organizational trust”, Academy of Management Review, Vol. 20 No. 3, pp. 709-734. Mehrabian, A. and Russell, J.A. (1974), An Approach to Environmental Psychology, The MIT Press, Cambridge, MA. Mende, M., Scott, M.L., van Doorn, J., Grewal, D. and Shanks, I. (2019), “Service robots rising: how humanoid robots influence service experiences and elicit compensatory consumer responses”, Journal of Marketing Research, Vol. 56 No. 4, pp. 535-556. Mitchell, V. (2018), “Gartner: why humans will still be at the core of great CX”, available at: https://www. cmo.com.au/article/642649/gartner-why-humans-will-still-core-great-cx/ (accessed 20 June 2018). Mollen, A. and Wilson, H. (2010), “Engagement, telepresence and interactivity in online consumer experience: reconciling scholastic and managerial perspectives”, Journal of Business Research, Vol. 63 Nos 9–10, pp. 919-925. Morgeson, F.P. and Humphrey, S.E. (2006), “The Work Design Questionnaire (WDQ): developing and validating a comprehensive measure for assessing job design and the nature of work”, Journal of Applied Psychology, Vol. 91 No. 6, pp. 1321-1339. Murray, J., Elms, J. and Curran, M. (2019), “Examining empathy and responsiveness in a high-service INTR context”, International Journal of Retail and Distribution Management, Vol. 47 No. 12, 32,2 pp. 1364-1378. Nass, C., Steuer, J. and Tauber, E.R. (1994), “Computers are social actors”, Proceedings of the SIGCHI Conference on Human Factors in Computing Systems: Celebrating Interdependence, ACM, Boston, MA, pp. 72-78. € € Onkal, D., Goodwin, P., Thomson, M., Gonul, S. and Pollock, A. (2009), “The relative influence of advice from human experts and statistical methods on forecast adjustments”, Journal of Behavioral Decision Making, Vol. 22 No. 4, pp. 390-409. Pise, R. (2018), “Chatbot market size is set to exceed USD 134 billion by 2024 - ClickZ”, available at: https://www.clickz.com/chatbot-market-size-is-set-to-exceed-usd-1-34-billion-by-2024/215518/ (accessed 6 July 2018). Piçarra, N. and Giger, J.C. (2018), “Predicting intention to work with social robots at anticipation stage: assessing the role of behavioral desire and anticipated emotions”, Computers in Human Behavior, Vol. 86, pp. 129-146. Podsakoff, P.M., MacKenzie, S.B., Lee, J.Y. and Podsakoff, N.P. (2003), “Common method biases in behavioral research: a critical review of the literature and recommended remedies”, Journal of Applied Psychology, Vol. 88 No. 5, pp. 879-903. Powell, P.A. and Roberts, J. (2017), “Situational determinants of cognitive, affective, and compassionate empathy in naturalistic digital interactions”, Computers in Human Behavior, Vol. 68, pp. 137-148. Pugh, S.D. (2001), “Service with a smile: emotional contagion in the service encounter”, Academy of Management Journal, Vol. 44 No. 5, pp. 1018-1027. Rose, S., Clark, M., Samouel, P. and Hair, N. (2012), “Online customer experience in e-retailing: an empirical model of antecedents and outcomes”, Journal of Retailing, Vol. 88 No. 2, pp. 308-322. Saldanha, T.J.V., Mithas, S. and Krishnan, M.S. (2017), “Leveraging customer involvement for fueling innovation: the role of relational and analytical information processing capabilities”, MIS Quarterly, Vol. 41 No. 1, pp. 267-286. Seeber, I., Bittner, E., Briggs, R.O., de Vreede, T., de Vreede, G.J., Elkins, A., Maier, R., Merz, A.B., Oeste-Reiß, S., Randrup, N., Schwabe, G. and S€ollner, M. (2020), “Machines as teammates: a research agenda on AI in team collaboration”, Information and Management, Vol. 57 No. 2, p. 103174. Shawar, B.A. and Atwell, E.S. (2005), “Using corpora in machine-learning chatbot systems”, International Journal of Corpus Linguistics, Vol. 10 No. 4, pp. 489-516. Shiau, W.L., Sarstedt, M. and Hair, J.F. (2019), “Internet research using partial least squares structural equation modeling (PLS-SEM)”, Internet Research, Vol. 29 No. 3, pp. 398-406. Starzyk, K.B., Holden, R.R., Fabrigar, L.R. and MacDonald, T.K. (2006), “The personal acquaintance measure: a tool for appraising one’s acquaintance with any person”, Journal of Personality and Social Psychology, Vol. 90 No. 5, pp. 833-847. Thatcher, J.B., Wright, R.T., Sun, H., Zagenczyk, T.J. and Klein, R. (2018), “Mindfulness in information technology use: definitions, distinctions, and a new measure”, MIS Quarterly, Vol. 42 No. 3, pp. 831-847. Thompson, C. (2018), “May AI help you?”, available at: https://www.nytimes.com/interactive/2018/11/ 14/magazine/ tech-design-ai-chatbot.html (accessed 14 November 2018). Tsai, W.C. and Huang, Y.M. (2002), “Mechanisms linking employee affective delivery and customer behavioral intentions”, Journal of Applied Psychology, Vol. 87 No. 5, pp. 1001-1008. Turing, A. and Haugeland, J. (1950), “Mind design II: philosophy, psychology, and artificial intelligence”, in Haugeland, J. (Ed.), Computing Machinery and Intelligence, MIT Press, Cambridge, MA, pp. 29-56. van Doorn, J., Lemon, K.N., Mittal, V., Nass, S., Pick, D., Pirner, P. and Verhoef, P.C. (2010), “Customer Consumers’ engagement behavior: theoretical foundations and research directions”, Journal of Service response to Research, Vol. 13 No. 3, pp. 253-266. text-based Wirtz, J., Patterson, P.G., Kunz, W.H., Gruber, T., Lu, V.N., Paluch, S. and Martins, A. (2018), “Brave chatbots new world: service robots in the frontline”, Journal of Service Management, Vol. 29 No. 5, pp. 907-931. Wood, S.L. (2001), “Remote purchase environments: the influence of return policy leniency on two- stage decision processes”, Journal of Marketing Research, Vol. 38 No. 2, pp. 157-169. Wu, Y.L. and Li, E.Y. (2018), “Marketing mix, customer value, and customer loyalty in social commerce: a stimulus-organism-response perspective”, Internet Research, Vol. 28 No. 1, pp. 74-104. Yuan, L. and Dennis, A.R. (2019), “Acting like humans? Anthropomorphism and consumer’s willingness to pay in electronic commerce”, Journal of Management Information Systems, Vol. 36 No. 2, pp. 450-477. Zhang, M., Jin, B., Wang, G.A., Goh, T.N. and He, Z. (2016), “A study of key success factors of service enterprises in China”, Journal of Business Ethics, Vol. 134 No. 1, pp. 1-14. Złotowski, J., Yogeeswaran, K. and Bartneck, C. (2017), “Can we control it? Autonomous robots threaten human identity, uniqueness, safety, and resources”, International Journal of Human Computer Studies, Vol. 100, pp. 48-54. Appendix Constructs and measurement items References Empathy (1) Text-based chatbots in e-commerce usually understand Cheng et al. (2018), Czaplewski et al. (2002), the specific needs of the customers de Kervenoael et al. (2020) (2) Text-based chatbots in e-commerce usually give customers individual attention (3) Text-based chatbots in e-commerce are available whenever it’s convenient for customers (4) If a customer requires help, the text-based chatbots would do their best to help (dropped) Friendliness (1) The text-based chatbot had a kind service during our Tsai and Huang (2002) interaction (2) The text-based chatbot provides the service in a friendly manner (3) The text-based chatbot treats me nicely Task complexity (1) The task I communicate with the customer service are Morgeson and Humphrey (2006) simple and uncomplicated (2) The task I communicate with the customer service comprises relatively uncomplicated tasks (3) The task I communicate with the customer service involves performing relatively simple tasks Table A1. Construct measurements (continued) INTR Constructs and measurement items References 32,2 Disclosure of the chatbot (1) The identity of the text-based chatbot is disclosed before Jensen and Yetgin (2017), Luo et al. (2019), conversation Starzyk et al. (2006) (2) The e-commerce platform hides the identity of text-based chatbot from me 516 (3) I am informed with the identity of the text-based chatbot before conversation (dropped) Trust toward the chatbot (1) The text-based chatbot is honest and truthful Cyr et al. (2009), Everard and Galletta (2005) (2) The text-based chatbot is capable of addressing my issues (3) The text-based chatbot’s behavior and response can meet my expectations (4) I trust the suggestions and decisions provided by the text- based chatbot (dropped) Reliance on the chatbot (1) I depend on the text-based chatbot for decision making Edwards et al. (2003) (2) I depend on the text-based chatbot for service in e- commerce (3) I depend on the text-based chatbot for suggestions in purchasing process Resistance to the chatbot (1) I prefer to use the human customer service rather than use Kim and Kankanhalli (2009), Lapointe and the text-based chatbot Rivard (2005) (2) I will try to avoid using the service provided by text-based chatbot in e-commerce (3) I do not agree with the suggestions provided by the text- based chatbot in e-commerce Table A1. Note(s): Items with (dropped) were removed in the data analysis About the authors Xusen Cheng is a professor of Information Systems in the School of Information, Renmin University of China, Beijing, China. He has obtained his PhD degree in Manchester Business School in the University of Manchester in the UK. His research focuses on information system and management, e-commerce and sharing economy, AI and behavior. His research paper has been accepted/appeared in journals such as MIS Quarterly, Journal of Management Information Systems, European Journal of Information Systems, Decision Sciences, Tourism Management, Information & Management, Information Processing and Management, Group Decision and Negotiation, Information Technology and People, British Journal of Educational Technology, Computers in Human Behavior, Electronic Commerce Research, Internet Research, among others. He serves as senior editor for Information Technology and People, associate editor for Information Technology for Development, Editorial Board for Electronic Commerce Research and Editorial Review Board for Internet Research, guest editor for several journals such as British Journal of Educational Technology, Electronic Markets, Journal of Global Information Technology Management and serves as track cochair and minitrack cochair for many conferences. His papers have also been presented in the leading conferences such as International Conference of Information System (ICIS), Hawaii International Conference of System Science (HICSS) and American Conference of Information Systems (AMCIS). Ying Bao is a PhD candidate in the School of Information Technology and Management in the University of International Business and Economics, Beijing, China. Her interests focus on user behaviors and trust in the sharing economy and corporate social responsibility. Her research papers have appeared in Information Processing and Management, Internet Research, Group Decision and Negotiation. Her papers have also been presented in the leading conferences such as Hawaii International Conference of System Science (HICSS) and American Conference on Information Consumers’ System (AMCIS). response to Alex Zarifis is a research associate in the School of Business and Economics at the University of text-based Loughborough, UK. His first degree was a BSc in Management with Information Systems from the University of Leeds, followed by an MSc in Business Information Technology and a PhD in Business chatbots Administration both from the University of Manchester. His research focuses on the user’s perspective in the areas of e-commerce, trust, privacy, artificial intelligence, blockchain, fintech, leadership, online collaboration and online education. His research has featured in journals such as Information Processing and Management, Computers in Human Behavior, Information Technology and People, International Journal of Electronic Business and Electronic Commerce Research. He served as the guest editor of British Journal of Education Technology and Electronic Markets. His research has been presented at conferences such as the European Conference on Information Systems (ECIS) and Hawaii International Conference on System Sciences (HICSS). Wankun Gong is a postgraduate student in College of Humanities and Law in Beijing University of Chemical Technology, China. His research interests include e-government, e-service and e-commerce, human behavior and management, public and personal behavior analysis, trust and technology-enabled innovation. Jian Mou is an associate professor of School of Business in Pusan National University, South Korea. His research interests include e-commerce, human–computer interaction, trust and risk issues in e-service and the dynamic nature in systems use. His research has been published in the Journal of the Association for Information Systems, Internet Research, Information and Management, International Journal of Information Management, Computers in Human Behavior, Electronic Commerce Research, Information Processing and Management, Information Development, Behaviour and Information Technology, Journal of Retailing and Consumer Services, Electronic Commerce Research and Applications, Industrial Management & Data Systems, Journal of Global Information Management, IT & People, International Journal of Human–Computer Interaction and Information Technology and Tourism, as well as in the proceedings such as ICIS, ECIS, PACIS and AMCIS. He serves as senior editor for Information Technology and People, associate editor for Internet Research, and Behaviour and Information Technology, Editorial Board for Electronic Commerce Research and guest editor for several journals such as Information Technology for Development, Electronic Markets and Journal of Global Information Technology Management. Jian Mou is the corresponding author and can be contacted at: jian.mou@pusan.ac.kr For instructions on how to order reprints of this article, please visit our website: www.emeraldgrouppublishing.com/licensing/reprints.htm Or contact us for further details: permissions@emeraldinsight.com

Journal

Internet ResearchEmerald Publishing

Published: Mar 15, 2022

Keywords: Text-based chatbot; Trust; Consumers’ response; Task complexity; Identity disclosure

There are no references for this article.