Future Directions For Chatbot Research: An Interdisciplinary Research Agenda
Future Directions For Chatbot Research: An Interdisciplinary Research Agenda
Future Directions For Chatbot Research: An Interdisciplinary Research Agenda
https://doi.org/10.1007/s00607-021-01016-7
REGULAR PAPER
Received: 3 September 2020 / Accepted: 15 September 2021 / Published online: 19 October 2021
© The Author(s) 2021
Abstract
Chatbots are increasingly becoming important gateways to digital services and infor-
mation—taken up within domains such as customer service, health, education, and
work support. However, there is only limited knowledge concerning the impact of
chatbots at the individual, group, and societal level. Furthermore, a number of chal-
lenges remain to be resolved before the potential of chatbots can be fully realized.
In response, chatbots have emerged as a substantial research area in recent years. To
help advance knowledge in this emerging research area, we propose a research agenda
in the form of future directions and challenges to be addressed by chatbot research.
This proposal consolidates years of discussions at the CONVERSATIONS workshop
series on chatbot research. Following a deliberative research analysis process among
B Asbjørn Følstad
[email protected]
123
2916 A. Følstad et al.
the workshop participants, we explore future directions within six topics of interest: (a)
users and implications, (b) user experience and design, (c) frameworks and platforms,
(d) chatbots for collaboration, (e) democratizing chatbots, and (f) ethics and privacy.
For each of these topics, we provide a brief overview of the state of the art, discuss
key research challenges, and suggest promising directions for future research. The six
topics are detailed with a 5-year perspective in mind and are to be considered items
of an interdisciplinary research agenda produced collaboratively by avid researchers
in the field.
1 Introduction
123
Future directions for chatbot research… 2917
research agenda has the overall aim to motivate and guide research to establish req-
uisite knowledge for fully realizing the potential of chatbots as a powerful means of
accessing information and services and for understanding the impact of chatbots at the
individual, group, and societal level. As the research on chatbots is rapidly evolving,
we hold that deriving a research agenda from collaborations and discussions among
avid researchers and practitioners, who keep abreast of the ongoing developments of
the area, is a more effective approach as compared, for example, with a mapping study
or systematic literature review. Furthermore, this collaborative approach enables us
to gain insights from different perspectives to address opportunities, challenges, and
perceived research needs within the field. The research agenda serves as a concise
research roadmap, offering links to pertinent studies for those readers who are inter-
ested in delving further into specific fields.
In the following, we first present relevant background on chatbot research before we
detail the need for a consolidation of future directions. We then present our approach
and proposed set of directions. Finally, we discuss our proposal and the way forward.
2 Background
The emerging chatbot research area has its historical roots in several research fields
addressing different aspects of conversational computer systems—the most prominent
of these with decades of research and efforts at industrial applications. Within the field
of dialogue systems [77], researchers have since the sixties and seventies worked on
text based [12] and later spoken [59] conversational user interface to support users
with specific tasks. Other streams of research preceding and relevant to current chatbot
research have addressed conversational interaction with physical social robots [15],
and embodied virtual agents [18]. There has also been a long-term research initiative
addressing computer systems for open-domain small talk [98], including the develop-
ment of the artificial intelligence markup language [111] used to power chatbots for
social chit-chat. Conversational computer systems have also had a long and, at times,
winding path through various commercial applications—particularly automated solu-
tions for customer service, sales and support [72], including interactive voice response
(IVR) systems for phone-based self service [23].
The recent substantial increase in chatbot research can be seen as a direct response
to the uptake of so-called virtual assistants by big tech companies, specifically the
inclusion of Siri as part of Apple operating system in 2011, Amazon’s promotion
of Alexa since 2014 and the conversational turn of Facebook, Microsoft and Google
in 2016 [25]. Piccolo et al. [86] concluded that chatbot research has followed in the
trail of the industrial uptake of conversational computer systems rather than being at
the driver seat. In consequence, the contribution of this burgeoning research area is as
much to understand the emerging application, uses, and implications of conversational
computing systems, as to improve on their technological underpinnings and methods
for design and development. Consequently, the chatbot research area has a broader
scope and disciplinary coverage than the fields at its historical roots.
123
2918 A. Følstad et al.
As noted by McTear [77], research streams such as those of dialogue systems, embod-
ied conversational agents, and social robotics, are now converging in a common aim
for developing and improving on conversational user interfaces to computer systems.
However, there is still a wide variety of terms in use in reference to the object of this
converging research interest. Since the recent industrial uptake of conversational com-
puting systems, these have increasingly been referred to as chatbots within industry
and media [25] and also in research. To demarcate the research area driven by the
industrial uptake of conversational computer systems, and to signify the attention of
this area towards emerging patterns of use, as well as broader business and societal
implications, we refer to this area as chatbot research.
In line with this scoping of the research area, we understand chatbots as conver-
sational agents providing access to information and services through interaction in
everyday language—an understanding which is in line with the definitions by Føl-
stad and Brandtzaeg [39] and Hobert and Meyer von Wolff [54]. This use of the
term chatbot encompasses conversational agents for goal-oriented task completion,
informational purposes, entertainment, and social chatter. It also encompasses agents
supporting interactions through text, voice, or both. The use of the term is in reference
to the object of our research interest—current and future design, development, and
implications of information and services provided through conversational computer
systems—rather than in reference to a specific set of technologies or approaches.
In consequence, our use of the term chatbot is broader than what may be found
in other research streams. For example, some distinguish between voice-based and
text-based conversational agents, using the term chatbot to refer to the latter, (e.g.
Ashktorab et al. 6). Others distinguish between conversational agents for goal com-
pletion versus social chatter, referring only to the latter as chatbots (e.g. Jurafsky and
Martin 61). However, in consequence of the rapid evolvement both in technology, ser-
vices, and patterns of use, we find such attempts at principled scoping of the chatbot
term challenging. For example, there is often no clear distinction between social chat-
ter and goal-orientation in conversational agents—as seen by the importance of social
responses for customer service chatbots [114]. Likewise, the distinction between text
and voice is less than clear-cut as the same conversational agents may make use of
different modalities [97].
Current chatbots are enabled by a large range of technologies and services [97] at
varied levels of sophistication. Dialogue management may be enabled through simple
rule-based approaches, statistical data-driven systems, or neural generative end-to-end
approaches [77], and many systems employ hybrid models [50]. Whereas early chat-
bots for social chit-chat may exemplify rule-based approaches (e.g., Weizenbaum 112)
current statistical data-driven systems—such as chatbots for customer service—have
user intents and corresponding chatbot responses identified on the basis of training
of machine learning models based on example user data [66]. Generative chatbots
123
Future directions for chatbot research… 2919
Chatbot research is currently evolving within and across a range of disciplines and has
a strong interdisciplinary character. Ground-breaking research has been presented in
fields as diverse as communication (e.g. Go and Sundar 42), health (e.g. Fitzpatrick
et al. 35), informatics (e.g. Adiwardana et al. 2), and business (e.g. Adam et al. 1). While
dedicated workshops and conferences of relevance to chatbot research are emerg-
ing—such as CUI,6 CONVERSATIONS,7 and CAIR8 —in addition to established
venues—such as SIGDIAL,9 IVA,10 IWSDS,11 and INTERSPEECH12 —research
findings are typically presented in a broad range of journals and conferences. Research
related to chatbots is also conducted in multiple communities with varying degrees
of exchange among them. These communities may not label their area of interest as
chatbot research but, for example, research addressing conversational agents [79],
dialogue systems [59] or social robotics [93]. The research objectives within these
1 DialogFlow, https://cloud.google.com/dialogflow.
2 Microsoft Bot Framework, https://dev.botframework.com/.
3 Pandorabots, https://home.pandorabots.com/.
4 Rasa, https://rasa.com/.
5 Mycroft, https://mycroft.ai/.
6 CUI 2021—Conversational User Interfaces, https://www.conversationaluserinterfaces.org/2021/.
7 CONVERSATIONS 2021—international workshop on chatbot research, https://conversations2021.
wordpress.com/.
8 CAIR 2020—Conversational Approaches to Information Retrieval, https://sites.google.com/view/cair-
ws/cair-2020.
9 SIGDIAL—Special interest Group on Discourse and Dialogue, https://www.sigdial.org/.
10 IVA 2021—21st ACM International Conference on Intelligent Virtual Agents, https://sites.google.com/
view/iva2021/.
11 IWSDS 2021—12th International Workshop on Spoken Dialogue Systems Technology, https://www.
iwsds.tech/.
12 INTERSPEECH 2021—The 21st Annual Conference of the International Speech Communication Asso-
ciation, https://interspeech2021.org.
123
2920 A. Følstad et al.
4 Approach
The proposed future research directions are based on the collaborative work conducted
as part of the CONVERSATIONS workshops. CONVERSATIONS is an international
workshop series for chatbot research, where researchers, students, and practitioners
with interest in chatbots gather to present their work, discuss, and collaborate. The
first workshop in this series was organised in 2017 and it has since been a yearly event,
advancing from being arranged in conjunction with a research conference the two first
years to now being a 2-day stand-alone event. The most recent workshop in 2020
[41], arranged as a virtual event due to the COVID-19 pandemic, involved about 150
registered participants from more than 30 countries and 80 different organizations,
including more than 20 paper presentations. The participants represent disciplines
such as computer science, information systems, human–computer interaction, com-
munication studies, linguistics, psychology, marketing, and design.
123
Future directions for chatbot research… 2921
Through the CONVERSATIONS workshop series, six overarching topics for future
chatbot research have been identified. In the following, we detail each of these based
on the CONVERSATIONS output, with particular concern for the state of the art,
research challenges, and future research directions. An overview of the six topics and
associated future research directions is provided in Table 1.
Given the current evolving use and emerging use cases for chatbots, important
questions to ask concern chatbot users and their contexts of use. This includes inves-
tigating antecedents for chatbot use—namely individual characteristics, motivations
and boundary conditions for choosing, accepting or even preferring to interact with
conversational agents. Furthermore, it is necessary to explore and discuss implications
of chatbot use on individuals, groups, organizations and society at large.
123
2922 A. Følstad et al.
123
Future directions for chatbot research… 2923
may in part be due to the substantial impact on the level of organizations and society
is assumed to be seen in the future more so than the present.
While we have substantial knowledge on current chatbot users, important topics lack
sufficient coverage. Two warrant particular mention: (a) broader chatbot uses and user
groups and (b) implications of chatbot use, both detailed below.
For the broader chatbot uses and user groups, the rich literature needs to be con-
tinuously updated, especially when it comes to user motivations and behaviour of
emerging user groups. This includes knowledge on specific demographics, for exam-
ple, vulnerable users, such as children, elderly and users with special needs, as well
as user groups within particular application areas. Moreover, research still needs to
assess whether there are systematic differences in the adoption and usage of chatbots
driven by socio-demographic characteristics.
Implications of chatbot use entail a range of exciting research challenges, as knowl-
edge is needed on how the uptake of chatbots may impact groups, organizations,
businesses, and society at large. For example, as chatbots are taken up by different
sectors and industries, chatbots may transform service provision and work processes.
Another example is our need for knowledge on how the interaction patterns that
emerge between human users and chatbots may spill over to our interaction with
other people: Will the demanding communication style we learn to use for virtual
assistants, such as Alexa and Siri, impact our communication style with our partners
or collaborators? How will the companionship offered by social chatbots influence
users’ social lives and desires, and how chatbots may enter the social fabric of groups
or organizations?
Based on the current state of the art and identified research challenges, two future
research directions emerge as particularly promising in the area of chatbot user and
communication studies.
(a) Emerging chatbot user groups and behaviours. While there exists knowledge on
current chatbot user groups, this needs to be updated as technology, services,
and patterns of use evolve. Furthermore, there is a need to move from studies of
chatbot users in general to studies of chatbot users and behaviours for particular
demographics, domains, or contexts. We are beginning to see this for domains
such as health, education, and business, but given the uptake of chatbots in new
contexts and domains, this is an area of research which will be in continuous need
of update.
(b) Social implications of chatbots. The study of social implications of chatbots is
an area where we expect to see substantial research interest in the near future.
Knowledge of the social implications of chatbot use will be of importance to
guide also future development and design of chatbot services. Possibly, a string
of research on the broader social implications could be motivated from the broader
123
2924 A. Følstad et al.
discourse on implications of AI for labour and business (e.g. [37, 76]). It will be
beneficial to accommodate for research on unintended social consequences of
chatbots or how chatbots are shaped in response to its uptake in society.
Chatbot user experience and design concerns how users perceive and respond to chat-
bots, and how chatbot layout, interaction mechanisms and conversational content may
be designed so as to manage these perceptions and responses. To gather insight into
users’ perceptions and responses, and how these are impacted by chatbot design, user-
centred evaluations of chatbots is necessary; that is, assessments of users’ perceptions
and responses to chatbots conducted through established methods.
Chatbot user experience has been a key theme in recent research efforts, for voice-based
virtual agents [75] and text-based applications [4]. This has helped identify factors
contributing to positive or negative user experience [118] and has addressed specific
aspects such as trust [119], perceived social support [71], human likeness [4], and how
these aspects are impacted by chatbot design [42]. There is also a growing base of
research to inform design of chatbot interactions, whether this concerns conversational
design [6], personalization of chatbots [69], the use of interactive elements in chatbots
[57], or the use of social cues to indicate social status and capabilities [32]. Recently, a
number of textbooks (e.g. [48, 79, 97]) and industry guidelines (e.g. by Google13 and
Amazon14 ) have also been published on chatbot interaction design and conversational
design. Textual and acoustic properties of users’ dialogue input are gradually being
applied as outcomes in empirical research for studying engagement and experience
with conversational agents [52, 67]. Furthermore, there exists an extensive body of
research on emotion detection through speech (e.g. [95]) and non-verbal behaviour
[27] of high relevance to chatbot user experience and design.
There is also a grown body of knowledge on methods and measures for evaluat-
ing chatbot user experience. User-centred evaluation has been key to research within
several of the disciplines at the roots of current chatbot research, such as studies of
social presence in social robotics [82] and the use of user satisfaction measures in
evaluations of dialogue systems [28]. Evaluation in chatbot research is conducted by
instruments for users’ self reports of user experience [63], through user observation
and interviews [75] and analyses of chatbot interaction [66], and also by physiolog-
ical measurements [22]. A range of evaluation approaches are employed, including
experiments by self-administered online studies [5] or in the lab [22], observational
studies in the wild [64], and investigations of long-term interactions with established
services [73].
123
Future directions for chatbot research… 2925
While there is a growing body of research available on chatbot user experience there
still is a lack of knowledge on how to leverage the findings from this research in
chatbot designs that consistently delight and engage users. Users still experience issues
in chatbot interaction, both in terms of pragmatic experiences—where chatbots fail
to understand or to help users achieve their intended goals [75]—and in terms of
hedonic experiences—where chatbots fail to engage users over time [117]. These
issues may in part be seen as due to the more general challenge of designing human-
AI interaction [116]. There are indeed indications that these challenges are being
mitigated, for example in the case of improvements in customer service chatbots [80]
and in the uptake of social chatbots such as Replika [103]. However, the strengthening
of chatbot user experiences remains a key research challenge.
Related to the challenge of strengthening chatbot user experience, is the challenge
of measuring and assessing chatbots in terms of user experience and from a more
holistic perspective to determine whether chatbots are actually beneficial. Relevant
aspects for this are, for instance, usefulness, efficiency and process support. While
there is a large number of studies on chatbot user experience available, there is a
lack of common definitions, metrics and validated scales for key aspects of chatbot
evaluations [63]. Furthermore, while a broad range of approaches are employed there
is a lack of commonly applied approaches to evaluation.
Future research should be directed at addressing the identified key research challenges.
Specifically, the following two directions are proposed.
(a) Design for improving chatbot user experience. Future research on chatbot user
experience needs to evolve from exploring and assessing aspects of user experi-
ence and effects of chatbot design elements, towards studying how this knowledge
may impact and improve chatbot user experience in industrial applications.
Specifically, to translate findings of theoretical interest to conclusions of practi-
cal impact on design. This is not to say that research to build theory on chatbot
user experience is not needed, but this research may need to take up also more
design-oriented objectives—so as to condense current research and knowledge to
guidelines that may directly inform conversational design or interaction design.
(b) Modelling and evaluating chatbot user experience. To advance future research
on chatbot user experience, there is also a need for convergence of chatbot user
experience models, measurements, and approaches to evaluation. While diver-
sity in definitions and operationalizations is to be expected in an emerging field
of research interest, there may now be the time for seeking agreement and con-
sistency in the use of terminology and definitions of user experience constructs,
and also to identify and apply standardized measurements (benchmarks) for these
constructs. While such convergence should not be done in a way that hampers the-
oretical advancement and method innovation, there is clearly a benefit in including
common measurements across studies so as to enable cross-study comparison and
123
2926 A. Følstad et al.
aggregation, and to be able to track progress over time. For this purpose, estab-
lished evaluation approaches from fields such as human-computer interaction or
the tradition of dialogue systems may be beneficial.
This area of chatbot research concerns the current and future frameworks and plat-
forms for chatbot development and delivery. That is, the technological underpinnings
of chatbot implementations such as solutions for natural language processing, data
extraction, storage, and access, as well as mechanisms to identify and adapt chatbot
interactions to context and user profile.
The advances in chatbot frameworks and platforms are key enablers of the current
interest in chatbot applications. As noted in the background Sect. 6 myriad platforms
and frameworks are available to support design and development of chatbots. Key
advances include the application of supervised machine learning for classification
and information retrieval—enabling, for example, intent prediction and identifica-
tion of user sentiment [20], which are critical to support task-oriented conversations.
Furthermore, the use of generative approaches has seen substantial progress, where
end-to-end dialogue systems are applied to predict suitable responses to user input
based on models built from large conversational datasets [2, 109]. Finally, the intro-
duction of the Transformer [107] as a dominant and highly effective architecture for
natural language processing along with high-quality open-source libraries [113] have
lowered the barrier to entry and make it possible to build conversational models that
exhibit high generalization and coherence [90].
In this regard, large-scale generative models are becoming increasingly impactful,
enabling a wide range of tasks that can benefit chatbot development [36]. Models such
as GPT-3 [16] by OpenAI and BERT (Bidirectional Encoder Representations from
Transformers) [29] by Google leverage massive amounts of data and computational
power that would not be available to smaller players. Indeed, GTP-3 currently uses
175 billion parameters, and it is estimated to have cost 12 million US dollars to
train [36]. Thus, opening up these powerful models to the public has the potential
to accelerate chatbot development even further. It is important to note, however, that
criticism around large models has been growing lately [9], especially ethical concerns
regarding undesirable and often inscrutable societal biases percolating the models [9,
120], carbon footprint [9, 99], misuse and misinterpretation [9], privatization of AI
research [99], and even research opportunity costs [49].
While substantial advances have been made in chatbot frameworks and platforms, a
number of challenges remain. Specifically, we lack the needed technological underpin-
nings to support some key aspects of chatbot applications. We see four such challenges
123
Future directions for chatbot research… 2927
The area of chatbots for collaboration concerns how we may understand and design
chatbots in the context of networks that comprise humans and intelligent agents, for
example for social networking, teamwork, or service provision. While the current
123
2928 A. Følstad et al.
research on chatbots typically addresses dyadic interactions between one chatbot and
one user, we foresee that chatbots in collaborative relations involving more people and
bots will become more prominent as chatbots mature further. In addition, we consider
that collaborative relations can be addressed to a chatbot’s relations with external
online services in the form of application programming interfaces (APIs) and other
artificial agents.
Chatbots for collaboration concerns chatbots involved in interactions with humans and
possibly with other chatbots in networks larger than dyads. While not as prominent as
chatbots for simpler dyadic interaction, chatbots for collaboration have been developed
and implemented in a range of contexts and for various purposes, for example, to
support group processes in education [43], at work [11] and organizational settings
[104], as well as in gaming communities [96].
Types of collaboration with chatbots may include (a) one human collaborating with
one chatbot as an extension of human abilities, for example for analysis, gaming, as part
of a service-related inquiry, or as learning partner (e.g. [53]), (b) chatbots supporting
human collaboration, for example by taking notes, documenting, or task management
(e.g. [104]), and (c) chatbots collaborating with other services for example in multi-
agent models, networks of chatbots, or external web services (e.g. [108]).
Chatbots may be integrated into collaborative processes forming what Grudin and
Jacques [45] refer to as humbots, that is, human-chatbot teams which handle chal-
lenging service queries better than chatbots alone and more efficiently than humans
alone. The concept of humbots assumes a tiered approach to service provision where
the chatbots constitute an initial service contact point, and customers are escalated to
human helpers only if the chatbot is unable to help. Such human–chatbot teams draw
on the concept of human-in-the-loop [24] from the human factors literature, sensitiz-
ing system managers to the need for a collaborative setup allowing sufficient situation
awareness to the human part of the team to provide quality takeover if need be. In
health-care context, human-in-the-loop concepts for conversational agents supporting
hospital nurse teams has proved beneficial [13]. Likewise, the notion of escalation in
customer service chatbots is a practical application of the human-in-the-loop concept
for robust application of chatbots in consumer service provision [83].
123
Future directions for chatbot research… 2929
than an established actor within the described social structure (e.g. [104]). Or collab-
oration is addressed as merely a technical feature when the agent is collaborating with
other artificial agents and external web services (e.g., [108]).
While a range of chatbots for collaboration have been developed, there is relatively
scarce research on the characteristics of collaboration with chatbots. That is, we lack
models or theories helping us to conceptualize collaboration involving intelligent con-
versational agents. While this problem of human–machine collaboration is addressed
in more generic terms, for example in actor-network theory [70], there is a lack of
models to characterize conversational collaboration involving agents. Related to the
challenge of conceptualizing chatbot collaboration, there is a need for research on the
different roles chatbots and humans should take in the human–chatbot collaboration,
and what the implications of these may be. Should for example, the relation be based on
assistance or mutual collaboration? Should chatbot participation be reactive or active?
Should the chatbot be submissive or take charge? And what would the implications
of these choices be?
Drawing on the above state of the art and research challenges, the following research
directions are found to be particularly promising.
123
2930 A. Følstad et al.
123
Future directions for chatbot research… 2931
Recent studies suggest that while chatbots may indeed serve as a low-threshold inter-
face to information, services, and societal participation, they may also face challenges
regarding bias and inclusion. Besides, there is a lack of more systematic or structured
investigations of universal and inclusive design of chatbots. Inclusive and responsible
design of chatbots requires an understanding of various linguistic elements of conver-
sation and an awareness of broader social and contextual factors. For example, studies
are needed on barriers to onboarding and barriers to the use of chatbots. The aim of
using chatbots for strengthening democratization, reducing bias, and facilitating uni-
versal design has been included in the vision of chatbots for social good [40], which
may be a useful scope for addressing this set of challenges.
Furthermore, while available platforms and frameworks are promoted as low-
threshold means of chatbot design and development, there is a lack of knowledge
regarding how these are actually employed to democratize chatbot development and
design. Also, knowledge is needed on what challenges users with limited technol-
ogy skills meet when trying to use these platforms and frameworks, and how such
challenges may be overcome through changes, for example, in design and training of
machine learning models.
In light of the background and research challenges mentioned above, the following
broad directions of future research are identified.
(a) Chatbots for social good. To realize the potential of chatbots as vehicles for bridg-
ing digital divides and strengthening accessibility, availability, and affordability
of services and information, chatbots for social good may be leveraged as an
alternative perspective on chatbot research and design. In this perspective, sys-
tematic studies are needed to gain insight into current barriers in chatbot use how
these could be employed for social good. In this way, it will be possible to seek
to overcome the existing barriers with standardized solutions and follow user-
centered design processes focusing on user needs. Finally, research is needed on
the normative and ethical implications of the adoption of chatbots in particular
contexts, as also outlined in the next section.
(b) Inclusive design with and for diverse user groups. Parallel to the research direction
of chatbots for social good, we foresee research and development continuing the
work towards making the underlying platforms and frameworks for chatbot design
and development more easily applicable also for users without strong technical
skills. Here, we foresee studies of current opportunities and challenges faced by
chatbot creators, followed by development and design stages, aiming to follow up
or mitigate these. Removing the need for complex configuration and simplifying
or eliminating coding is probably the easiest way to serve the needs of the small
business and research groups—but also the needs of large enterprises that may
have domain experts creating chatbots. Furthermore, developing platforms that
facilitate the implementation of chatbots and recommend using best practices
during the design process will surely raise the quality level of the final products.
123
2932 A. Følstad et al.
The final research topic concerns ethical and privacy implications of chatbots. Specifi-
cally, how to reflect ethical and privacy concerns in the design of chatbots, recognising
the implications that different chatbot use cases and design choices may have for users’
trust in chatbots, and how we may identify and address unethical chatbot use.
AI has recently been the objective of substantial interest from policy-making and reg-
ulatory bodies, as well as in discussions and reflections on ethics, privacy management
and trust [21]. This concern for ethics in AI is motivated by its disruptive character
and potential for changes in the job market and misuse by malevolent actors, as well
as issues pertaining to accountability and bias [58]. Ethical concerns arising from
the design and deployment of AI technology have motivated a number of initiatives
[47]—such as the Ethical Guidelines for Trustworthy AI by the European Commission
expert group on AI, and Microsoft’s FATE: Fairness, Accountability, Transparency,
and Ethics in AI—addressing issues including mitigating bias and discrimination in AI
systems and fairness in the use of AI systems [81]. Chatbots are a prominent AI-based
technology, and as such in principle addressed by the broader concern for ethics and
privacy in technology research in general and in AI-based technology in particular.
Nevertheless, as noted in a review of the chatbot literature, there has been an initial lack
of ethical discussion in chatbot research [102]—though noteworthy exceptions to this
exist, such as the exploration of ethical and social considerations for conversational
AI by Ruane et al. [91]. The ethical discussion in chatbot research may, however, be
gaining traction motivated, for example, by Bender et al.’s [9] critical overview of
ethical risks pertaining to large language models.
The interest and discussion concerning ethics and privacy in AI have been partic-
ularly impactful in Europe, where the General Data Protection Regulation (GDPR) is
now used to govern privacy in technology-based systems and services. Furthermore,
based on the advice of a high-level expert group on AI, a European set of ethics
guidelines for trustworthy AI has been presented [30]. According to these guidelines,
it is of paramount importance for trustworthy AI to be aligned with (a) legal regu-
lations and (b) ethical principles and values, and also (c) be robust from a technical
perspective given its particular social context. From these principles, the European
Commission expert group has identified seven key requirements for ethical AI appli-
cations, including human agency and oversight, privacy and data governance, and
diversity, non-discrimination, and fairness. Finally, a proposed European set of reg-
ulations for AI, the AI Act, will help strengthen aspects of ethical concern in AI
systems, including legal requirements for human oversight, accuracy, robustness, and
security. Of particular relevance for chatbots is the proposed requirement for trans-
parency which will make it an obligation for service providers to ensure users are
aware when are interacting with machine agents and not human operators [94].
123
Future directions for chatbot research… 2933
Ethical and privacy challenges permeate the field of chatbot research, but specifically
where the context is sensitive or high-stakes or the users are marginalised or vulnerable;
for example, in designing chatbots for health and education, or in designing chatbots
to support asylum seekers or children. There is a large and growing body of ethical and
privacy knowledge to draw on, and an emerging set of guidelines and regulations on
ethics and privacy for digital systems in general, and AI-based systems in particular.
Nevertheless, we lack research and theorising around ethics and privacy specifically
for conversational user interfaces. This is problematic, as the conversational charac-
ter of chatbots may conceivably introduce a range of specific ethical problems, for
example the ethical implications of human-like and socially present chatbot interac-
tion, issues of consent, the privacy implications of third-party interactions and the
implications of emotional effect on children and vulnerable users. Research is needed
to better understand and address these, and other, emergent problems of ethics and
privacy.
Drawing on the above, we accentuate the following two directions for future
research—though other directions could be possible and maybe equally relevant.
(a) Understanding chatbot ethics and privacy. Future research should facilitate
reflections on ethical implication of chatbots, for example through identification
of ethical and privacy issues in chatbot design and implementation—includ-
ing design intentions, practical mitigation of known issues and exploration of
unforeseen implications. These could be domain-specific issues, such as ethical
implications for research and education, media or marketing and commerce, but
these could also be general issues such as how interaction with chatbots may
motivate oversharing in users, helping spread misinformation and hate speech, or
induce potential negative consequences as a result of over-humanizing chatbots.
(b) Ethics by design. In parallel with work on chatbot ethics, there will be a need for
research on the pragmatic and material issues of how to honour ethical guide-
lines and principles in the design of chatbot technology and applications. With
reference to the principle of privacy by design, we refer to this as ethics by
design—where privacy is subsumed as one of several aspects to consider as part
of an ethics discussion and subsequent design challenge. Important challenges
may include research on how to avoid biases in chatbots, how to avoid chatbot
discrimination and redlining, and how to mitigate the ethical issues introduced
by the black-box approach to machine learning underpinning aspects of chatbot
functioning, as well as to avoid misuse and weaponization of chatbot technology.
A useful starting point for an exploration of ethics by design, could be to refine
the generic European expert group requirements for ethical AI [30] to the context
of conversational AI.
123
2934 A. Følstad et al.
6 Discussion
Two of the identified research directions concern studies of users and implications, as
well as how to design for desirable chatbot use. As chatbots become more pervasive in
the coming years, and communication with non-human agents increasingly become
part of our daily routines, it becomes even more pressing to expand our knowledge
on the antecedents, contents and consequences of human–machine communication.
In doing so, this stream of research needs to explore the cognitive, affective and
behavioural dimensions of engagement with these agents, the extent to which there
are systematic differences between individuals, groups or contexts of use, and the
individual, group and societal implications of this phenomenon. Moreover, as the field
progresses, there is a growing need to consolidate the existing knowledge, updating
and extending overarching theoretical frameworks and models. Work within a wide
variety of disciplines can serve as an inspiration in that regard, such as the studies of
Sundar [100] on the psychology of human–agent interaction and Guzman and Lewis
[46] on human–machine communication.
This evolution in our understanding of conversational user experiences should be
accompanied with the proper support from platforms and frameworks. We can see the
support as increasingly creating abstractions that would facilitate the design, testing,
integration and development of chatbots, as it has historically happened with other
software artefacts. Current efforts are already moving in that direction, providing
development resources that promise anyone with enough motivation, regardless of
their background, to deliver human-like interactive experiences. While this has the
potential to bring substantial value to societies, empowering communities to develop
their own solutions, it can also bring unintended consequences, as we cannot expect
users of these platforms to have knowledge about complexities of modelling proper
Human-AI experiences [116]. On the other hand, abstractions can also hide underlying
information about machine learning models, AI decision-making, as well as latent bias
in the training data (e.g., [101]) that can translate into social biases (e.g., [120]).
Human–chatbot collaboration is foreseen as an increasingly important aspect of
chatbot research and applications. We hold that such collaboration will benefit from
being implemented while reflecting on human collaboration, and in line with relevant
empirical evidence of chatbot research—in line with reflections by Grudin and Jacques
[45]. Considering the meaningful value of collaboration for decision making and
productivity in professional and organizational settings, tasks assigned to chatbots in
these collaborative interactions can vary in complexity and involvement. Such tasks can
123
Future directions for chatbot research… 2935
123
2936 A. Følstad et al.
The identified research topics and corresponding future research directions may guide
research so as to contribute to the fundamental understanding of the chatbot technology
and the corresponding user interaction and engagement. However, to generate added
value in specific application areas—such as customer service, health, education, office
work, and home applications—further reflections about the respective use cases are
needed. In particular, researchers need to analyse how chatbots may be leveraged
and taken up in different application areas, how knowledge and research may be
transferable across different application areas, and whether distinct research agenda
should be established.
Many aspects of the outlined on our future research agenda are valid for any appli-
cation area. For example, results concerning chatbot communication, user experience,
design, and technology are the basis for applying chatbots in specific application
areas. However, further analysis efforts are needed to understand the characteristics
of each application area in more detail. For instance, requirements in the health sector
concerning privacy aspects, ethics, and trust may be significantly more demanding
than similar requirements in other sectors as they might have severe impacts on the
users and concern highly sensitive personal information. In business contexts such as
corporate customer support scenarios, the potential impacts may be less severe, but
specific corporate regulations and norms need to be considered. In contrast to that, the
use of chatbots in personal settings, e.g. the use of chatbots for social relationship, is
often mainly driven by motivations for engagement and meaning-making. In contrast
to health and business contexts, personal benefits are often not measured in mone-
tary terms, but the main focus of personal usage is the improvement of daily life or
wellbeing.
Regardless of differences among the diverse application areas in application-
oriented research, many research studies exist in specific domains that could possibly
be transferred to others. For instance, studies focusing on information provision in
business contexts can most likely be applied in the health sectors as well, e.g., pro-
vision of product information will likely be similar to explaining healthy nutrition.
However, to enable a transfer of research results across application areas, commonali-
ties and differences of the involved application areas need to be identified and assessed.
If the main characteristics of both are similar, transfer of the research results seems
viable. Based on such an analysis and comparison, a generalization of the research
across application areas seems possible. This procedure for future research on chatbot
application areas could lead to a substantial increase in the body of knowledge as
many research results from existing pilot studies and prototypes for specific applica-
tion areas exist and may be reused as the basis for transfer and generalization (i.e.,
general design guidelines) to further application areas.
The presented challenges may serve as a step in the direction of contributing to the
body of knowledge about chatbot usage and challenges, the frameworks and platforms
123
Future directions for chatbot research… 2937
underpinning chatbot applications, as well as needed future work on the broader impli-
cations of chatbots to work and society.
The proposed future research directions are intended as a response to the current
lack of coherence in the emerging field of chatbot research, which may in part be
observed by the broad range of journals and conferences in which findings from
chatbot research are presented, and also the lack of commonly agreed key constructs,
models, and measurement instruments. While this may be expected in an emerging
research area, future research will benefit from a greater degree of coherence and
cohesiveness in the field.
Nevertheless, there may be topics that have been omitted in the process leading up
to our proposition, and relevant state-of-the-art and current research challenges may
have been left out. Furthermore, as the field evolves, it is necessary to update the set of
topics and research directions regularly. In consequence, continued interdisciplinary
discussion and collaboration are needed to validate and refine the proposed set of
future research directions.
One limitation deserving particular mention concerns the context of this discussion.
The findings are based on discussions at the CONVERSATIONS workshop and mainly
involve researchers from European organizations. While we assume the proposed
directions hold broad international relevance and interest, it may be fruitful to test this
assumption through discussion in the field—a discussion which we hope this paper
will spur.
In further discussion and collaboration on chatbot research directions, care should
be taken to involve the broadest possible set of interests and perspectives. For example,
it will be beneficial to involve both researchers and practitioners, as well as the emerg-
ing and established set of research communities with an interest in conversational
computer systems, to make sure that different enabling technologies and knowledge
resources needed in future development and design of chatbots are represented. While
research on conversational systems and user interfaces has been conducted for decades,
chatbot research and design are still in its relative infancy. A consolidation of the field
is needed, and we hope the proposed research agenda, with its directions for future
research, may serve as a step towards such consolidation.
Funding Open access funding provided by SINTEF AS. Funding supporting the work conducted by the
first author was provided by Norges Forskningsråd (Grant No. 270940).
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give
appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence,
and indicate if changes were made. The images or other third party material in this article are included
in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If
material is not included in the article’s Creative Commons licence and your intended use is not permitted
by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the
copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
123
2938 A. Følstad et al.
References
1. Adam M, Wessel M, Benlian A (2020) AI-based chatbots in customer service and their effects on
user compliance. Electron Markets. https://doi.org/10.1007/s12525-020-00414-7
2. Adiwardana D, Luong MT, So DR, Hall J, Fiedel N, Thoppilan R, Le QV (2020) Towards a human-like
open-domain chatbot. arXiv preprint. arXiv:2001.09977
3. Amershi S, Weld D, Vorvoreanu M, Fourney A, Nushi B, Collisson P, Teevan J (2019) Guidelines for
human–AI interaction. In: Proceedings of the CHI 2019 (paper no. 3). ACM, New York
4. Araujo T (2018) Living up to the chatbot hype: the influence of anthropomorphic design cues and
communicative agency framing on conversational agent and company perceptions. Comput Hum
Behav 85:183–189
5. Araujo T (2020) Conversational agent research toolkit: an alternative for creating and managing
chatbots for experimental research. Comput Commun Res 2(1):35–51
6. Ashktorab Z, Jain M, Liao QV, Weisz JD (2019) Resilient chatbots: repair strategy preferences for
conversational breakdowns. In: Proceedings of CHI 2019 (paper no. 254). ACM, New York
7. Baez M, Daniel F, Casati F (2019) Conversational web interaction: proposal of a dialog-based natural
language interaction paradigm for the Web. In: Chatbot research and design. Third international
workshop, CONVERSATIONS. Springer, Cham, pp 94–11
8. Bavaresco R, Silveira D, Reis E, Barbosa J, Righi R, Costa C, Moreira C (2020) Conversational agents
in business: a systematic literature review and future research directions. Comput Sci Rev 36:100239.
https://doi.org/10.1016/j.cosrev.2020.100239
9. Bender EM, Gebru T, McMillan-Major A, Mitchell M (2021) On the dangers of stochastic parrots: can
language models be too big? In: Proceedings of the 2021 ACM conference on fairness, accountability,
and transparency. ACM, New York, pp 610–623
10. Bickmore T, Picard RW (2005) Establishing and maintaining long-term human–computer relation-
ships. ACM Trans Comput Hum Interact 12(2):293–327. https://doi.org/10.1145/1067860.1067867
11. Bittner E, Shoury O (2019) Designing automated facilitation for design thinking: a chatbot for support-
ing teams in the empathy map method. In: Proceedings of the 52nd Hawaii international conference
on system sciences. Scholar Space, Honolulu, pp 227–236
12. Bobrow DG, Kaplan RM, Kay M, Norman DA, Thompson H, Winograd T (1977) GUS, a frame-
driven dialog system. Artif Intell 8(2):155–173
13. Bott N, Wexler S, Drury L, Pollak C, Wang V, Scher K, Narducci S (2019) A protocol-driven, bedside
digital conversational agent to support nurse teams and mitigate risks of hospitalization in older adults:
case control pre-post study. J Med Internet Res 21(10):e13440
14. Brandtzaeg PB, Følstad A (2017) Why people use chatbots. In: Proceeedings of the international
conference on internet science—INSCI 2017. Springer, Cham, pp 377–392
15. Breazeal C (2003) Toward sociable robots. Robot Auton Syst 42(3–4):167–175
16. Brown TB, Mann B, Ryder N, Subbiah M, Kaplan J, Dhariwal P, Amodei D (2020) Language models
are few-shot learners. arXiv preprint. arXiv:2005.14165
17. Candello H, Pinhanez C, Pichiliani M, Cavalin P, Figueiredo F, Vasconcelos M, Carmo DH (2019)
The effect of audiences on the user experience with conversational interfaces in physical spaces. In:
Proceedings of CHI 2019 (paper no. 90). ACM, New York
18. Cassell J, Bickmore T, Billinghurst M, Campbell L, Chang K, Vilhjálmsson H, Yan H (1999) Embod-
iment in conversational interfaces: Rea. In: Proceedings of CHI ‘99. ACM, New York, pp 520–527
19. Catania F, Di Nardo N, Garzotto F, Occhiuto D (2019) Emoty: an emotionally sensitive conversational
agent for people with neurodevelopmental disorders. In Proceedings of the 52nd Hawaii international
conference on system sciences. Scholar Space, Honolulu, pp 2014–2023
20. Chen H, Liu X, Yin D, Tang J (2017) A survey on dialogue systems: recent advances and new frontiers.
ACM SIGKDD Explor Newsl 19(2):25–35
21. Chung H, Iorga M, Voas J, Lee S (2017) Alexa, can I trust you? Computer 50(9):100–104
22. Ciechanowski L, Przegalinska A, Magnuski M, Gloor P (2019) In the shades of the uncanny valley:
an experimental study of human–chatbot interaction. Future Gener Comput Syst 92:539–548
23. Corkrey R, Parkinson L (2002) Interactive voice response: review of studies 1989–2000. Behav Res
Methods Instruments Comput 34(3):342–353
24. Cranor LF (2008) A framework for reasoning about the human in the loop. In :UPSec’08: Proceedings
of the 1st conference on usability, psychology, and security. USENIX Association, Berkeley
25. Dale R (2016) The return of the chatbots. Nat Lang Eng 22(5):811–817
123
Future directions for chatbot research… 2939
26. Daniel G, Cabot J, Deruelle L, Derras M (2020) Xatkit: a multimodal low-code chatbot development
framework. IEEE Access 8:15332–15346
27. DeVault D, Artstein R, Benn G, Dey T, Fast E, Gainer A, Morency LP (2014) SimSensei Kiosk: a
virtual human interviewer for healthcare decision support. In Proceedings of the 2014 international
conference on autonomous agents and multi-agent systems, pp 1061–1068
28. Deriu J, Rodrigo A, Otegi A, Echegoyen G, Rosset S, Agirre E, Cieliebak M (2021) Survey on
evaluation methods for dialogue systems. Artif Intell Rev 54:755–810
29. Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: Pre-training of deep bidirectional transformers
for language understanding. arXiv preprint, arXiv:1810.04805
30. European Commission (2019) Ethics guidelines for trustworthy AI. Retrieved from: https://digital-
strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai
31. Farooq U, Grudin J (2016) Human–computer integration. Interactions 23(6):26–32
32. Feine J, Gnewuch U, Morana S, Maedche A (2019a) A taxonomy of social cues for conversational
agents. Int J Hum Comput Stud 132:138–161
33. Feine J, Gnewuch U, Morana S, Maedche A (2019b) Gender bias in chatbot design. In: Chatbot
research and design. Third international workshop CONVERSATIONS 2019. Springer, Cham, pp
79–93
34. Ferrara E (2020) What types of COVID-19 conspiracies are populated by Twitter bots? First Monday.
https://doi.org/10.5210/fm.v25i6.10633
35. Fitzpatrick KK, Darcy A, Vierhile M (2017) Delivering cognitive behavior therapy to young adults
with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a
randomized controlled trial. JMIR Mental Health 4(2):e19
36. Floridi L, Chiriatti M (2020) GPT-3: Its nature, scope, limits, and consequences. Mind Mach
30(4):681–694
37. Frey CB, Osborne MA (2017) The future of employment: How susceptible are jobs to computerisa-
tion? Technol Forecast Soc Chang 114:254–280
38. Følstad A, Brandtzæg PB (2017) Chatbots and the new world of HCI. Interactions 24(4):38–42
39. Følstad A, Brandtzaeg PB (2020) Users’ experiences with chatbots: findings from a questionnaire
study. Quality User Exp. https://doi.org/10.1007/s41233-020-00033-2
40. Følstad A, Brandtzaeg PB, Feltwell T, Law ELC, Tscheligi M, Luger E (2018) Chatbots for social
good. In: Extended abstracts of the 2018 CHI conference on human factors in computing systems
(paper no. SIG06). ACM, New York
41. Følstad A, Araujo T, Papadopoulos S, Law EL-C, Luger E, Goodwin M, Brandtzaeg PB (eds) (2020)
Chatbot research and design 4th international workshop, CONVERSATIONS 2020. Springer, Cham
42. Go E, Sundar SS (2019) Humanizing chatbots: the effects of visual, identity and conversational cues
on humanness perceptions. Comput Hum Behav 97:304–316
43. Goel AK, Polepeddi L (2016) Jill Watson: A virtual teaching assistant for online education. Georgia
Institute of Technology. Retrieved from https://smartech.gatech.edu/handle/1853/59104
44. Gorwa R, Guilbeault D (2020) Unpacking the social media bot: a typology to guide research and
policy. Policy Internet 12(2):225–248
45. Grudin J, Jacques R (2019) Chatbots, humbots, and the quest for artificial general intelligence. In:
Proceedings of CHI 2019 (paper no. 209). ACM, New York
46. Guzman AL, Lewis SC (2019) Artificial intelligence and communication: a human–machine com-
munication research agenda. New Media Soc 22(1):70–86
47. Hagendorff T (2020) The ethics of AI ethics: An evaluation of guidelines. Mind Mach 30(1):99–120
48. Hall E (2018) Conversational Design. A Book Apart, New York
49. Hao K (2020) We read the paper that forced Timnit Gebru out of Google. Here’s what it says.
MIT Technology Review. Retrieved from https://www.technologyreview.com/2020/12/04/1013294/
google-ai-ethics-research-paper-forced-out-timnit-gebru/
50. Harms JG, Kucherbaev P, Bozzon A, Houben GJ (2018) Approaches for dialog management in
conversational agents. IEEE Internet Comput 23(2):13–22
51. Harwell D (2018) The accent gap. The Washington Post. Retrieved from https://www.washingtonpost.
com/graphics/2018/business/alexa-does-not-understand-your-accent/
52. Ho A, Hancock J, Miner AS (2018) Psychological, relational, and emotional effects of self-disclosure
after conversations with a chatbot. J Commun 68(4):712–733
123
2940 A. Følstad et al.
53. Hobert S, Berens F (2019) Small talk conversations and the long-term use of chatbots in educational
settings–experiences from a field study. In: Chatbot research and design. Third international workshop
CONVERSATIONS 2019. Springer, Cham, pp 260–272
54. Hobert S, Meyer von Wolff, R (2019) Say hello to your new automated tutor—a structured literature
review on pedagogical conversational agents. In: Proceedings of the 14th international conference on
Wirtschaftsinformatik. Association for Information Systems, pp.301–313
55. Hoffman G, Breazeal C (2004) Collaboration in human–robot teams. In: AIAA 1st intelligent systems
technical conference. https://doi.org/10.2514/6.2004-6434
56. Hsieh T, Chaudhury B, Cross ES (2020) Human–robot cooperation in prisoner dilemma games: people
behave more reciprocally than prosocially toward robots. In: Companion of the 2020 ACM/IEEE
international conference on human–robot interaction. ACM, New York, pp 257–259
57. Jain M, Kumar P, Kota R, Patel SN (2018) Evaluating and informing the design of chatbots. In:
Proceedings of the 2018 designing interactive systems conference. ACM, New York, pp 895–906
58. McTear MF (2002) Spoken dialogue technology: enabling the conversational user interface. ACM
Comput Surv (CSUR) 34(1):90–169
59. Jokinen K, McTear M (2009) Spoken dialogue systems. Synthesis lectures on human language tech-
nologies. Morgan & Claypool Publishers, Williston
60. Jovanovic M, Baez M, Casati F (2020) Chatbots as conversational healthcare services. IEEE Internet
Comput. https://doi.org/10.1109/MIC.2020.3037151
61. Jurafsky D, Martin JH (2020) Dialogue systems and chatbots. In: Speech and language processing.
Draft of December 30, 2020. https://web.stanford.edu/~jurafsky/slp3/24.pdf
62. Kim S, Lee J, Gweon G (2019) Comparing data from chatbot and web surveys: effects of platform
and conversational style on survey response quality. In: Proceedings of CHI 2019 (paper no. 86).
ACM, New York
63. Kocaballi AB, Laranjo L, Coiera E (2019) Understanding and measuring user experience in conver-
sational interfaces. Interact Comput 31(2):192–207
64. Kopp S, Gesellensetter L, Krämer NC, Wachsmuth I (2005) A conversational agent as museum
guide–design and evaluation of a real-world application. In: International workshop on intelligent
virtual agents. Springer, Heidelberg, pp 329–343
65. Kretzschmar K, Tyroll H, Pavarini G, Manzini A, Singh I (2019) Can your phone be your therapist?
Young people’s ethical perspectives on the use of fully automated conversational agents (chatbots) in
mental health support. Biomed Inform Insights. https://doi.org/10.1177/1178222619829083
66. Kvale K, Sell OA, Hodnebrog S, Følstad A (2019) Improving conversations: lessons learnt from
manual analysis of chatbot dialogues. In: Chatbot research and design. Third international workshop
CONVERSATIONS 2019. Springer, Cham, pp 187–200
67. Laban G, George J, Morrison V, Cross ES (2021) Tell me more! Assessing interactions with social
robots from speech. Paladyn J Behav Robot 12(1):136–159
68. Laban G, Araujo T (2019) Working together with conversational agents: the relationship of perceived
cooperation with service performance evaluations. In: Chatbot research and design. Third international
workshop CONVERSATIONS 2019. Springer, Cham, pp 215–228
69. Laban G, Araujo T (2020) The effect of personalization techniques in users’ perceptions of conver-
sational recommender system. In: IVA ’20: Proceedings of the 20th ACM international conference
on intelligent virtual agents (paper no. 34). ACM, New York
70. Latour B (2005) Reassembling the Social - An introduction to actor-network-theory. Oxford Univer-
sity Press, Oxford, UK
71. Lee S, Choi J (2017) Enhancing user experience with conversational agent for movie recommendation:
effects of self-disclosure and reciprocity. Int J Hum Comput Stud 103:95–105
72. Lester J, Branting K, Mott B (2004) Conversational agents. In: Singh MP (ed) The Practical Handbook
of Internet Computing. Chapman & Hall/CRC, Boca Raton
73. Liao QV, Geyer W, Muller M, Khazaen Y (2020) Conversational interfaces for information search.
Understanding and improving information search. Springer, Cham, pp 267–287
74. Liao QV, Mas-ud Hussain M, Chandar P, Davis M, Khazaeni Y, Crasso MP, Geyer W (2018) All
work and no play? In Proceedings of CHI 2018 (paper no. 3). ACM,New York
75. Luger E, Sellen A (2016) “Like having a really bad PA”—the gulf between user expectation and
experience of conversational agents. In: Proceedings of CHI 2016. ACM, New York, pp 5286–5297
76. McAfee A, Brynjolfsson E (2017) Machine, platform, crowd: harnessing our digital future. W.W.
Norton & Company, New York
123
Future directions for chatbot research… 2941
77. McTear M (2021) Conversational AI: Dialogue Systems, Conversational Agents, and Chatbots. Mor-
gan & Claypool, Williston
78. Meyer von Wolff R, Hobert S, Masuch K, Schumann M (2020) Chatbots at digital workplaces - A
grounded-theory approach for surveying application areas and objectives. Pac Asia J Assoc Inf Syst
12(2):64–102
79. Moore RJ, Arar R (2019) Conversational UX design: a practitioner’s guide to the natural conversation
framework. Morgan & Claypool, Williston
80. Nordheim CB, Følstad A, Bjørkli CA (2019) An initial model of trust in chatbots for customer
service—findings from a questionnaire study. Interact Comput 31(3):317–335
81. Ntoutsi E, Fafalios P, Gadiraju U, Iosifidis V, Nejdl W, Vidal ME, Ruggieri S, Turini F, Papadopou-
los S, Krasanakis E, Kompatsiaris I (2020) Bias in data-driven artificial intelligence systems—an
introductory survey. Wiley Interdiscipl Rev Data Min Knowl Discov 10(3):e1356
82. Oh CS, Bailenson JN, Welch GF (2018) A systematic review of social presence: definition,
antecedents, and implications. Front Robot AI 5:114
83. Paikens P, Znotiņš A, Bārzdiņš G (2020) Human-in-the-loop conversation agent for customer service.
In: Proceedings of the 25th international conference on applications of natural language to information
systems, NLDB 2020. Springer, Cham, pp 277–284
84. Pérez JQ, Daradoumis T, Puig JMM (2020) Rediscovering the use of chatbots in education: a sys-
tematic literature review. Comput Appl Eng Educ 28(6):1549–1565
85. Perski O, Crane D, Beard E, Brown J (2019) Does the addition of a supportive chatbot pro-
mote user engagement with a smoking cessation app? An experimental study. Digital Health
5:2055207619880676
86. Piccolo LS, Mensio M, Alani H (2018) Chasing the chatbots. In Internet Science. INSCI 2018
international workshops. Springer, Cham, pp 157–169
87. Porcheron M, Fischer JE, Reeves S, Sharples S (2018) Voice interfaces in everyday life. In: Proceed-
ings of CHI 2018 (paper no. 640). ACM,New York
88. Press G (2019) AI Stats News: 62 % of US Consumers like using chatbots to interact with busi-
nesses. Forbes. Retrieved from https://www.forbes.com/sites/gilpress/2019/10/25/ai-stats-news-us-
consumers-interest-in-using-chatbots-to-interact-with-businesses-rise-to-62/
89. Rodríguez-Gil L, García-Zubia J, Orduña P, Villar-Martinez A, López-De-Ipiña D (2019) New
approach for conversational agent definition by non-programmers: a visual domain-specific language.
IEEE Access 7:5262–5276
90. Roller S, Dinan E, Goyal N, Ju D, Williamson M, Liu Y, Boureau YL (2020) Recipes for building an
open-domain chatbot. arXiv preprint. arXiv:2004.13637
91. Ruane E, Birhane A, Ventresque A (2019) Conversational AI: social and ethical considerations. In:
Proceedings for the 27th AIAI Irish conference on artificial intelligence and cognitive science—AICS
2019, pp 104–115. CEUR Workshop Proceedings, Vol-2563. http://ceur-ws.org/Vol-2563/
92. Rubin VL, Chen Y, Thorimbert LM (2010) Artificially intelligent conversational agents in libraries.
Library Hi Tech 28(4):496–522
93. Salem M, Kopp S, Wachsmuth I, Rohlfing K, Joublin F (2012) Generation and evaluation of com-
municative robot gesture. Int J Social Robot 4(2):201–217
94. Schaake M (2021). The European commission’s artificial intelligence act. Policy Brief, Stanford HAI.
Retrieved from https://hai.stanford.edu/sites/default/files/2021-06/HAI_Issue-Brief_The-European-
Commissions-Artificial-Intelligence-Act.pdf
95. Schuller BW (2018) Speech emotion recognition: two decades in a nutshell, benchmarks, and ongoing
trends. Commun ACM 61(5):90–99
96. Seering J, Luria M, Ye C, Kaufman G, Hammer J (2020) It takes a village: integrating an adaptive
chatbot into an online gaming community. In: Proceedings of CHI 2020 (paper no. 579). ACM,New
York
97. Shevat A (2017) Designing bots: creating conversational experiences. O’Reilly Media, Boston
98. Shum HY, He XD, Li D (2018) From Eliza to XiaoIce: challenges and opportunities with social
chatbots. Front Inf Technol Electron Eng 19(1):10–26
99. Strubell E, Ganesh A, McCallum A (2019) Energy and policy considerations for deep learning in
NLP. arXiv preprint, arXiv:1906.02243
100. Sundar SS (2020) Rise of machine agency: a framework for studying the psychology of Human–AI
Interaction (HAII). J Comput Mediat Commun 25(1):74–88
123
2942 A. Følstad et al.
101. Suresh H, Guttag JV (2019) A framework for understanding unintended consequences of machine
learning. arXiv preprint, arXiv:1901.10002
102. Syvänen S, Valentini C (2020) Conversational agents in online organization–stakeholder interactions:
a state-of-the-art analysis and implications for further research. J Commun Manag 339–362
103. Ta V, Griffith C, Boatfield C, Wang X, Civitello M, Bader H, Loggarakis A (2020) User experiences
of social support from companion chatbots in everyday contexts: thematic analysis. J Med Internet
Res 22(3):e16235
104. Toxtli C, Monroy-Hernández A, Cranshaw J (2018) Understanding chatbot-mediated task manage-
ment. In: Proceedings of CHI 2018 (paper no. 58). ACM,New York
105. Vaidyam AN, Wisniewski H, Halamka JD, Kashavan MS, Torous JB (2019) Chatbots and conversa-
tional agents in mental health: a review of the psychiatric landscape. Can J Psychiatry 64(7):456–464.
https://doi.org/10.1177/0706743719828977
106. Van der Goot MJ, Pilgrim T (2019) Exploring age differences in motivations for and acceptance
of chatbot communication in a customer service context. In: Chatbot research and design. Third
international workshop CONVERSATIONS 2019. Springer, Cham, pp 173–186
107. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017)
Attention is all you need. In: Conference on neural information processing systems—NIPS 2017, pp
5998–6008
108. Vaziri M, Mandel L, Shinnar A, Siméon J, Hirzel M (2017) Generating chat bots from web API
specifications. In: Proceedings of the 2017 ACM SIGPLAN international symposium on new ideas,
new Pparadigm, and reflections on programming and software. ACM, New York, pp 44–57
109. Vinyals O, Le Q (2015) A neural conversational model. arXiv preprint, arXiv:1506.05869
110. Väänänen K, Hiltunen A, Varsaluoma J, Pietilä I (2019) CivicBots–Chatbots for supporting youth in
societal participation. In: Chatbot research and design. Third international workshop CONVERSA-
TIONS 2019. Springer, Cham, pp 143–157
111. Wallace R (2003) The elements of AIML style. Alice AI Foundation, 139
112. Weizenbaum J (1967) Contextual understanding by computers. Commun ACM 10(8):474–480
113. Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, Cistac P, Rault T, Louf R, Funtowicz
M, Brew J (2019) HuggingFace’s transformers: state-of-the-art natural language processing. arXiv
preprint, arxiv:1910.03771
114. Xu A, Liu Z, Guo Y, Sinha V, Akkiraju R (2017) A new chatbot for customer service on social media.
In: Proceedings of CHI 2017. ACM, New York, pp 3506–3510
115. Yaghoub-Zadeh-Fard MA, Zamanirad S, Benatallah B, Casati F (2020) REST2Bot: bridging the gap
between bot platforms and REST APIs. In: Companion proceedings of the web conference 2020.
ACM, New York, pp 245–248
116. Yang Q, Steinfeld A, Rosé C, Zimmerman J (2020) Re-examining whether, why, and how Human-AI
Interaction is uniquely difficult to design. In: Proceedings of CHI 2020 (paper no. 174). ACM,New
York
117. Zamora J (2017) I’m sorry, Dave, I’m afraid I can’t do that: Chatbot perception and expectations.
In: Proceedings of the 5th international conference on human agent Interaction. ACM,New York,pp
253–260
118. Zarouali B, Van den Broeck E, Walrave M, Poels K (2018) Predicting consumer responses to a chatbot
on Facebook. Cyberpsychol Behav Soc Netw 21(8):491–497
119. Zierau N, Engel C, Söllner M, Leimeister JM (2020) Trust in smart personal assistants: A systematic
literature review and development of a research agenda. In: Proceedings of the 15th international
conference on Wirtschaftsinformatik—WI2020. Association for Information Systems,pp 99–11
120. Zou J, Schiebinger L (2018) AI can be sexist and racist—it’s time to make it fair. Nature 559:324–326
Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.
123