Introduction
Human-robot trust determines human acceptance and hence the willingness to collaborate with robots [1]. Research has shown that trust is not only malleable but a critical determinant of human acceptance of technology and eventual performance [2]. Based on the modified IBM technology acceptance model (TAM), perceived usefulness and perceived ease of use have positive impacts on human trust, which influences the behavioral intention and serves as an indicator of the actual system use (see Fig. 1 for an illustration). There is an inverse correlation between trust and monitoring of automation [3]. Work can be delegated to robotic systems only if they can be trusted without constant monitoring. Improper/non-calibrated trust leads to misuse (e.g., acceptance of all of the robot's operations without questioning), under-utilization (e.g., not assigning a robot tasks that it is capable of completing), or disuse (e.g., not utilizing the robot at all) [4], [5], [6]. More specifically, overtrust leads to misuse and hence system performance degradation, while under trust leads to under-utilization/disuse and hence human stress and cognitive overload. Fig. 2 illustrates the relationship between trust, trustworthiness, and robot usage.
Despite the increasing attention in human-robot trust research [7], [8], social studies on interpersonal trust [9], [10] and social-network based trust [11], [12], [13], and human-machine trust research over the past four decades [3], [4], [5], there still lacks a good understanding on computational models of human's trust in robots. Many extant works remain at a descriptive and qualitative level since the trust between humans and robots is latent, fuzzy, and complex to measure and quantify. Human-robot trust model has yet to be elucidated from a computational perspective to capture its temporal, dynamic, and uncertain nature. The quantification of human trust when interacting with robots may be utilized to design robot motion planning and control algorithms, predict and optimize the dynamic allocation of autonomous functions, provide objective metrics for the selection of levels of autonomy (LoAs), and hence is of critical importance to the development of human-robot interaction (HRI) and human-robot teaming strategies [14]. The purpose of this paper is therefore threefold:
Provide an overview of human trust in robots, especially a survey of most representative quantitative methods and computational models of human trust in robots;
Investigate the utilization of computational human trust models in control and/or robotics applications;
Discuss limitations, open questions, and gaps in the extant literature, and explore future research directions in the field.
Different from other recent surveys [15], [16], [17], we focus on human-robot trust models and will provide a clear classification of different trust models, a detailed introduction of the trust formulations, and compare their pros and cons. Furthermore, we will emphasize the control and robotics applications of these trust models, seeking to explore their utilization and potential contributions to control system technologies.
Our working definition of robot in this paper is: “A robot is an autonomous machine capable of sensing its environment, carrying out computations to make decisions, and performing actions in the real world” [18]. Compared to a machine, a robot can sense, think (in a deliberative, non-mechanical sense), and act [19]. We refer to robots in general, which may include ground/aerial/underwater mobile robots, fixed/mobile factory robots, self-driving cars, medical robots, humanoids, autonomous decision aids, etc.
Two types of performance/competency-related trust have been identified in the literature to capture the belief in automation/robot's ability to achieve an individual's goal [2], [20]: dispositional/initial trust, and history-based trust. Dispositional trust describes an individual's tendency to trust or distrust even without any interaction with the automation/robot. It is the inherent psychological trait of an individual, and hence unchanging. Dispositional trust is generally predictive of an individual's learned trust in that it is the starting point for an individual's interpersonal interactions with others. In contrast, history-based trust is a construct that dynamically adjusts cumulative interactions based on the external environment and context-dependent characteristics of the operator, which is the core type of trust in the computational trust models considered in this paper. Humans can either gain or lose trust based on the progress of the task [21]. Besides the well-acknowledged performance/competency-based trust, several studies [22], [23] also suggested that the intention-based trust should also be included in the human-robot trust framework. The former indicates that the human trusts that the robot is capable of performing a task, while the latter indicates the human trusts that the robot has good intentions to act for his/her benefit. Although intention-based trust typically occurs among human teams, the human anthropomorphism of robots makes it an unignorable component in the human-to-robot team.
Trust is a continuum spanning from distrust to trust [24], rather than a binary discrete variable of being either trustful or untrustful [24]. While dispositional trust may be static, history-based trust has a dynamic nature [2], [9], [25] (see Fig. 3). Although factors that influence dispositional trust (e.g., propensity to trust) may be static, they still impact human decision-making during interactions. Trust has inertia and initial bias [26] and it is easy to break but hard to recover [27]. Trust is individual specific [10]. While generalized trust models for a population provide guidelines that may be suitable for an average person, individualized trust models are also important for more accurate evaluation and prediction in human-robot teaming and collaboration [28]. Trust is domain (task-, situation-, context-) specific [10]. For example, it is not the same to trust an autonomous car on the road than to a factory manufacturing robot in assembly tasks [29]. Furthermore, trust is not completely transitive, i.e., “
\begin{equation*}
T(x,y)>T(y,w)\,{\not\Rightarrow {}}T(x,y)>T(x,w).
\end{equation*}
In Desai's thesis [32], trust models for machines and automation were classified into five categories [33]: regression models [34], [35], time-series models [3], [4], [5], [26], [36], [37], [38], neural net models [39], qualitative models [33], [40], and argument-based probabilistic models [33], [41]. Among these models, the time-series, regression, and neural net models are quantitative. The regression models establish quantitative relations between independent variables (trust impacting factors) and the dependent variable (trust) but did not capture the dynamic variance in trust development. The neural nets model discovers the relationship between inputs (trust impacting factors) and output (trust). However, they rely on large datasets collected with human subjects and can not reveal the inner mechanism of trust due to the nature of neural networks. The time-series models characterize the dynamic relationship between human-to-robot trust and the independent variables, which require prior knowledge of the factors that impact human trust. Nevertheless, these are statistical models that fit the data without consideration of human psychology [26].
A lot of computational models have also been developed to evaluate trust in multi-agent social networks [11], [12], which can refer to the trust relationship among a society of people, robots, the Semantic Web, a recommendation system, an electronic commerce site (e.g., eBay, Amazon), etc. Kuter and Golbeck introduced a probabilistic network model for social networks and developed a trust inference algorithm called SUNNY that uses the trust inference procedure from TidalTrust [11] based on information sources with the highest confidence estimates [12]. Avesani et al. developed another algorithm to infer trust values in networks for a real-world application, namely moleskiing.it [42]. Pinyol and Sabater-Mir reviewed the most representative computational models for trust and reputation in multi-agent systems [29]. Reputation is regarded as a collective measure of trustworthiness therein [43]. Both cognitive models built based on beliefs, desires, and intentions, and numerical models based on game theory were introduced. Similar review papers can be found in [24], [43], [44]. In Mikulski's report [45], six categories of computational trust models for networks were summarized: scaled value models [46], multi-faceted representation models [47], [48], logical metric models [49], [50], [51], direct trust models [52], [53], recommendation trust models [54], [55], [56], [57], [58], and hybrid trust models [59], [60], [61], [62]. These works computed objective trustworthiness (in terms of reliability, utility, and risk) as a property of each agent in a network instead of an individual's actual emotion and cognition when interacting with networked multi-agent systems. These models are normative, which model the most ideal situations that might not often be found in practice. Their main utilization has been in economics and software development without the consideration of sociology and psychology nor verification with human subject test data.
Furthermore, none of the above models were personalized to each individual or customized to a specific scenario. There was also no discussion regarding leveraging trust models in control and/or robotics applications. In this paper, we seek to provide a review of works that attempt to address such problems. Furthermore, instead of reviewing all work for trust in machines/automation or general agent-agent, we will focus on computational models for human-robot trust and their control/robotics applications. As this field is still in its early stage, we hope our work may help guide future developments in human-robot trust models and steer research towards trust-aware robotic control for improved HRI.
The organization of the rest of the paper is as follows. Section II discusses the definitions and differences in interpersonal trust, human trust in machines/automation, and human trust in robots. We also summarize impacting factors of human-robot trust, existing trust measurement approaches, and corresponding scales. Section III surveys the state-of-the-art computational trust models, together with their utilization in control and robotics applications if any. We also discuss the merits and shortcomings of each type of trust model. Section IV discusses the limitations of the current trust modeling work and future directions in human-robot trust. Section V concludes the paper.
Human Trust in Robots: Definition, Impacting Factors, & Measurements
A. Definitions of Human-Robot Trust
Cho et al.’s survey on trust modeling [24] summarized the concept of trust based on the common themes across disciplines (Sociology, Philosophy, Economics, Psychology, Organizational Management, International Relations, Automation, Neurology, and Computing & Networking) as “the willingness of the trustor (evaluator) to take risk based on a subjective belief that a trustee (evaluatee) will exhibit reliable behavior to maximize the trustor's interest under uncertainty (e.g. ambiguity due to conflicting evidence and/or ignorance caused by complete lack of evidence) of a given situation based on the cognitive assessment of past experience with the trustee.”
Mayer et al. defined interpersonal/organizational trust as the “willingness of a party to be vulnerable to the actions of another party, based on the expectation that the other will perform a particular action important to the truster, irrespective of the ability to monitor or control the other party” [10]. The authors indicated that this definition of trust applies to a relationship between a trustor and another identifiable party perceived to act and react with volition toward the trustor. There are also many other definitions for interpersonal trust and they are not unified. For years the definition of trust proposed by Mayer et al. was the most accepted and used one [5]. Similarly, Lee and See defined trust in automation or another person as “the attitude that an agent will help achieve an individual's goals in a situation characterized by uncertainty and vulnerability” [5]. This definition has been adopted by many human automation trust studies, and is considered the most thorough [15]. However, Rempel et al. argued that the interpersonal trust development process starts from observation-based predictability, followed by dependability, then to faith in the relationship [63]. While human trust in automation/machines tends to develop in an opposite way, which is initiated from the machine function description (faith) and then refined by the user experience (dependability) and observation (predictability) [3]. Furthermore, as Ekman et al. pointed out in [64], people tend to consider human-human interaction more in terms of trust than distrust while the order in human-machine is different.
As Scott et al. pointed out, human-to-robot trust shares similarity with human-to-human trust as humans tend to anthropomorphize robots [23]. Johnson-George and Swap pointed out “Willingness to take risks may be one of the few characteristics common to all trust situations,” and the possibility to fail is necessary for trust to be present [64]. Nevertheless, there are still challenges to extrapolating the concept of trust in people to trust in robots. We believe that human trust in robots differs from interpersonal trust, provided that robots as trustees differ from humans in ability, integrity, and/or benevolence. When considering human-to-human trust, the purpose, reliance, and transparency of the human is of great importance [70]. However, Lee and Moray's study [36] reveals that the performance and process of automation are the most important for humans to form trust. In contrast, the purpose of automation has been presupposed as they are designed to help humans and produce positive outcomes [71]. This is one of the differences between human-to-human trust and human-to-automation trust. We believe the same applies to human-to-robot trust.
Furthermore, human trust in robots is more complicated than trust in agents, machines, or automation [72]. For instance, humans may not interact physically with agents, which are otherwise more explicit in automation/machines and robots. Robots are mobile, while automation/machines are traditionally fixed-base and static. Robots are required to perform autonomous tasks and make human-like decisions in uncertain and complex environments, while automation operation conditions are relatively simple, predictable, and constrained. Robots can have self-governance and learning capabilities to adapt to new tasks, while automation is pre-programmed to perform a set of pre-defined actions [73]. Robots collaborate with humans more like a teammate than a tool. Furthermore, the types of interactions, the number of systems interacting with the human, the designated social and operative functions, and the environment all differentiate HRI from human-computer or human-machine interaction [74], [75].
Remark 1:
In the literature, the overall trust is considered as the combination of subjective and objective trust [24]. But it is more about the perceived trustworthiness of the trustee [14], which means its value can be affected by the trustor (human) factors. In contrast, trustworthiness is considered an intrinsic property of the trustee. In other words, the value of trustworthiness cannot be changed with respect to different trustors. It is objective and needs to be verified through observed evidence. There is no uniform agreement on the compositions of trustworthiness. It is also not rare to find that reliability and trustworthiness are interchangeable in some literature [76]. Besides received referrals and personal experience, the emotional aspects of the human also contribute to the distinctions between human trust in robot and the robot trustworthiness [43]. The distinctions between trust and trustworthiness are further categorized into the “over-trust” situation when the subjective trust is greater than the trustworthiness and the “under-trust” situation when the trust value is below the trustworthiness [77] (see Fig. 2). Intuitively, trustworthiness can be used as a good objective reference to judge human trust decisions.
Not surprisingly, there are many definitions of human trust in robots in the literature, and there is a lack of consensus. Schaefer et al. provided a list of 37 definitions for human trust in automation to get implications for HRI [78]. They suggest that this diversity in definitions of trust, as seen in other domains, supports the lack of a universal definition of trust. Table 1 lists the definitions of human-robot trust we found in the literature. These papers were published from 2007 to 2021. Among these works, the definition by Hancock et al. [22] seems yet one of the most referenced definitions of trust in robots. With the above studies, we may conclude that human-robot trust differs from interpersonal trust, human-agent trust, and human-machine/automation trust, mainly due to their different contexts, properties, and tasks. Next, we provide a short survey of impact factors that primarily affect human-robot trust.
B. Impacting Factors of Human-Robot Trust
Interpersonal trust usually involves the beliefs of ability, integrity, and benevolence [10]. For HRI, given that a robot has good intentions to collaborate with the human, human trust in robots can be influenced by factors within the robotic system, contextual (e.g., task and environmental related), and nature and characteristics of the human working with the robotic systems [6], [22], [79]. Hancock et al.'s 2011 and 2021 meta-analyses [8], [22] further suggested that robot performance is the most important factor that influences the level of trust and there are significant individual precursors to trust within each category. This analysis reinforces the notion that robot trustworthiness/competence predicts trust in robots, and trust, in turn, predicts the use of robots. Though the meta-analysis found only a moderate influence of environmental factors on trust and little effect of human-related factors, the authors mention that the pool of studies examining the effect of robot attributes on trust was severely limited [7]. The survey by Hoff and Bashir [2] further emphasized that this finding of the insignificance of human-related trust factors can be attributed to the shortage of experiments focused on individual differences and should be the focus of further investigation in the coming years. Accordingly, the works studying different antecedents of human trust in robots in the literature can be categorized into robot-related (including performance-based factors and attribute-related factors), human-related (including ability-based and characteristic-based factors), and contextual (including team collaboration-related factors and task-based factors) factors.
The following factors can be categorized as robot-related factors:
Robot performance-based factors: robot reliability-related factors (i.e., reliability [80], [81], [82], dependability [7], [83], failure rate [84], false alarm rate [7], predictability [83], [84], robot consistency [83], robot timeliness [83]), robot competence and capability [28], [85], robot behavior [75], [86], robotic system transparency [87], [88], [89], robotic system safety- and security-related factors (i.e., robot operational safety, privacy protection, data security) [90], and level of automation (LOA) [22].
Robot attribute-related factors: robot appearance [91], robot proximity [92], [93], anthropomorphism [79], and robot physical attributes [94], [95].
The following factors are human-related:
Human ability-based factors: workload [2], training [5], human mental model [64], and situational awareness [96].
Human characteristic-based factors: self-confidence [66], faith [3], [4], stereotypical impressions [97], [98], personality [73], [75] (such as propensity to trust [99]), gender [25], attitudes [100], culture [101], experience [102], age [103], situation management [85], locus of control [85], perception of robot attributes [104], [105], and social influence [106].
The following factors can be categorized as contextual factors:
C. Trust Measurements and Scales
In contrast to the computational human-robot trust models we focus on in this paper, there are numerous trust measurement approaches for human trust in automation/robots based on the scales of the factors mentioned in the above section. Existing trust measurement approaches can be classified into two categories, i.e., posterior trust measurement and real-time trust measurement [111].
The posterior trust measurement approaches quantify human trust after interactions with robots/automation. A well-designed questionnaire is the most typical approach. The acknowledged questionnaire designs include [40], [80], [83], [112]. For example, Muir's 7-point Likert scale trust questionnaire [40], [113] has been adopted for human-robot trust measure as shown in Table 2. Rosemarie et al. developed the HRI Trust Scale based on five dimensions: team configuration, team process, context, task, and system [83]. Corresponding to different scenarios, questionnaires can be modified to accommodate specific features of the robots or automation [114] [115]. A set of scales corresponding to different factors, e.g., self-confidence and reliance, can be extracted from the questionnaire. In addition to the questionnaire method, which derives the value of trust based on all relevant factors from well-designed questionnaires, some work directly requires the participants to provide their levels of trust via a simple continuous scale, e.g., from 0 (fully distrust) to 100 (fully trust) [116] or measuring via objective behavioral actions such as use-choice [117]. Besides considering these scales as a multi-dimensional representation of the trust [118], sometimes it is preferable to combine the set of scales into binary [119], discrete [24], ordinal [120], or continuous values [121]. For example, Schaefer developed a 40-item trust scale [86] using a 0-100 percentage score. This way, a single-valued trust measure can be used as a metric for post decision-making or design purposes. The trust values obtained by these posterior measurement approaches provide valuable guidelines for robot/automation design. However, they do not offer insight into the trust evolution during the task execution. It is also skeptical that human operators can accurately and timely quantify the level of trust in their mental states [122]. They are hence less attractive to developing real-time robot motion planning and control algorithms. In addition, the drawback of high variances in the results is also mentioned in the literature [123].
Real-time trust measurement approaches have emerged in recent years. The psycho-physical signals, such as Electroencephalography (EEG) and blood oxygenation signals, are usually used to determine the trust values [111], [124], [125]. Although EEG sensors are good at emotion recognition, questions like how accurate this classification is, or whether the EEG sensors can be used to measure emotion signals quantitatively, remain unanswered.
Although the above trust measurement approaches have been proved and widely adopted by industry or academia, the lack of trust quantification and prediction and/or its transition process limits their applications in robot control and motion planning. As a result, this motivates the current paper on computational trust models and their robot control and robotics applications.
Computational Human-Robot Trust Models
In this section, we provide a survey of computational human-robot trust models and a discussion regarding the respective merits and drawbacks of these models. Based on their respective modeling techniques, we categorize the models into five major classes: performance-centric algebraic, time-series, Markovian Decision Process (MDP)/Partially Observable MDP (POMDP), Gaussian-based, and dynamic Bayesian network (DBN)-based trust models. Note that other types of human-robot trust models also exist, but it is often one model for one category and hence we are not elaborating these models in this paper. We will introduce each category of models in a subsection. Each subsection will start with a brief introduction of the modeling method, followed by the detailed formulations of some representative models in the literature utilizing the modeling technique to describe human-robot trust and their corresponding robot control and robotics applications, and conclude with an analysis of the pros and cons of the trust model category. We will also briefly discuss the various approaches used to learn the model parameters. Although we summarized many factors in the previous sections, only a small portion of these factors have been adopted in the following computational trust models. Also note that, to keep notation consistency, we will use
A. Performance-Centric Algebraic Trust Models
In this subsection, we introduce performance-centric algebraic trust models. These models are deterministic and algebraic. Consistent with Hancock et al.’s meta analyses [8], [22], these models consider robot performance or human-robot team performance as the main impacting factor of trust and use observations to evaluate performance. Algebraic formulations are used to compute performance and human-robot trust. Next, we review some representative performance-centric algebraic trust models and their robotics and robot control applications in the literature.
Floyd et al. proposed a performance-centric computational trust model for a simulated wheeled unmanned ground robot working cooperatively with a simulated human agent in movement and patrol tasks. It would be ideal if a robot is guaranteed to operate in a trustworthy manner. However, it may be impossible to elicit a complete set of rules for trustworthy behaviors if the robot needs to handle changes in teammates, environments, or mission contexts. Therefore, this work developed a trust model and a case-based reasoning framework to enable a robot to determine when to adapt its behavior to be more trustworthy for the team [126]. Behaviors that the robot itself can directly control were defined as the modifiable component, e.g., speed, obstacle padding, scan time, scan distance, etc. Define different modifiable component sets as
\begin{equation*}
B := \langle c_{1}, c_{2},\ldots, c_{m} \rangle,
\end{equation*}
\begin{equation*}
Trust_{B} := \sum _{j=1}^{n} w_{j} \times cmd_{j},
\end{equation*}
\begin{align*}
CB =& \lbrace \langle \mathcal {E}_{\text {past},1}, B_{\text {final},1}\rangle, \langle \mathcal {E}_{\text {past},2}, B_{\text {final},2}\rangle, \ldots,\\
&\langle \mathcal {E}_{\text {past},l}, B_{\text {final},l}\rangle,\ldots \rbrace.
\end{align*}
A similar kind of binary observations mechanism was adopted in the RoboTrust algorithm [127] capturing human trust in unmanned ground vehicles (UGVs) platoon to report their driving conditions under cyber attacks. The UGV's average performance over different time intervals was calculated, and the most conservative (worst) average performance was then used to update trust. The trust update equation is,
\begin{equation*}
RoboTrust(k)=\min \left(\begin{array}{c}\frac{\sum _{j=k-\tau _{T}}^{k}OB(j)}{\tau _{T}+1} \\
\frac{\sum _{j=k-\tau _{T}-1}^{k}OB(j)}{\tau _{T}+2} \\
\vdots \\
\frac{\sum _{j=k-c_{T}}^{k}OB(j)}{c_{T}+1} \end{array} \right),
\end{equation*}
Xu and Dudek developed an event-centric trust model to model and predict trust change after the HRI session [104]. Within each interaction session, multiple events can occur, e.g., cold start and glitches caused by low reliability. The trust model was represented as a weighted linear sum with the trust output in a discrete level,
\begin{equation*}
\Delta T = \text {Quantization}\left(\sum _{i, j, k}\left(\omega _{0} \! + \! \omega _{i} \epsilon _{i}^{pe}(W)\! + \!\omega _{j} \epsilon ^{s}_{j} \! + \!\omega _{k} a_{k}^{s}\right)\right)
\end{equation*}
Overall, performance-centric algebraic trust models offer an efficient trust evaluation mechanism based on observational evidence. They are created based on selective human-robot trust impacting factors, which includes robot performance as the major component [45], [109], [126]. Due to this performance-centric perspective, these works usually mix trust with trustworthiness, which is inaccurate per Remark 1 in Section II-A. Besides performance, many other robot-related, contextual, and human-related factors could be considered, as discussed in Section II-B. These deterministic and algebraic models also neglected the uncertain, stochastic, and dynamic nature of human feeling and emotion. Most of the time, no human subject tests were conducted to collect real human data for learning these models. There is so far no unified framework or quantitative metrics for modeling either robot performance or trust. The reason why each specific model is adopted in a particular task or application is not very well justified. Furthermore, the above models are algebraic and do not capture the dynamic evolution of trust and the causal relationship with the impacting factors. As a consequence, performance-centric trust models are better suited for scenarios where long-term interactions effect can be ignored, and the robots' performance is the dominant consideration in HRI.
B. Time-Series Trust Models
We first introduce the time-series models in general. A time series is a series of data points ordered in time. A time-series model is a dynamic system used to fit the time-series data.
Definition 1:
[129] An autoregressive model of order
\begin{equation*}
y(k) = \sum _{i=1}^{p} \phi _{i} y(k-i) + w(k),
\end{equation*}
The autoregression polynomial involves regression of its own lagged (past) values. Note that other probability distributions, e.g., gamma, Laplace, may also be assumed [130].
Definition 2:
[129] A moving average model of order
\begin{equation*}
y(k) = w(k) + \sum _{j=1}^{q} \theta _{j} w(k-j),
\end{equation*}
The moving-average polynomial involves a linear combination of the error terms in time.
Definition 3:
[129] A time-series data
\begin{equation*}
y(k) = \sum _{i=1}^{p} \phi _{i} y(k-i) + w(k) + \sum _{j=1}^{q} \theta _{j} w(k-j) \tag{1}
\end{equation*}
An autoregressive moving average vector form (ARMAV) is the multivariate form of ARMA and allows the use of multiple time series to model a system's input/output relationship [131], [132]. In (1), vectors
Definition 4:
[133] An ARMA with exogenous input, abbreviated ARMAX, is
\begin{align*}
y(k) =& \sum _{l=1}^{r}\gamma _{l} u(k-l) + \sum _{i=1}^{p} \phi _{i} y(k-i)\\
&+ \sum _{j=1}^{q} \theta _{j} w(k-j) + w(k)
\end{align*}
For example, according to Definition 4, a simple dynamic regression model can be given as
\begin{equation*}
y(k)=\phi y(k-1)+\gamma u(k)+w(k).
\end{equation*}
Similarly, a multivariate ARMAX model can be constructed using vector sequences of
Lee and Moray developed an ARMAV of the time-series trust model to capture the dynamic evolution of human trust in a machine performing a juice pasterization task in a simulated supervisory control task [36]. The authors manipulated the reliability of the automation and measured overall operator performance and trust in the automation. They found that a time-series model that incorporated automation performance and the fault or failure rate of the automation was most predictive of eventual trust. However, the noise part was neglected in the final time-series trust model, probably because the model fitting returned a zero coefficient for the noise error term. This means that the model does not consider past forecast errors to explain the current value, which could indicate a lack of autocorrelation in the data. However, since the fitting of time series models involves estimation and statistical inference, the parameter might not be exactly zero due to estimation errors. Therefore, this model may not be accurate and have reduced estimation and prediction accuracy. Inspired by this model, Sadrfaridpour et al. proposed a time-series human trust in robot model for human-robot collaborative assembly tasks by further considering human performance and fault occurrence [137]:
\begin{align*}
T(k)=AT(k-1)+B_{1}P_{R}(k)+B_{2}P_{R}(k-1)+C_{1}P_{H}(k)\\
+\,C_{2}P_{H}(k-1)+D_{1}fault(k)+D_{2}fault(k-1)\tag{2}
\end{align*}
\begin{equation*}
P_{R}=P_{R,\max }-|V_{H}(k)-V_{R}(k)|,
\end{equation*}
\begin{equation*}
P_{H}(k)= \frac{F_{max,iso}(k)-F_{th}}{MVC-F_{th}},
\end{equation*}
\begin{align*}
F_{max,iso}(k)=& F_{max,iso}(k-1) - C_{f} F_{max,iso}(k-1) \\
&\cdot\frac{F(k-1)}{MVC} + C_{r} (MVC - F_{max,iso}(k-1))
\end{align*}
\begin{equation*}
F_{th}=MVC \frac{C_{r}}{2C_{f}}\left(-1+\sqrt{1+\frac{4C_{f}}{C_{r}}}\right)
\end{equation*}
Azevedo-Sa et al. proposed a linear time-invariant (LTI) system model and estimated the dynamics of drivers' trust in level-3 autonomous driving [145]. The authors considered the influence of Boolean event signals of the alarm system in the autonomous driving system. Gaussian variables were used to model trust
\begin{align*}
T\left({k+1}\right)&=\mathbf {A} T(k)+\mathbf {B}\left[\begin{array}{c}L(k) \\
F(k) \\
M(k) \end{array}\right]+u(k) \\
\left[\begin{array}{l}\varphi (k)\\
v(k)\\
\pi (k) \end{array}\right]&=\mathbf {C} T(k)+w(k)
\end{align*}
Zheng et al. further extended time-series models to a human multi-robot trust model and used the model to encode human intention into the multi-robot motion tasks in offroad environments [147]. Each robot indexed with
\begin{align*}
\mathbf {x}_{1:I}(k) &= \bm {B}_{0} \mathbf {x}_{1:I}(k-1) + \sum _{m=1}^{M}\bm {B}_{m} \mathbf {z}_{1:I,m}(k) + \mathbf {b} + \boldsymbol {\epsilon }_{w}(k),\tag{3a}\\
\mathbf {y}_{1:I}(k) &= \mathbf {x}_{1:I}(k) - \mathbf {x}_{1:I}(k) + \boldsymbol {\epsilon }_{v}(k), \tag{3b}
\end{align*}
\begin{align*}
\bm {B}_{0} = \begin{bmatrix}\beta _{0} & & \\
{\beta ^{\prime }\beta _{0}} & \beta _{0} & \\
& \ddots & \ddots & \\
& & {\beta ^{\prime }\beta _{0}} & \beta _{0} \end{bmatrix}_{I \times I}
\end{align*}
\begin{align*}
\bm {B}_{m} = \begin{bmatrix}\beta _{m} & & \\
{\beta ^{\prime }\beta _{m}} & \beta _{m} & \\
& \ddots & \ddots & \\
& & {\beta ^{\prime }\beta _{m}} & \beta _{m} \end{bmatrix}_{I \times I}
\end{align*}
Time-series trust models can capture the dynamic nature of trust and different causal factors that affect human-robot trust, including previous trust value that gives information about the memory of trust, current and previous robot and human performance, and other robot-related, human-related, and contextual factors. The parameter for each factor can be identified using model identification methods and corresponding toolboxes based on human subjective test data. Compared to the performance-centric algebraic trust models in Section III-A, the time-series trust models show the trust change due to performance increase or faults as a gradual process capturing inertia in an individual's level of trust instead of instantaneous change. The time-series trust models can also accommodate the uncertainty in data and find parameters that yield the best estimation and forecast accuracies. However, it is also important to note the limitations of the time-series trust models. Many time-series trust models assume a linearity form, which may oversimplify the relation between trust and its impacting factors. Similar to performance-based trust models, there lacks a unified framework for modeling robot performance and trust in different application scenarios. As a result, time-series trust models can be a suitable choice when it is necessary to consider the dynamics and multi-dimensionality of trust while maintaining a reasonable level of complexity to facilitate further trust dynamics analysis, especially for linear systems.
C. Markov Decision Process (MDP)-Based Trust Models
In this subsection, we first give an introduction to the general modeling approach.
Definition 5:
[150] A Markov Decision Process (MDP) models an agent's stochastic decision-making process in discrete time and can be represented by a tuple
is a set of states, which is called the state space;S is a set of actions, which is called the action space;A is the state transition function representing the probabilityT(s,a,s^{\prime }): S\times A\times S\to [0,1] of resulting in state\Pr (s_{k+1}=s^{\prime }|s_{k}=s,a_{k}=a) after taking actions_{k+1}=s^{\prime } at statea_{k}=a , wheres_{k}=s ;s, s^{\prime }\in S, a\in A is the reward received after transitioning from stateR(s,a,s^{\prime }): S\times A\times S\to \mathbb {R} to states_{k}=s due to actions_{k+1}=s^{\prime } ;a_{k}=a is the discounting factor that discounts the weight of future rewards.\gamma \in [0,1]
The reward
\begin{align*}
&\Pr (s_{k+1},r_{k+1}|s_{k},a_{k},r_{k},s_{k-1},a_{k-1},\ldots,r_{1},s_{0},a_{0}) \\
&=\Pr (s_{k+1},r_{k+1}|s_{k},a_{k}).
\end{align*}
Definition 6:
[154] Given two MDPs
is the Cartesian product ofS_{1} \times S_{2} andS_{1} and represents the state set ofS_{2} ;\mathcal {M} is union ofA_{1} \cup A_{2} andA_{1} and represents the action set ofA_{2} ;\mathcal {M} ,T_{1,2}((s_{1}, s_{2}), a,(s_{1}^{\prime }, s_{2}^{\prime }))=T_{1}(s_{1}, a, s_{1}^{\prime }) T_{2}(s_{2}, a, s_{2}^{\prime }) , ifR_{1,2}((s_{1}, s_{2}), a)=R_{1}(s_{1}, a)+R_{2}(s_{2}, a) and botha \in A_{1} \cap A_{2} andT_{1}(s_{1}, a, s_{1}^{\prime }) are defined;T_{2}(s_{2}, a, s_{2}^{\prime }) ,T_{1,2}((s_{1}, s_{2}), a,(s_{1}^{\prime }, s_{2}))=T_{1}(s_{1}, a, s_{1}^{\prime }) ifR_{1,2}((s_{1}, s_{2}), a)= R_{1}(s_{1}, a) anda \in A_{1} \backslash A_{2} is defined;T_{1}(s_{1}, a, s_{1}^{\prime }) ,T((s_{1}, s_{2}), a,(s_{1}, s_{2}^{\prime }))=T_{2}(s_{2}, a, s_{2}^{\prime }) ifR_{1,2}((s_{1}, s_{2}), a)= R_{2}(s_{2}, a) anda \in A_{2} \backslash A_{1} is defined.T_{2}(s_{2}, a, s_{2}^{\prime })
Definition 7:
[155] A partially observable Markov decision process (POMDP) is an extension of MDP where the state information is incomplete and can only be partially observed. A POMDP can be represented by a tuple
is a set of states;S is a set of actions;A is the state transition probability;T(s,a,s^{\prime }): S\times A\times S\to [0,1] is a set of observations;O is the reward;R(s,a,s^{\prime }): S\times A\times S\to \mathbb {R} is the emission probability that gives the likelihood of observationP_{o}(s,o): S\times O\to [0,1] given the current stateo , i.e.,s , where\Pr (o_{k}=o|s_{k}=s) ;s\in S, o\in O is the discounting factor.\gamma \in [0,1]
In POMDP, because the actual state is hidden, decisions need to be made under the uncertainty of the actual state. The belief of the state can be updated based on the observation and action taken (see (4)). POMDP is often computationally intractable and can be approximated using point-based value iteration methods [156] or reinforcement learning [157]. Next, we review some representative MDP and POMDP trust models in the literature and their robotics and robot control applications.
Bo et al. introduced an optimal task allocation framework for repetitive industrial assembly tasks under human-robot collaboration [154]. The human trust MDP model
\begin{equation*}
M = M^{w} \Vert M^{r} \Vert M^{t} \Vert M^{f}.
\end{equation*}
Nam et al. studied the human robotic swarm trust model for the supervisory control of a target foraging/search mission [94], [95]. Trust is analyzed in three different LOAs: manual, mixed-initiative (MI), and fully autonomous LOA. In the manual LOA, the heading directions of a swarm are re-guided by an operator in changing the target search region. In the autonomous LOA, the swarm robots direct the heading angles by themselves. In the MI LOA, a switch mode allows the LOA to change between manual and autonomous operations based on the target search rate. Participants were asked to command the heading directions of the swarm in a simulator. Trust ratings were queried from participants based on their subjective feelings, which ranged from strongly trust to strongly distrust. Meanwhile, a set of data was collected, including the mean and variance of the robots' heading angles, the length of the velocity vector from the user command input, the convex hull area taken by the swarm, and the number of targets found. The results found that human trust was affected more by the swarm's physical characteristics, such as heading angles and convex hull area, than task performance. Fig. 4 illustrates the swarm's physical characteristics. Trust is modeled as an MDP
whereS=\lbrace s=(h_{k},c_{k},i_{k},t_{k})|h_{k} \in \mathbb {R}^{\geq 0}, c_{k} \in \mathbb {R}^{+}, i_{k} \in \lbrace 0,1\rbrace, t_{k} \in [0,1]\rbrace is the heading variance at timeh_{k} ,k is the convex hull area,c_{k} indicates if an intervention command is given or not, andi_{k} denotes the set of finite trust states;t_{k} denotes the finite set of actions;A is a reward function that describes the weighted sum of featuresR(s_{k},a,s_{k+1})=\sum _{j=1}^{n} \alpha _{j} \phi _{j} ;\phi _{j} is the state transition function;T(s_{k},a,s_{k+1}) is the discount factor.\gamma \in [0,1]
Variations of the trust model were developed for different LOAs. An inverse reinforcement learning (IRL)6 algorithm was then utilized to learn the reward function, which was used in the MDP to predict trust in the three different LOAs. A swarm simulator with 32 homogeneous robots was deployed, and 30 participants were recruited to operate the robot simulator and provide trust feedback. Results showed that the model could predict trust in real-time without incorporating user input that provided ground truth values of trust. Compared to the DBN trust model [159] to be introduced in Section III-E, the MDP trust model reduced prediction error in all three LoAs. The MDP model concatenated human trust in swarm and swarm parameters as the state and captured its dynamic change. The model evaluated the trust of the human in the swarm with reference to the operator's real-time rating of swarm behavior in the training process. However, the explicit relation between trust and swarm parameters remains unknown. The multi-dimensional metrics in the state of MDP can be challenging for the system to accommodate a larger size of state space for the proposed trust model to perform online adaptation. These metrics were presented in relatively coarse levels to reduce the computation burden in the experiment.
Akash et al. established a POMDP model [160], [161], [162] to capture the dynamic change of human trust and workload for the collaboration between humans and an intelligent decision-aid of a teleoperated robot. The teleoperated robot that was equipped with a camera and a chemical sensor can detect dangers in buildings and provide a recommendation for the human companion on whether or not to take on protective gear. Humans can choose to either comply with or disregard the robot's recommendation. The system's reliability, transparency of the user interface, and presence/absence of danger were manipulated to observe humans' changes in trust and workload. Human compliance to the decision aid was utilized to implicitly infer their trust, while the human response time to the decision aid's recommendation was used to infer workload. Accordingly, a trust-workload POMDP was constructed as in Definition 7 with the finite state set
\begin{equation*}
a^* := \text{argmax}_{a} \sum _{s\in S} b(s) Q_\text {MDP}(s,a), \tag{4}
\end{equation*}
\begin{align*}
b^{\prime }(s^{\prime }) & := \Pr (s^{\prime }|o,a,b)\\
& := \frac{\Pr (o|s^{\prime },a)\sum _{s\in S}\Pr (s^{\prime }|s,a)b(s)}{\mathop {\sum }_{s^{\prime } \in S} (\Pr (o|s^{\prime },a)\sum _{s\in S}\Pr (s^{\prime }|s,a)b(s))},
\end{align*}
\begin{align*}
Q_\text {MDP}(s,a) &:= \sum _{s^{\prime } \in S}T(s,a,s^{\prime })(R(s^{\prime }|s,a) + \gamma V(s^{\prime })), \\
Q^{\tau }(s,\tau) &:= \sum _{S_{A},E}Pr(S_{A},E)Q_\text {MDP}(s, a := [S_{A}, E, \tau ]), \\
V(s) &:= \text {max}_{\tau }Q^{\tau }(s,\tau).
\end{align*}
Zahedi et al. proposed a Meta-MDP trust model for the human supervision task [163]. This model introduced the concept of trust transition based on the distance
\begin{equation*}
\mathbb {P}\left(s_{k+1}=s_{i-1}|s_{k}=s_{i}, \pi \right)=\omega (s_{i})(1 - \mathcal {P}(EX(\pi))),
\end{equation*}
\begin{equation*}
\mathbb {P}\left(s(k+1)=s_{i}|s(k)=s_{i}, \pi \right) = \omega (s_{i})\mathcal {P}(EX(\pi)).
\end{equation*}
\begin{equation*}
\mathbb {P}\left(s(k+1)=s_{i+1}|s(k)=s_{i}, \pi \right) = 1 -\omega (s_{i}).
\end{equation*}
\begin{equation*}
\mathcal {C}\!\left(s(k), \pi \right) = \left(1 - \omega \left(s(k)\right)\right) C_{r}(\pi) + \omega (s(k)) C_{s} (\pi),
\end{equation*}
Overall, the (PO)MDP models consider the dynamic and probabilistic nature of human trust and give a systematic approach to present the transition of robots' states and actions for seeking trust-associated optimal policies. They can describe the relation between human trust and robot behaviors/performance with corresponding probabilistic functions. On the other hand, these models are trained using a limited set of finely-tuned robot behaviors, typically unable to encompass the full spectrum of behaviors encountered during real-world operations. The reliability of human trust and decision-making when operating under such a trained model may be compromised by unforeseen robot behaviors. The generated optimal policy also lacks interpretability regarding the underlying human decision-making mechanism. In (PO)MDP, trust is usually modeled together with other variables such as workload and intervention as the state, which makes it difficult to reveal the relation between trust and these variables, which are often trust-impacting factors. Furthermore, due to the need to accommodate system and state changes, it is necessary to have an efficient online algorithm to find the optimal policy dynamically. In addition, these POMDP models are designed and trained based on a group of people. Thus, they represent a general prediction of human trust and corresponding optimal policy. That is, it is not an individualized model for a specific person that takes into account personality, cognitive abilities, etc. Last but not least, although (PO)MDP can be used for continuous state and action spaces, most existing (PO)MDP-based trust models merely consider discrete action and state spaces with a limited number of values. In practice, human trust evolves continuously instead of jumps/drops discretely. Therefore, such models can only approximate trust changes to some extent. In general, (PO)MDP trust models can be considered a superior choice for scenarios where the environment and tasks are relevantly well-defined with transition and observation uncertainties, and the state-action data pairs are readily accessible. The resulting model can integrate the trust state with robot behaviors, which is advantageous for optimizing the overall system performance.
D. Gaussian-Based Trust Models
In this subsection, we first briefly introduce the Gaussian distribution and Gaussian process (GP) for continuous random variables and processes.
Definition 8:
A Gaussian distribution, also known as the normal distribution, is a continuous probability distribution for a real-valued random variable, which can be represented mathematically as
\begin{equation*}
\mathcal {N}(x; \mu, \sigma)=\frac{1}{\sqrt{2\pi }\sigma }exp\left(-\frac{1}{2}\left(\frac{x-\mu }{\sigma }\right)^{2}\right)
\end{equation*}
Many nature phenomena tend to approximate a Gaussian distribution. Graphically, a normal distribution looks like a “bell curve”. The cumulative distribution function (CDF)
\begin{equation*}
\Phi (x)=\frac{1}{\sqrt{2 \pi }} \int _{-\infty }^{x} \exp \left(-\frac{t^{2}}{2}\right) \text {d} t. \tag{5}
\end{equation*}
\begin{align*}
&\mathcal {N}(\bm {x};\bm {\mu },\bm {\Sigma }) \\
&=\frac{1}{(2\pi)^{m/2}|\bm {\Sigma }|^{1/2}}exp\left(-\frac{1}{2}(\bm {x}-\bm {\mu })^{T}\bm {\Sigma }^{-1}(\bm {x}-\bm {\mu })\right)
\end{align*}
Definition 9:
A Gaussian process (GP) [166] is a stochastic process, i.e., a collection of random variables over time or space, which is fully represented by a mean function
\begin{align*}
f_{GP}(\bm {z})&\sim \mathcal {GP}(m(\bm {z}),k(\bm {z},\bm {z}^{\prime })) \\
m(\bm {z})&=\mathbf{E}[f_{GP}(\bm {z})] \\
k(\bm {z},\bm {z}^{\prime })&=\mathbf{E}[(f_{GP}(\bm {z})-m(\bm {z}))(f_{GP}(\bm {z}^{\prime })-m(\bm {z}^{\prime })]
\end{align*}
Any finite collection of these random variables
Chen et al. [113], [167] introduced a linear Gaussian equation to build human trust in a robot
\begin{equation*}
\theta _{k+1}\sim \mathcal {N}(\alpha _{e_{k+1}}\theta _{k}+\beta _{e_{k+1}},\sigma _{e_{k+1}}),
\end{equation*}
\begin{equation*}
e_{k+1} = \text{performance}(x_{k+1},x_{k},a^{R}_{k},a^{H}_{k}),
\end{equation*}
\begin{equation*}
v(x_{0}|\pi ^{R},\pi ^{H}) = \mathop {\mathbf {E}}_{a^{R}_{k} \sim \pi ^{R}, a^{H}_{k} \sim \pi ^{H}}\sum _{k=0}^{\infty }\gamma ^{k} r(x_{k},a^{R}_{k},a^{H}_{k}),
\end{equation*}
\begin{equation*}
\pi ^{R}_* = \text{argmax}_{\pi ^{R}} \mathop {\mathbf {E}}_{\pi ^{H}}v(x_{0}|\pi ^{R},\pi ^{H}).
\end{equation*}
Soh et al. modeled trust as a latent dynamic function
\begin{equation*}
\tau ^{a}_{t}(\bm {x}_{k}) =\int P(c_{k}^{a} | f^{a},\bm {x})p_{k}(f^{a})df^{a},
\end{equation*}
\begin{equation*}
p_{k}\left(f^{a} \mid \bm {x}_{k-1}, c_{k-1}^{a}\right)=\frac{P\left(c_{k-1}^{a} \mid f^{a}, \bm {x}_{k-1}\right) p_{k-1}\left(f^{a}\right)}{\int P\left(c_{k-1}^{a} \mid f^{a}, \bm {x}_{k-1}\right) p_{k-1}\left(f^{a}\right) d f^{a}} \tag{6}
\end{equation*}
\begin{equation*}
P(c^{a}_{k}|f^{a},\bm {x}_{k})=\Phi \left(\frac{c^{a}_{k}(f^{a}(\bm {x}_{k}) - m(\bm {x}_{k}))}{\sigma ^{2}_{n}}\right),
\end{equation*}
There are also other works using different probability distributions to model trust. For example, Azevedo-Sa et al. proposed a bidirectional trust model that takes into account the difference between the trustee's capabilities
\begin{equation*}
\tau (t) = \int _\Lambda p \left(\Omega (\lambda, \bar{\lambda }, t) = 1 |\lambda \right) bel(\lambda, t-1)d\lambda, \tag{7}
\end{equation*}
\begin{equation*}
p \left(\Omega (\lambda, \bar{\lambda }, t)=1|\lambda \right) = \prod ^{n}_{i=1}\left(\frac{1}{1 + \exp (\beta _{i}(\bar{\lambda }_{i} - \lambda _{i}))}\right)^{\zeta _{i}},
\end{equation*}
\begin{equation*}
\tau (t) = \prod _{i=1}^{n} \psi (\bar{\lambda }_{i}),
\end{equation*}
\begin{equation*}
\psi (\bar{\lambda }_{i}) = {\begin{cases}1 & \text {if } 0 \leq \bar{\lambda }_{i} \leq l_{i} \\
\frac{u_{i} - \bar{\lambda }_{i}}{u_{i} - l_{i}} & \text {if } l_{i} \leq \bar{\lambda }_{i} \leq u_{i}\\
0 & \text {if } u_{i} \leq \bar{\lambda }_{i} \leq 1\\
\end{cases}}.
\end{equation*}
\begin{equation*}
\hat{l}_{i}, \hat{u}_{i} = \arg \min _{[0,1]^{2}}\int _\Lambda \Vert \tau -\hat{\tau }\Vert ^{2}d\lambda \tag{8}
\end{equation*}
As another example, Guo and Yang proposed a Beta distribution-based human-robot trust model [171]. The mean of the Beta distribution was interpreted as the predicted trust, and both parameters of the Beta distribution were updated based on prior trust information. Hence, this model captures the nonlinearity of trust dynamics. The authors tested the proposed model with 39 human participants interacting with four drones in a simulated surveillance mission. The ANOVA results demonstrated that the proposed Beta distribution model had smaller root mean square errors in predicting human trust than the ARMAV and OPTIMo (to be introduced in Section III-E) trust models under their scenario.
Gaussian-based (or in general, probability distribution based) trust models provide a means to model trust as a continuous random variable or stochastic process, which may better capture the probabilistic and continuous nature of human trust in robots in practice. The trust value or distribution also gets updated dynamically through the updates of robot performance or the availability of a new observation. Furthermore, the Gaussian-based trust models are flexible and can be integrated into other decision-making and learning frameworks, such as POMDP and Bayesian inference. Human subject data can be used to learn the model parameters, resulting in either personalized or population-based models. However, it is unclear if the human trust does follow Gaussian, Beta, or another type of distribution, which might limit the reliability and prediction accuracy of the trust models. This further challenges the use of a rational Bayesian framework for updating trust. Similar to other trust models in the literature, only robot performance was used as the trust impacting factor in the existing Gaussian-based trust models, eliminating the effects such as task difficulty, risk, etc. The robot performance models used in these trust models may also be overly simplified and did not fully leverage the task features and information. Overall, assuming the appropriate underlying distribution, a Gaussian-based (or probability distribution based) trust model offers an intriguing option for capturing the randomness and stochastic nature of the trust process without imposing significant data demands.
E. Dynamic Bayesian Network (DBN)-Based Trust Models
In this subsection, we first introduce the Dynamic Bayesian Network (DBN). Denote
Definition 10:
[172] A DBN is a probabilistic graphical model that contains
An initial Bayesian network (BN) consisting of (a) an initial directed acyclic graph (DAG)
containing the variables inG_{0} and (b) an initial probability distribution\bm {X}(0) of these variables.P_{0} A transition BN which is a template consisting of (a) a transition DAG
containing the variables inG_\to and (b) a transition probability distribution\bm {X}(k)\cup \bm {X}(k+1) which assigns a conditional probability to every valuePr of\bm {x}(k+1) given every value\bm {X}(k+1) of\bm {x}(k) ,\bm {X}(k) \begin{equation*} Pr(\bm {X}(k+1)=\bm {x}(k+1)\mid \bm {X}(k)=\bm {x}(k)). \end{equation*} View Source\begin{equation*} Pr(\bm {X}(k+1)=\bm {x}(k+1)\mid \bm {X}(k)=\bm {x}(k)). \end{equation*}
The DBN consisting of (a) the DAG composed of the DAG
and forG_{0} the DAG0 \leq k \leq K - 1 evaluated atG_{\to } and (b) the following joint probability distribution:k \begin{equation*} {\kern-14.45377pt}Pr(\bm {x}(0), \ldots \bm {x}(K))=P_{0}(\bm {x}(0))\prod _{k=0}^{K-1} Pr(\bm {x}(k+1)\mid \bm {x}(k)). \end{equation*} View Source\begin{equation*} {\kern-14.45377pt}Pr(\bm {x}(0), \ldots \bm {x}(K))=P_{0}(\bm {x}(0))\prod _{k=0}^{K-1} Pr(\bm {x}(k+1)\mid \bm {x}(k)). \end{equation*}
Fig. 6 shows some example DBNs. DBN is a generalization of linear state-space models (e.g., Kalman filters), linear and normal forecasting models (e.g., ARMA), and simple dependency models (e.g., hidden Markov models (HMMs)) [173]. The expectation maximization (EM) approach can be used to learn the parameters of a DBN [174]. When the inference of the model becomes intractable, Bayesian inference [174] can provide approximate inference for learning. Next, we summarize some DBN-based human-robot trust models in the literature.
Xu and Dudek proposed a DBN-based computational trust model called OPTIMo [159], which relates human's current degree of trust
\begin{align*}
\mathcal {P}&(t_{k},t_{k-1},p_{k},p_{k-1}):= Pr(t_{k}|t_{k-1},p_{k},p_{k-1})\\
&\approx \mathcal {N}(t_{k};t_{k-1}+\omega _{tb}+\omega _{tp}p_{k}+\omega _{td}(p_{k}-p_{k-1}),\sigma _{t}), \end{align*}
\begin{align*}\mathcal {O}_{i}&(t_{k},t_{k-1},i_{k},e_{k}) := {\begin{cases}Pr(i_{k}=1|t_{k},t_{k-1},e_{k})\\
Pr(i_{k}=0|t_{k},t_{k-1},e_{k}) \end{cases}}\\
&={\begin{cases}\mathcal {S}(\omega _{ib}+\omega _{it}t_{k} +\omega _{id}\Delta t_{k}+\omega _{ie}e_{k})\\
1-\mathcal {S}(\omega _{ib}+\omega _{it}t_{k} +\omega _{id}\Delta t_{k}+\omega _{ie}e_{k}) \end{cases}}
\end{align*}
\begin{align*}
&{O}_{c}(t_{k},t_{k-1},c_{k}) := {\begin{cases}Pr(c_{k}=1|t_{k},t_{k-1}) \\
Pr(c_{k}=-1|t_{k},t_{k-1}) \\
Pr(c_{k}=0|t_{k},t_{k-1}) \end{cases}} \\
&Pr(c_{k}=1|t_{k},t_{k-1})=\beta _{c}+(1-3\beta _{c})\cdot \mathcal {S}(\kappa _{c}[\Delta t_{k}-o_{c}]) \\
&Pr(c_{k}=-1|t_{k},t_{k-1})=\beta _{c}+(1-3\beta _{c})\cdot \mathcal {S}(\kappa _{c}[-\Delta t_{k}-o_{c}]) \\
&Pr(c_{k}=0|t_{k},t_{k-1})= 1\! -\! 2\beta _{c}\! -\! (1-3\beta _{c})(\mathcal {S}(\kappa _{c}[\Delta t_{k}-o_{c}])\\
&\qquad\qquad+\mathcal {S}(\kappa _{c}[-\Delta t_{k}-o_{c}])),
\end{align*}
\begin{equation*}
\mathcal {O}_{f}(t_{k},f_{k}) := Pr(f_{k}|t_{k})\approx \mathcal {N}(f_{k};t_{k},\sigma _{f})
\end{equation*}
OPTIMo: A DBN-based model for dynamic, quantitative, and probabilistic trust estimates.

The smoothed trust belief at any time
\begin{align*}
bel_{s}(t_{k-1}) &=Pr(t_{k}|p_{1:K},i_{1:K},e_{1:K},c_{1:k},f_{1:k},t_{0})\\
&= \int \frac{ \overline{bel}(t_{k},t_{k-1})}{\int \overline{bel}(t_{k},t_{k-1})dt_{k-1}} bel_{s}(t_{k})dt_{k}
\end{align*}
\begin{align*}
\overline{bel}(t_{k},t_{k-1}):=\mathcal {O}_{i}(t_{k},t_{k-1},i_{k},e_{k}) \mathcal {O}_{c}(t_{k},t_{k-1},c_{k}) \\
\cdot \mathcal {O}_{f}(t_{k},f_{k}) \mathcal {P}(t_{k},t_{k-1},p_{k},p_{k-1}) bel_{f}(t_{k-1}).
\end{align*}
Data was collected from 21 users on a visual navigation task, where an aerial robot was autonomously controlled to track boundaries, and the human operator could intervene and provide trust feedback. The learned trust model was shown to be able to predict the human dynamic trust state more accurately than the benchmark models. In sum, OPTIMo introduced a probabilistic trust model capable of inferring a human operator's degree of trust in a robot as a probability distribution at each time step
Similar DBN trust models with different robot performance models can be found for the applications in runtime verification10 of multiple quadrotors' motion planning [176] and human-robot cooperative manipulation (co-manipulation) [177].11 Other works made some variations but generally followed a similar structure [180], [181], [182]. Following a similar structure as in Fig. 7, a DBN trust model for multi-robot symbolic robot motion planning12 with a human-in-the-loop was developed [180], [181]. Besides the robot performance, the DBN trust model also included human performance and faults as causal inputs. The resulting computational trust model was used as a metric for multi-robot task decomposition and allocation. Multi-robot simulations with direct human input and trust evaluation were conducted to implement the overall framework. Zheng et al. built a DBN human trust model for multi-robot bounding overwatch [184] tasks in offroad environments [182]. In such a scenario, trust was used to evaluate the reliability of autonomous overwatch robots in providing situational awareness to human-operated robots and making bounding decisions. The trust model used terrain traversability and visibility as the causal inputs to the latent trust state, which is associated with task specification, robot transition model, and the discrete mission environment. Only numerical simulations were conducted to demonstrate the overall framework.
Mangalindan et al. developed a POMDP framework for trust-seeking policies of a mobile manipulator to seek assistance from a human supervisor in object collection tasks [185]. The robot can perform the object collection task autonomously (corresponding to action
\begin{equation*}
R_{o,E}^{a} = \left\lbrace \begin{array}{ll}+2 & \text{if $(a,o,E)=(a^-,o^+,E^+)$};\\
+1 & \text{if $a=a^+$};\\
0 & \text{if $(a,o)=(a^-,o^-)$};\\
-3 & \text{if $(a,o,E)=(a^-,o^+,E^-)$}. \end{array} \right.
\end{equation*}
\begin{align*}
Pr(s_{k+1}|s_{k})=&Pr(T_{k+1}|E_{k+1},C_{k},a_{k}) \\
&\cdot Pr(C_{k+1})Pr(E_{k+1}|T_{k},C_{k},a_{k}). \tag{9}
\end{align*}
\begin{align*}
R(s,a^-)=&R_{o^+,E^+}^{a^-} Pr(E^+|o^+,C,a^-)Pr(o^+|T,C,a^-) \\
&+R_{o^+,E^-}^{a^-} Pr(E^-|o^+,C,a^-)Pr(o^+|T,C,a^-) \\
&+R_{o^-,E^+}^{a^-} Pr(o^-|T,C,a^-), \\
R(s,a^+)=&R_{\star,\star }^{a^+}.
\end{align*}
Mahani et al. developed a DBN-based trust model for multi-robot systems [188]. Fig. 9 shows the structure of the trust DBN model for human-multi-robot teams. This model treated a set of degrees of human trust in a robot team
\begin{align*}
&Pr(\bm {T}_{1:K},\bm {Y}_{1:K} \mid \bm {P}_{1:K}) \\
=&Pr(\bm {T}_{\bm {1}}) \prod _{k=1}^{K}Pr(\bm {T}_{k} \mid \bm {T}_{k-1}, \bm {P}_{k})Pr(\bm {Y}_{k} \mid \bm {T}_{k}), \\
=&Pr(\bm {T}_{\bm {1}})\prod _{k=1}^{K}\prod _{r=1}^{\bar{r}}{Pr(T_{k}^{r} \mid T_{k-1}^{r}, P_{k}^{r})} \\
&\cdot Pr(I_{k} \mid \bm {T}_{k})\prod _{r=1}^{\bar{r}}{Pr(F_{k}^{r} \mid T_{k}^{r})},
\end{align*}



A 2-layer neural network representing three variables
DBN-based trust models are able to capture the dynamic property of trust and characterize the time-based trust evolution through the state transition distributions. Although both discrete random variables and continuous random variables can be accommodated in DBN with their corresponding parameter learning approaches, most existing DBN-based trust models assumed discrete levels of trust. DBNs with continuous random variables allow the modeling of trust as a continuous dynamic process, which is more consistent with human subject studies and can be further explored. Furthermore, since the DBN-based trust models use graphs to represent the causal relations among trust impacting factors and trust measurements/feedback, they are more interpretable. DBN trust models can also describe non-symmetry in trust in multi-robot systems by formulating different likelihood functions to represent human trust in each robot. However, the accuracy of DBN models is highly subject to the choice of initial probability distributions. The DBN-based trust models can only capture human trust updates well under the prerequisite that initial bias is accurately described. It is challenging work to model a suitable initial bias without a good understanding of the problem and experience. Consequently, with a reliable estimate of the initial bias, the DBN-based trust model offers high interpretability and can be a preferable choice for scenarios with identified trust evolution dynamics and casual relations between trust impacting factors, the hiddden trust state, and trust observations/feedback.
Limitations and Discussions on Future Research Directions
This paper is by no means exhaustive, and we only intend to show some most representative categories of trust models in the literature. It can be observed from existing work that quantitative and computational trust modeling is an interdisciplinary area that integrates knowledge in system modeling, machine learning, statistics, human factors, and psychology with various HRI applications. For different HRI tasks, different performance models were considered which led to different metrics or calculations of the trust values even following similar trust dynamics. Out of various models, the time-series, DBN, and MPD/POMDP trust models seem to have received the most attention, which resulted in more development and extension. On the other hand, although a handful of studies in trust measures and modeling can be found thanks to the increasing attention in this field and its wide-spanned applications across different disciplines and applications, the current literature is still lacking in several aspects. In this section, we will discuss the shortcomings of the current trust modeling literature based on our observations. Correspondingly, we believe a few directions are worth pursuing for future research.
A. Missing Pieces in Current Trust Models
We believe there are several missing components in the current trust modeling works. Although many factors impacting human trust in robots were identified in the literature as summarized in Section II-B, only a small portion of these factors (e.g., robot performance, false alarm rate, human performance, fault, task difficulty) have been adopted in these trust models. These models may be considered trust-like instead of actually modeling trust in that they capture certain psychological qualities of trust rather than seeking to model all aspects of trust development accurately. Many works also restricted trust as a discrete random variable that can only take values from a finite set, or just binary. The main reason would probably be the computation required to consider larger continuous trust state space and the corresponding parameter learning algorithms. This calls for a more general model that can accommodate more important trust-impacting factors, and efficient parameter regression/estimation, approximation, or model learning algorithms that can handle large state space.
As pointed out in [10], [29], trust implies a decision in risk-taking. Only when a trustor is willing to make himself/herself vulnerable by taking risks, does there arise a situation that needs trust. That is, risk is requisite to trust. Furthermore, risk propensity, i.e., an individual's tendency to be either risk-seeking or risk-averse, will also affect his/her trust, making a personal difference. When a trustor's risk-taking decision leads to a positive outcome, then the trustor's trust in the trustee will be enhanced and vice versa. Hence, it does not make much sense to model trust in isolation from the perceived risk level. However, very few trust models considered the risk in HRI and took it into account in modeling. On the other hand, the relationship between trust and risk has already been well studied in business and psychology [10], [190]. Risk quantification and risk-based decision-making have also been well-adopted in the robotics literature [191]. There are also works focusing on quantifying human risk attitudes in HRI [192]. These may shed some light on computational trust modeling from the risk perspective.
Also, almost all the current work models human trust in robots assuming that the robot has goodwill and its main impacting factor is robot reliability and performance. However, in practice, a robot may be subjected to cyber security issues and not have benevolence. As discussed in [22], [23], intention-based trust should also be considered. A malicious robot may be associated with a high-risk level, which should cause a significant decrease in trust if detected. Even under the assumption of good-intent robots, the performance of a robot may degrade due to its capability limitations, which in turn may cause a sudden drop in trust and gets recovered slowly if its performance improves. However, current works fail to capture such dynamics. On the other hand, there is a vast literature on modeling agent reputation and trustworthiness using probabilistic measures such as the Beta and Dirichlet distributions [60], [62], [193], [194], [195] for detecting unreliable members in a multi-agent system. Although these models have not been used in human-robot trust modeling, they may have great potential in evaluating an agent that may have performance degradation or even be malicious.
Most importantly, trust should be calibrated and maintained over time instead of maximized. Trust calibration is the correspondence between a human's trust in a robot and this robot's capabilities [34], [110], [196]. In the HRI context, trust calibration is referred to one's perception of the robot and the actual competency and capability of the robot [197]. Overtrust (over-reliance) or undertrust (under-reliance) in robots results in negative effects on robot performance and hence either misuse or disuse of the robots [198], [199]. It was further experimentally verified in [113] that maximization of trust does not lead to optimal performance of the human-robot team. Despite the importance of calibrated trust, only a few trust models [109], [113], [167] have considered how to adjust subjective trust to match the robot's trustworthiness. Shafi formulated the trust calibration problem as the minimization of the difference between human trust in robot and robot trustworthiness [109]. Chen et al. embedded trust in a POMDP framework, where the human trust in the robot can be adjusted by the action taken by the robot [113], [167].
Furthermore, humans may be unable to accurately and timely quantify the level of trust in their mental states [122]. In such cases, neither trust questionnaires nor direct human trust feedback can reveal the actual human emotion. Correspondingly, the human-robot trust models that utilize such data for model parameter learning might lead to results that deviate from the ground truth and provide inaccurate predictions. On the other hand, as discussed in Section II-C, EEG or other psycho-physical signals, may offer an additional measure in addition to human feedback and can be integrated into the existing models to improve the estimation and prediction accuracy.
Moreover, trust can be easily transferred from one context to another. The idea of trust transferability has been adopted in many business settings [200]. Human-subjects studies in organization science motivated the investigation of trust differences in different contexts, e.g. across tasks, environmental or situational factors [200]. More recently, Soh et al. [28] started the investigation on how human trust in robot capabilities transfers across tasks. The authors modeled trust in robots as a context-dependent latent dynamic function. The human-subjects study showed that perceived task similarities, difficulty, and observed robot performance have an influence on trust transfer across tasks. The result further showed that a trust model accounting for the trust transfer led to better trust predictions than existing approaches on unseen participants. We expect the study on trust transferability can potentially make a computational trust model applicable to different tasks in different domains with different participants to improve the generality of the current models.
Trust models for multi-human multi-robot systems also become increasingly necessary to better understand human-robot teaming. Although there are a lot of multi-agent trustworthiness models available in the literature [201], very few current literature extend the paradigm to model the trust feeling involving the human aspect. Except for some initiatives in modeling an individual's trust in multi-robot systems [127], [147], [188] and swarms [94], [95], there does not seem to have any effort on modeling a group's trust in a robot or multiple robots yet. As a future research direction, it could also be interesting to explore a model that can seamlessly integrate human-multi-robot trust and inter-human trust.
Last but not least, data acquisition is a common challenge in all trust modeling. Because trust is a psychological construct, data gathered from human subject tests are necessary for training and parameter estimation of trust models. Given human subject tests are expensive to run, both in terms of money and time, it is always difficult to acquire sufficient human data for trust modeling. First of all, the required sample size of human subject tests is a critical question. In statistical tests, power analysis [202], [203] and Monte Carlo analysis [204] can be used to obtain statistically defensible numbers of required human subjects. However, so far, there has not emerged a consensus on the ideal sample size to result in a satisfactory training and parameter estimation results of the trust modeling. Lessons may be learned from related research areas. For example, in computer-human-interaction (CHI) research, the local standard for sample size of human subjects is around 12 [203]. Second, the appropriate number of trials per human subject is another important issue. For example, within-subject designs require a larger sample size than between-subject designs and are more statistically powerful [205]. However, the selection of trials number per human subject will largely depend on the specific human-robot task and the desired accuracy of the trust model. There lack standards and metrics regarding the optimal number of trials. Nevertheless, data scarcity remains a great challenge for trust modeling. We believe potential ways to circumvent these challenges will be to build standard datasets for trust measurement (please see the discussion on ImageNet in Section IV-C) and to explore strategies for constructing an identifiable model with limited human data.
B. Control & Robotics Applications of Computational Trust Models
Despite a handful of literature studies on trust modeling, the majority of the work focuses on building and learning the model itself and do not necessarily explain the utilization of the computational trust models in various application domains, especially how trust quantification can be utilized in robot control designs. Some initial attempts have been made to integrate trust models into robot velocity and force control. For example, Sadrfaridpour and Wang implemented an NMPC for a manipulator end-effector velocity control taking into account trust constraints to improve the human operator's interaction experience with the robot [140]. Considering trust as a hint of the human operator's intent, Sadrfaridpour et al. also studied trust-based impedance control for human-robot co-manipulation, switching the robot behavior and the corresponding impedance control laws depending on its corresponding trust level [177]. To enhance human acceptance in the haptic teleoperation system, Saeidi et al. developed trust-based shared control for mobile robots, which dynamically adjusts the control authority to the robot based on trust computation [206]. To guarantee appropriate human utilization of the proposed interface, Vinod et al. formulated a human-automation system as a multiple-input-multiple-output (MIMO) LTI system and cast the user interface design problem as selecting an output matrix where the level of human trust was used as a constraint [207]. However, no explicit trust model was considered in this work. To design more flexible safe control considering different individuals, Ejaz and Inoue [208] modeled the trust of an ego vehicle towards surrounding pedestrians as a safety margin indicator and used the trust computation to influence control barrier functions16 designed to constrain the ego vehicle's states in a safe set while the ego vehicle ran the MPC algorithms. Note that this work considered robot-to-human trust instead of human-to-robot trust.
However, we believe there is much space in control system theory and technology to be explored related to the computational trust models. For instance, both time-series models with inputs, states, and outputs and value-based (not symbol-based) DBN models that describe a time-series effect can be converted into the state-space form [129], [210], which can further enable system analysis and controller design in control theory. The trust dynamics may also be coupled with the human-robot physical interaction dynamics in an augmented state-space form, where control theory can be leveraged to analyze the overall human-robot system. Moreover, trust may be integrated into the reachability and safety analysis of the HRI system, using approaches such as control barrier function (CBF) [209] and Hamilton-Jacobi (HJ) reachability [211]. On the one hand, by conceptualizing trust as a state of interest, the reachability and safety analysis can proactively safeguard against overtrust and/or undertrust. On the other hand, by relating trust to human's perception of the uncertainty and reliability of the interacting robot, we may adjust the rigidity of the safety/reachability control laws to achieve higher flexibility and efficiency [208]. Various robot control designs can be developed to regulate or optimize the system's performance. The rich literature in control designs for physical HRI [212], [213], [214], [215] can also be leveraged and extended. In addition, we think the trust calibration process may be cast as a set point regulation (assuming static trustworthiness), tracking (dynamic trustworthiness), or MPC (extending the framework in [109] by adding robot dynamics constraints) problem by controlling the robot inputs.
Furthermore, many works on trust measures and qualitative analysis suggested meaningful ways to understand and utilize trust for autonomous systems and robots. We expect these works to benefit even more with a computationally trust model that has the prediction ability. For example, quantified trust value can be used to determine what information to be presented through human-machine interface (HMI). One objective of HMI is to calibrate users' trust to its appropriate level. Knowing the dynamic evolution of the trust is crucial because it determines the useful information to be presented through HMI [64]. Trust can be used to determine the control allocation and motion planning of robots. Based on the study of Muir et al. [40], human trust is directly related to the use of automation systems. A lot of literature quantifies the level of human operator trust in their experiments based on this argument [96]. In such experiments, the automation/robot will provide at least autonomous and semi-autonomous modes. The trust level of the human operator is considered to decrease if he/she chooses to switch away from autonomous mode and vice versa. A possible solution is to adjust robot behaviors based on the predicted trust level. For instance, whenever trust is too low, robot behaviors can be changed to gain trust. Whenever trust is too high, the robot may deliberately perform untrustworthy behaviors and check if the human is overtrusting.
C. The Role of AI in Human-Robot Trust
Despite the prevailing attention to AI and deep neural networks in recent years and previous work on neural network-based human trust models in automation [32], neural network-based human-robot trust models are surprisingly rare in the literature. Soh et al. proposed a recurrent neural network (RNN)-based model to capture the task-specific trust dynamics [28]. This is the only neural network-based trust model we could find in the literature. The human subject experiment in this work showed the RNN-based model performed better in predicting trust on unseen tasks than the GP models summarized in Section III-D (from [28]) in automated driving scenarios. The RNN model has high flexibility in processing task and performance representations. Compared with the GP models, it can generate more accurate trust predictions.
We think this may be a new direction with a lot of potential in the future, especially where enough data is available for model training and learning. Neural network-based trust models offer a powerful approach to utilizing data to model complicated nonlinear relations between trust impacting factors and trust variables. Due to the strong computation capability of deep learning, a neural network-based trust model can potentially include a wide range of trust impacting factors as feature inputs to the neural network, far beyond what other types of models in the literature can handle. Furthermore, RNN can be used to capture the temporal dynamic behavior of the trust evolving process. But what has severely limited the prosperity of neural networks in the domain of trust models?
In general, data-driven AI models belong to black-box models. Compared to time-series and DBN trust models, the process lacks interpretability for decision-making based on observations and transparency needed for HRI. In addition, there lacks systematic collections of trust data. The current situation in trust modeling is very similar to the era of computer vision before ImageNet [216] when diverse hard-coded image processing algorithms dominated the field and neural networks remained silent. No standardized datasets are available for trust modeling. The measurement metrics vary across different laboratories and across different projects. Trust measurements are also in small quantities, expensive to collect, and seldom shared and reused in the community. Creating an online platform to host, administrate, and redistribute the data might benefit the entire research community. Furthermore, data may be biased or skewed. To fully leverage the strength of data and machine learning, neural networks-based trust models may be combined with the above models and human knowledge.
Lastly, human trust in AI-enhanced robots may be another new research direction. Recent advancements in artificial intelligence and deep learning enable rapid developments in robots as embodied AI agents. On the one hand, the powerful performances of AI-enhanced robots will likely improve human trust and hence the willingness to use and collaborate with robot partners. On the other hand, due to the lack of interpretability, AI-enabled robot behaviors may be unpredictable, making it challenging to build trust. Improving the transparency and explainability of AI-enhanced robots may mitigate this issue [217]. Furthermore, human trust in AI, robots, and robots as embodied AI may differ, which deserves scrutiny to realize trustworthy AI. One may also consider developing artificial robot-to-human trust. For example, Azevedo-Sa et al. created a bidirectional human-robot trust model and the utilization of non-performance-based metrics allows the proposed model to predict a robot's trust in humans [170]. Freedy et al. developed a measure of “goodness” for a specific operator [66]. A relative expected loss score (REL) was established to collect the observed human task allocation decision behavior, risk, and observed robot performance. Ejaz and Inoue [208] modeled vehicle trust to pedestrians. The momentary trust was the linear combination of pedestrians' phone usage, eye contacts, and pose fluctuation. Values of these impacting factors were extracted from images using neural network tools. Since trust is mutual, we may study the dyadic relation between human-to-robot trust and artificial robot-to-human trust [218]. Some initiatives in human-robot mutual trust modeling in the literature include [170], [206], [219], [220].
Conclusion
With the fast and widespread adoption of robots in almost every section of our daily lives, it is pivotal to understand and analyze how humans feel about robots, especially the level of belief an individual would place in a robot, in order to design robots and control their behaviors in a way that we are willing to utilize and collaborate with. Therefore, in this paper, we attempted to survey some most representative computational models of human-trust-to-robot. We first investigated the definition of human-robot trust compared to interpersonal, social, human-automation, agent-agent trust, and trustworthiness. We also summarized the trust-impacting factors, current quantitative trust measurement approaches, and their corresponding trust scales. We categorized the existing computational trust models into five types: performance-centric algebraic, time-series, (PO)MDP-based, Gaussian-based, and DBN-based, trust models. Detailed formulations of these trust models were elucidated with an analysis of their advantages and shortcomings. Despite their respective advantages and limitations, these trust models may be applied to robot control, task allocation, motion planning, etc., in human-centric frameworks with enhanced HRI and situational awareness. We also provided our observations on the limitations of the extant trust models and discussed possible future research directions to fill these gaps. Given that the field of computational trust model is still in its early stages of development, the primary aim of this paper is to offer a comprehensive overview of the current state-of-the-art in computational trust models. As a future work, a rigorous guideline for trust model design/selection based on the specific task can be very promising.