Summary
This article provides a perspective on estimation and control problems in cyberphysical human systems (CPHSs) that work at the intersection of cyberphysical systems and human systems. The article also discusses solutions to some of the problems in CPHSs. One example of a CPHS is a close-proximity human–robot collaboration (HRC) in a manufacturing setting. The issue of the joint operation’s efficiency and human factors, such as safety, attention, mental states, and comfort, naturally arise in the HRC context. By considering human factors, robots’ actions can be controlled to achieve objectives, including safe operations and human comfort. Alternately, questions arise when robot factors are considered. For example, can we provide direct inputs and information to humans about an environment and the robots in the area such that the objectives of safety, efficiency, and comfort can be satisfied by considering the robots’ current capabilities?
The article discusses specific problems involved in HRC related to controlling a robot’s motion by taking the current actions of the human in the loop with the robot’s control system. To this end, two main challenges are discussed: 1) inferring the intention behind human actions by analyzing a person’s motion as observed through skeletal tracking and gaze data and 2) a controller design that keeps robot motion constrained to a boundary in a 3D space by using control barrier functions. The intention inference method fuses skeleton-joint tracking data obtained using the Microsoft Kinect sensor and human gaze data gathered from red-green-blue Kinect images. The direction of a human’s hand-reaching motion and a goal-reaching point is estimated while performing a joint pick-and-place task. The trajectory of the hand is estimated forward in time based on the gaze and hand motion data at the current time instance. A barrier function method is applied to generate safe robot trajectories along with forecast hand movements to complete the collaborative displacement of an object by a person and a robot. An adaptive controller is then used to track the reference trajectories using the Baxter robot, which is tested in a Gazebo simulation environment.
Graphical Abstract

The prospect of a collaborative work environment between humans and robotic automation in a manufacturing setting [1] provides the motivation for finding innovative solutions to human-in-the-loop control for safe, efficient, and trustworthy human–robot collaboration (HRC), or HR interaction (HRI), in cyberphysical human systems (CPHSs) [2], [3]. Studies in [4] show that collaborative automation can be beneficial to 90% of approximately 300,000 small-to-medium-scale enterprises in the United States. In the paradigm of human-centered automation [5], human safety, ergonomics, and the collaborative efficiency of the work are given the utmost importance. Traditional methods to ensure the safety of humans around factory robots involve the use of cages. Recent work looked beyond cage-based safety to provide robot control and sensing-driven solutions for human safety around robots [6]–[8].
The purpose of this article (see “Summary”) is to provide a tutorial on human-in-the-loop estimation and control methods to achieve worker safety in the context of close-proximity HRC. Examples of current close-proximity interaction include collaborative assembly, the cooperative carrying of loads, and robot assistants working near humans in manufacturing plants in industries such as automotive, aerospace, and electronics.
The article is broadly divided into two main sections. The first describes methods for human-action intention inference based on sensor data, and the second describes safe robot-control policy generation based on an inferred human-action intention. State-of-the-art methods in both human-action intention and safe robot control are first discussed, and the mathematics behind them are provided in adequate detail (supported by simulation and experimental results in certain cases). An overview of intention estimation is first provided by presenting existing literature in a tutorial manner, followed by the approaches that focus and expand on prior work. Modifications to the algorithms are mentioned at the appropriate locations. New experiments are conducted, and the results are added to the manuscript.
Compared to the prior work in [9], a robot control design method for generating safe robot-reference trajectories is discussed that uses control barrier functions (CBFs) in the HRC context. Experiments are conducted using a Kinect sensor and the Baxter research robot platform, where a human’s trajectory forecast is employed to generate safe courses for robots to follow.
A block diagram of human-in-the-loop control with human-intention estimation and safe trajectory tracking of a robot for collaborative tasks is shown in Figure 1. Use cases of such collaborative tasks include the following:
robot-assisted assembly, where different components require various installation methods (distinct models); a robot observes a human to determine the current model and hence the robot’s own desired actions
wire harness assembly, in which a human and a robot complete certain steps in the fabrication process; for instance, for a wire stripping task, a human may reach for a tool, indicating to a robot that it must grasp wire and take it to the human
assisted construction/surgery, where a robot is responsible for acquiring the correct tool for the task the human is currently trying to perform
collaboratively moving loads (an experiment is performed for this use case)
the repair of vehicles in uninhabitable environments (such as underwater and in space), where a robot may assist a human who has limited ability given the conditions, and the robot has to interpret the human’s intention from motion profiles.
FIGURE 1.

A block diagram of human-in-the-loop control, with human-intention estimation and safe trajectory tracking for robots participating in collaborative tasks. The human dynamics block uses skeletal tracking via red-green-blue-depth sensors. Based on the skeletal tracking of one arm, the human trajectory estimator generates an approximation of a person’s reaching action intentions. The robot-reference trajectory is produced using the estimated human trajectory so that the robot’s movement is safe.
INTENTION ESTIMATION
Providing sensing-based human-action intention inference solutions is one of the important steps toward achieving safety during HRC in manufacturing automation, automotive applications (self-driving cars) [10], space robotics [11], and assistive robotics [12]–[16]. Studies in psychology suggest that when two humans interact, those people infer one another’s intended actions for safe interaction and collaboration [17], [18]. An optimal control model of human response and its applications is examined in [19] and [20]. Based on these studies about the principles of human interactions, the safety, operational efficiency, and task reliability in HRC could be greatly improved if robots were provided the capability to infer human-action intentions. For instance, in [12] and [21]–[23], inferring a human partner’s intention is shown to improve the overall performance of tasks requiring HRC.
Human intention is inferred via various sensing modalities by measuring different cues captured by sensors that are on a person’s body (wearable sensors) or mounted elsewhere (nonwearable sensors). Using wearable devices, physiological information measured through heart rate, skin response [24], [25], and electromyography sensors [26] is typically used for gauging human motion intention. In [24], human intention is represented using valence/arousal characteristics that are measured by physiological signals, such as heart rate and skin response. The valence/arousal representation of human intention indicates only the degree of approval to a given stimulus. Using nonwearable sensors, cues such as human emotion [27], approval responses [24], body posture [28], gestures [29], eye gaze [30], [31], facial expressions [32], and skeletal movement [33], [34] are measured. Imaging sensors, such as red-green-blue (RGB) cameras, are also commonly used. In [35], a human’s intention to hand over an object is predicted by key features extracted from an RGB camera sensor.
For estimating human intention, the dynamics of human motion are modeled using probabilistic representations, where hidden Markov models (HMMs) [24], [36], [37], dynamic Bayesian networks [38], [39], growing HMMs [40], and conditional random fields (CRFs) [41]–[43] are employed. In [44], human activities are inferred by an HMM of people’s actions and interactions with autonomous mobile robots. In [45], a human intent estimation algorithm based on a fuzzy inference engine is presented. The intention-estimation problem is formulated only as a relationship between attention and physiological measurements, and online inference of human intention is not performed.
In [24], an online algorithm to estimate the affective state of a person is developed based on valence/arousal representation using an HMM. The intention in these methods is formulated as a classification problem. Since the future state prediction in methods based on HMMs and their variants is dependent only on the current state, these approaches quickly react to changes in intent. However, if the change in intent is not frequent, the Markov assumption can be overly restrictive, as it prevents these algorithms from becoming more certain of human intent through additional observations [46].
In another popular paradigm, human motion is represented using continuous/discrete dynamics that are parameterized using dynamic neural networks (NNs) in the case of deterministic modeling with noise and through Gaussian processes (GPs) and Gaussian mixture models (GMMs) in the case of probabilistic modeling. In [23], an NN is used to approximate the endpoint of a human hand for representing human motion and intent in physical HRC applications. In [33], a latent variable representation called the intention-driven dynamic model (IDDM) is proposed to infer intentions from observed human movements. Robot table tennis and human activity classification are demonstrated using a belief propagation algorithm coupled with the IDDM. In [47], a human-intention inference algorithm is developed using unsupervised GMMs, where the parameters of the GMMs are learned using the expectation-maximization (EM) algorithm.
The framework presented in [47] provides an unsupervised online learning algorithm, while the algorithms presented in [33] do not involve online learning. The aforementioned approaches use the entire observed trajectory in the prediction of future states. Thus, the certainty in the estimated intent tends to converge through time. These models are typically not good for detecting sudden changes in intention. Additionally, the models developed through the preceding methods are not all learned, considering the stability and convergence to the reaching point of human actions/motion.
In yet another paradigm, action plans are represented as policies in terms of state-action pairs. Inverse optimal control (IOC) and inverse reinforcement learning (RL) algorithms are used to model the intention-driven behavior, where the intended motion maximizes an unknown objective or reward function. An inverse linear quadratic regulator approach is developed in [48] to predict the intent and trajectory forecast of human motion. The applicability of human-intention estimation using inverse approaches can be found in various applications. For example, human motor-control intent estimation in rehabilitation applications is presented in [49] using an inverse model predictive control strategy. Predicting pilots’ behavior by modeling their goals through RL and game theory is presented in [50]. In [51], human motion during collaborative manipulation is predicted via an IOC approach. The IOC-based techniques typically require exploration of the state space; thus, they require large amounts of data for converging to the correct solutions. A categorization of intention estimation algorithms is given in Table 1.
TABLE 1.
A categorization of the intention estimation algorithms based on intention estimation problem formulation.
| Method | Sensor Type | Reference | |
|---|---|---|---|
| Intention parameter estimation | Neural network (NN) | Red-green-blue depth (RGB-D) | Ravichandar et al. [34] |
| Gaussian process (GP) | RGB | Wang et al. [33] | |
| Gaussian mixture model (GMM) | RGB-D | Mainprice and Berenson [52] | |
| NN | Force/torque | Li et al. [23] | |
| Action recognition and intention classification | Interacting multiple model | RGB-D | Ravichandar et al. [53] |
| Hidden Markov model (HMM) | Physiological | Kulic et al. [24] | |
| Anticipatory temporal conditional random field (CRF) | RGB-D | Koppula et al. [54] | |
| HMM | RGB + laser | Kelley et al. [44] | |
| HMM | Motion capture | Ding et al. [37] | |
| Hybrid dynamic Bayesian network (DBN) | RGB | Gehrig et al. [38] | |
| Hybrid DBN | Agnostic | Schrempf et al. [39] | |
| Growing HMM | RGB | Elfring et al. [40] | |
| Linear-chain CRF | RGB-D | Hu et al. [43] | |
| GP-latent CRF | RGB-D | Jiang et al. [42] | |
| GMM | Motion capture | Luo et al. [47] | |
| Inverse optimal control/inverse reinforcement learning (IRL)-based intention estimation | Path integral IRL | Motion capture | Mainprice et al. [51] |
| Inverse linear quadratic regulator | RGB-D | Monfort et al. [48] | |
| Inverse model predictive control | Encoder | Ramadan et al. [49] | |
| Miscellaneous | Fuzzy logic | Physiological | Kulic et al. [45] |
| Fuzzy logic | Force/torque | Carli et al. [36] |
The intention is defined as a goal location of reaching motions and an associated trajectory forecast. Knowing an estimate of the goal location is sufficient in many applications. In such cases, a maximum likelihood (ML) estimator can be used. When a distribution of the intention is required, a maximum a posteriori (MAP) estimator can be used. Thus, two approaches to human-intention estimation are discussed that are based on ML and MAP estimation techniques. The ML estimator requires more data, compared to the MAP estimator (which uses a priori information). When the priors are uniform, the MAP estimator gives an ML approximation. Hence, ML estimation is a special case of MAP estimation.
In the first technique, the goal-reaching intention is modeled as a parameter of the continuous dynamics. An ML estimation technique called the approximate EM algorithm is applied to estimate the goal-reaching intention. To accommodate changes in motion, an online method for NN weight learning is also developed. NN modeling of human motion can also accommodate personalized factors, such as age and other physical characteristics. The algorithm is called the adaptive neural intention estimator (ANIE) [34]. In addition to the ANIE algorithm presented in [34], gaze information is included to provide a more accurate initialization to the EM algorithm. In the second technique, individual goal-reaching motions are represented as multiple dynamical system (DS) models. A MAP estimation of intention is obtained through an interacting multiple-model (IMM) framework that computes probabilities for each goal-reaching motion model based on observations. Changes/switches in intention can also be detected using the IMM framework. A good prior to the IMM algorithms can improve their performance. Hence, additional cues, such as the direction of a person’s gaze, computed by estimating the head orientation from RGB images are used. The algorithm is called the gaze-based multiple-model intention estimator (G-MMIE) [9].
In the first method, based on ML estimation, the complex dynamic motion of the human arm is represented by a nonlinear system model. The positions and velocities of the joints are used as the states. The uncertain system dynamics are modeled using a dynamic NN [55] to represent the state propagation. Intentions are modeled as the goal locations of human arm-reaching motions, which are represented by the unknown parameters of the state-space model. The issue of intention inference is solved as a parameter inference problem, given noisy motion data through an approximate EM algorithm [56]. The NN estimation can potentially enable the consideration of user- and object-specific characteristics, such as the size and the shape of the object, to be included as a part of the dynamics.
Different humans may reach the same point in 3D space in various ways based on their physical characteristics. This brings a challenge to using the model learned from the demonstration data to represent joint position trajectories of other subjects. One way of updating the representation in real time is to use the EM algorithm by optimizing the Q function across the model parameters, along with the intention. A closed-form expression for the model update using EM exists if the model is linear or represented using a radial basis function (RBF) NN [57]. However, arm-motion dynamics are highly nonlinear, and the RBF may not always be the best choice for the basis functions of the NN to represent human movement. To overcome this challenge, an identifier system-based algorithm presented in [58] is used for online model updates. The identifier system is designed using a robust feedback term: the robust integral of the signum of the error (RISE) [59]. Based on the Lyapunov analysis, the parameter update laws for the model update are derived using the error between the state estimate generated by the identifier system and the state estimate from the original system model. The analysis ensures the asymptotic convergence of the state identification errors and their derivatives between the learned model and the true model.
The guarantees of online learning can be very useful in tasks where the training data are limited and predictions have to be made about new users with varying motion dynamics in novel environments. For instance, consider a wire harness fabrication task in a manufacturing environment, where the NN is trained using data obtained from a user assembling parts to build an object. If the trained NN is employed for a new user handling parts of a similar object, the NN approximation error is likely to be high, as the motion profiles will vary for different users assembling dissimilar parts. However, as new data become available, the presence of the feedback term enables the identifier system to implicitly learn the network weights and minimize the effects of NN approximation errors in real time [58]. An inference algorithm is then used with the updated model for early prediction of the intentions.
In the second method, based on MAP estimation, the organizing principles of motion generated in humans are used. The principles state that the generated motion is inherently closed-loop stable and smooth during various tasks [60]. The problem of learning the human arm’s motion dynamics is considered. It is formulated as a parameter learning problem under goal convergence constraints (derived using a contraction analysis of nonlinear systems [61]) that aid in learning stable nonlinear dynamics with respect to the reaching goal location. Details of the learning algorithm can be found in [62]. Such modeling of human motion is also useful when a person is represented as a point and his or her motion in the 2D/3D space is observed as a point motion reaching to varying locations (for instance, factory workers moving to different work stations).
Using the learned model, the intention inference is completed by the multiple-model intention estimator (MMIE) presented in [63]. The MMIE algorithm employs an IMM filtering approach in which the posterior probabilities of candidate goal locations are computed through model-matched filtering (see [64]). When the number of models is very high, it is well known that the performance of the IMM filter degrades. A variable-structure IMM (VS-IMM) filter can be applied in such cases [65]. A limiting case of the VS-IMM when the mode space is continuous is presented in [66]. In this article, probability priors of the finite number of available models are computed using a gaze-based prior computation that helps in reducing the number of possible candidates. A set of demonstrations capturing human arm joint position trajectories for reaching motions is collected by an RGB-depth (RGB-D) camera (Microsoft Kinect). Each recorded joint position trajectory is labeled according to the corresponding true intention, that is, the 3D goal location of the reaching motion. NN models are learned by the labeled demonstrations of the joint position trajectories.
HUMAN-IN-THE-LOOP SAFETY CONTROL OF ROBOTS
When a robot is collaborating with a human in close proximity, one of the important problems is to make the robot aware of the human’s movement intentions so that the robot can adapt its motion controller to operate safely and perform a collaborative task. In [21]–[23], inferring a human’s intention is shown to improve the overall performance of collaborative tasks. In [67], human goal intention is applied in conjunction with an admittance controller to achieve HRC. Many contributions to HRI have targeted motion-intention estimation for tasks where direct physical interaction between humans and robots is involved. For control strategies, the literature focuses on designing impedance control laws [12], [23] and admittance control laws [67], [68] for adapting the interaction forces exerted by a human on a robot when the robot is physically interacting with the person. In [69], a controller is developed for HR handover interaction based on dynamic movement primitives. In HRC (where there may not be direct contact with a human), the human motion intention estimation, robot path planning, and control design become a more important technical challenge.
There are studies in the literature that address the problem of robot motion control when humans are present in the vicinity of robots and autonomous systems [46], [70]. Most of these studies view the problem as a collision avoidance issue and solve it using the potential field approach [71]. These control actions are purely reactionary in nature. Anticipatory skills are required to improve efficiency and safety during collaborative tasks when humans work in close proximity to robots [46], [70]. To achieve proactiveness, the controller and motion planner must incorporate probabilistic information about the possible human actions. In [46], predictive modeling of pedestrian motions with changing intentions is proposed to plan safe robot trajectories. Integrated estimation and control for HRI involving an industrial robot was proposed in [71]. In [23], a feedback controller for HRI is developed based on an NN model of human intention. In [52], a stochastic trajectory optimizer for motion planning is employed for planning robot arm motion based on human intentions.
In [72], scheduling, planning, and control algorithms are presented that adapt to the changing preferences of a human coworker while providing strong guarantees for the synchronization and timing of activities. In [73], new hierarchical planners based on hierarchical goal networks are developed for assembly planning in the HR team. In [74], an empirical study of human–human interaction is conducted to investigate the ways in which human teammates perform in a coordinated behavior. This study is targeted toward scheduling and planning tasks rather than studying human behavior in the context of robot motion planning for safety and the proactive control of close proximity operations.
In control literature, stability studies of human-in-the-loop adaptive controllers are presented using the inner–outer loop control structure in [75]. Stability examinations of human-in-the-loop telerobotics with a time delay are presented in [76]. These works do not explicitly consider safety aspects of the human-in-the-loop systems. Providing safety guarantees in the learned controller of the machine/robot is typically achieved by adjusting the reference command via a prefilter called a reference governor [77]–[79] and by using optimal control under uncertainty in a differential game setting.
In [80], an RL method that guarantees stability and safety by exploring the state space to collect new data for learning is presented. In [81], a safe, online, model-free approach to path planning with Q-learning is discussed. A general safety framework for learning-based control using reachability analysis is introduced in [82]. In [83], a receding horizon safe-path planning approach using mixed integer linear programming is presented. Safe trajectory generation for autonomous operation of spacecraft using convex optimization formulation is proposed in [84]. When the region is non-convex, successive convexification can be performed [85]. A detailed survey of, and tutorial on, an -adaptive control architecture for safety critical systems appears in [86].
Other methods of achieving the safety property of controller synthesis are to employ a BF/certificate or a CBF, which ensures that the closed-loop system’s trajectories remain inside a prescribed safe set [87]. There are two candidates to construct BFs, namely, reciprocal BFs and zeroing BFs. Reciprocal BFs can be of the inverse type and the logarithmic type. Similar extensions to CBFs have also been developed in the literature. Applications of BFs and CBFs in many autonomous robotic systems (such as robot manipulators, autonomous vehicles, and walking robots) are shown in [88]–[91]. In [88], [90], and [92], BFs were successfully applied to DSs, where ensuring safety conditions is critical. In [92], time-varying BFs and CBFs for avoiding moving and static obstacles are derived, and their application to quadcopters that avoid unsafe obstacle regions is shown. Robustness properties of the CBFs are studied in [93], which shows that if a perturbation (or model error) makes it impossible to satisfy the invariance condition for a reciprocal BF, then the solution of the model must cease to exist because the control input becomes unbounded.
For zeroing CBFs, input-to-state stability results hold in the presence of model uncertainties. A concept of exponential BFs and CBFs is introduced in [94]. The method of CBFs is extended to position-based constraints with relative degree two in [95] to address the safety constraints for systems with a higher relative degree. Furthermore, a backstepping-based method to design CBFs with a higher relative degree is also introduced. However, achieving a backstepping-based CBF design for systems with a higher relative degree is challenging. In [94], a concept of exponential CBFs is presented that can handle state-dependent constraints for systems with a higher relative degree. In [96], a safety-aware RL framework using BFs is proposed. However, the application of BFs and CBFs to HRC is new, and there are still many unaddressed technical challenges. In the recent work presented in [97], a methodology is proposed to learn system dynamics that can be harnessed to generate a robot’s desired trajectories, which are strictly bounded within a prescribed safety set. An adaptive controller that accepts the human-intention estimation in the loop for trajectory synchronization appears in the recent work in [9].
In this article, a safe trajectory generation algorithm that uses the inferred human trajectory from the intention estimator is developed. The safe trajectory is then provided to the robot for tracking via a feedback adaptive controller for the nonlinear Euler–Lagrange (EL) system that adapts to the uncertainties in the model. EL dynamics are widely used to represent the dynamics of robot manipulators. In the controller stability analysis, errors due to the inferred reference trajectory of the robot (which is generated using human trajectory forecasts computed through the intention estimation algorithm) are considered. The controller is shown to be asymptotically stable under disturbances and uncertainties. A robot with an understanding of human intent is more equipped to act to improve the safety of joint tasks. An experiment is completed that computes the intention of a person performing a collaborative task of moving an object with a robot. To this end, a safe reference trajectory is generated for the robot, and an adaptive controller is designed to track the estimated safe reference trajectory.
There are many broader challenges in human-in-the-loop robot control from the perspectives of both estimation and control. The major challenges include the development of fundamental theoretical guarantees and limits on long-term intention estimation, using predictive forecasting methodologies for learning models of humans and/or robots, and the fusion of multiple sensing modalities (such as vision, ultrasound, and lidar in the context of autonomous vehicles and/or physiological signals in the context of manufacturing applications). Incorporating human trust, workloads [98]–[102], attention allocation [103], and cognitive factors [104] in the HR collaboration brings many potential avenues for research. Including human ergonomic preferences in CPHSs also has great potential to make the HRI more comfortable for humans [105], [106].
Some recent trends in control and decision making in HR collaborative systems are found in [107]. For the ML-based intention estimation technique of applying the approximate EM algorithm, various approximations for the E step can be considered to improve the performance of the estimator when the process and sensor noise characteristics are non-Gaussian. Integrating information provided by intention inference algorithms into safe control methodologies that use CBF formulation and address the uncertainty in estimation is another important area that requires further investigation. In addition to safety confirmation, there are temporal properties that need verification in the HRC setting, such as eventuality, avoidance, reachability (a problem related to the safety property), and the composition of these properties. Developing methods for corroborating these properties in the HRC context is also important.
Compensating for time delays during the communication of information between humans and robots brings many challenges, specifically in the context of networks of human and robot agents [108]. HRC in a virtual reality setting also presents many obstacles from an estimation and control perspective. For example, in a rehabilitation application, a human who is undergoing walking or biking rehabilitation can benefit from interactions with a virtual environment. The interaction with the virtual setting can provide important feedback to the human user via the robot he or she is interacting with.
The remainder of the article is organized as follows. First, a human intent estimation method based on ML estimation using an approximate EM algorithm is presented. Then, another intention approximation method based on a MAP estimator that employs an IMM filter (with priors computed using the human gaze map) is described. A human-in-the-loop control strategy that uses CBFs to compute a safe reference trajectory for a robot to follow is designed, and an adaptive robot control is developed for a robot manipulator to track a safe desired trajectory during a collaborative task. Simulations and experiments are conducted to validate the performance of the proposed intention estimation and control results. Note that, in the subsequent development, the dependency of variables on time is dropped for the compactness of notation unless it is necessary for clarity.
HUMAN-ACTION INTENTION ESTIMATION SCHEME
Intention Estimation as a Machine Learning Estimation Problem
In this section, the human goal-reaching intention is modeled as a parameter of the nonlinear dynamics. The motion of a human is modeled as a nonlinear differential equation approximated as an NN. An approximate EM algorithm is designed to estimate the reaching intention of the human’s action (see Figure 2). To facilitate the discussion, a problem scenario is first described.
FIGURE 2.

A block diagram of the approximate expectation-maximization (EM)-based intention estimation algorithm with gaze-based approximates that are used for initializing the M step of the EM. The gaze estimator block uses red-green-blue (RGB) images along with skeletal data to estimate the gaze map of a person, which determines the most probable objects that the person is looking at. The gaze map is applied to provide an initial estimate of the parameter for the EM algorithm. CNN: convolutional neural network; EKF: extended Kalman filter; ANN: artificial neural network.
Problem Scenario
Consider a 3D workspace with a human performing tasks, such as picking up objects placed on a table or a shelf or walking toward certain locations in the area. The human subject reaches out to different objects placed on a table, and a robot watches the person through a 3D camera sensor. The problem of inferring the human hand’s reaching goal location is addressed. Since the human motion is highly nonlinear and uncertain, an NN approximation of the nonlinear function of the dynamic system is used to model the movement. The NN is trained using a data set containing skeletal tracking of a human reaching for predefined target locations in a given workspace, observed using an RGB-D camera. When a set of new measurements becomes available, the trained NN is employed to estimate the reaching goal intention (the goal location in 3D) via an approximate EM algorithm. An online NN model weight learning method is also developed through an identifier-based algorithm to adapt to the variations of motions in different human subjects.
Human Motion Dynamic Model and Measurement Model
The dynamics of human arm motion are modeled using a continuous nonlinear dynamic model of joint positions, velocities as states, and intention parameters represented as the reaching goal location of the motion. The human intention is denoted by , where , and represents a 3D location of an object on a table. The true intention g is one of the goal locations , which can be finite (represented by a discrete variable) or very large (represented by a continuous variable). The state represents the positions and velocities of four points on the arm (shoulder, elbow, wrist, and palm) that describe the motion of the arm in the robot-reference frame, and denotes the measurement obtained from the RGB-D camera sensor data. The modeling of g as a continuous variable would be suitable in scenarios where it is not possible to obtain all possible object/goal locations.
State Transition Model
The state transition model is described by the following equation:
| (1) |
where is a zero-mean Gaussian random process with a covariance matrix . Let be a compact, simply connected set of . Here, is a Lipschitz continuous function. Since an explicit form of a nonlinear function for the human arm motion is not known, is approximated using a feedforward NN. There exist weights and biases such that the function can be represented by a three-layered NN as
| (2) |
where is the input vector to the NN; and are the bounded constant ideal weight matrices, that is, and ; is an activation function that can be represented by a vector sigmoid function, an RBF, or a rectilinear unit; is the ith element of the vector ; is the function reconstruction error; and is the number of neurons in the hidden layer of the NN.
Measurement Model
The measurements of human arm joint positions are gathered by using an RGB-D camera sensor. The measurements are obtained in the camera’s reference frame. Let pc(t) = (xc(t), yc(t), zc(t))T be a point in the camera reference frame and pr(t) = (xr(t), yr(t), zr(t))T be a point in the robot-reference frame. The points pc(t) and pr(t) are related by
| (3) |
where and are the rotation matrix and the translation vector, respectively, between the robot-reference frame and the camera reference frame. The camera sensor measures the 3D locations of the skeleton’s joints.
The measurement model is given by
| (4) |
where is the position of the skeletal joints of the arm in the camera reference frame, h(x(t)) = Hx(t) + b, , , and is zero-mean Gaussian noise with a covariance matrix . The measurement noise υ(t) is assumed to be independent of the process noise ω(t) defined in (1). The measurement model of the shifted measurement vector z(t) = y(t) − b is given by
| (5) |
For the RGB-D camera, the Gaussian assumption of the measurement noise υ(t) is a standard assumption in the literature [109], [110]. The measurement noise of zc(t) components of y(t) is assumed to be normally distributed in [109], and the variance is modeled as a quadratic function , where , , . The other components of Σz can be calculated from the calibration parameters; individual pixel variances , ; and depth variance .
Neural Network Model Training
The training of the NN is completed using the data consisting of the human arm’s joint locations, velocities, and accelerations along with the reaching goal locations. The baseline NN is trained through Bayesian regularization. The objective function applied to train the NN using Bayesian regularization is given by , where , are the estimated NN weight matrices, is the sum of squared errors, yi(t) is the target output, ai(t) is the network’s output, is the sum of the squares of the NN weights, and Kα and Kβ are the parameters of regularization that can be used to change the emphasis between reducing the reconstruction errors and the model complexity, respectively.
Approximate Expectation-Maximization Algorithm for Estimating the Intention
An approximate EM algorithm is presented to estimate the intention parameter [56] using the state transition model learned through the NN. The intention inference algorithm on an offline trained model is presented first. The extension of the intention estimation algorithm with the online model learning is discussed subsequently. We start by discretizing the continuous model using a first-order Euler approximation. It is assumed that the states are sampled at a high rate so that the first-order Euler approximation is valid. The discretization of the state transition model defined in (1) yields
| (6) |
where f(x(t − 1), g) = x(t −1) + WTσ(UTs(t − 1))δt and δt is the sampling period. Let ZT = {z(t = 0), …, z(t = T)} be a collective set of observations and XT = {x(t = 0), …, x(t = T)} be a collective representation of states from time t = 0 to t = T. To infer the intention, the posterior probability of ZT given the intention g is maximized using an ML criterion. The process noise of the discretized system in (6) is given by Q = δt2 Qc.
The log-likelihood function of the intention g is given by
| (7) |
which can be obtained after marginalizing the joint distribution p(XT, ZT|g) across XT. In general, analytically evaluating this integral is very difficult. In this article, an approximate EM algorithm is presented that uses state transition models trained using the NN. Given the fact that , the log likelihood defined in (7) is decomposed in the following way:
| (8) |
where is the expected value of the complete data log likelihood (given all the measurements and intentions), , is the expectation operator, and ĝ(t) is the estimate of g at time t.
It can be shown using Jensen’s inequality that H(g, ĝ(t)) ≤ H(ĝ(t), ĝ(t)). Thus, to iteratively increase the log likelihood, g must be chosen such that Q(g, ĝ(t)) ≥ Q(ĝ(t), ĝ(t)). The E step involves the computation of the auxiliary function Q(g, ĝ(t)), given the observations ZT and the current estimate of the intention ĝ(t). The M step involves the computation of the next intention estimate ĝ(t + 1) by finding the value of g that maximizes Q(g, ĝ(t)). The E step involves the evaluation of the expectation of the complete data log likelihood, which can be rewritten as
| (9) |
If v(t) and w(t) are Gaussian, the computation of Q(g, ĝ(t)) can be simplified. The M step involves the optimization of Q(g, ĝ(t)) across g, as described by
| (10) |
This step can be completed in two different ways, namely, numerical optimization and direct evaluation. One way to maximize the Q function is to use the gradient EM (GradEM) algorithm for the M step, which, in turn, uses the first iteration of Newton’s method [56]. Since Newton’s method often converges quickly, the local convergence properties of the GradEM algorithm are identical to the EM algorithm. More details of the convergence properties of the GradEM algorithm can be found in [111]. This method involves optimizing the Q function across . The update equation for ĝ, through the GradEM algorithm, is given by
| (11) |
where ĝ(t) is the estimate of g at the current time t of the optimization algorithm and and Δ(Q) are the Hessian and the gradient of the Q function, respectively. Note that numerical optimization methods need to run at every time step of the EM algorithm. For real-time implementations, the number of iterations for the optimization in (11) could be chosen based on computational capabilities. More details of the computation of the Hessian of the Q function and the expression for the gradient of the Q function can be found in [34]. General details of the EM algorithm are described in “Expectation-Maximization Algorithm.”
Another way to infer g is to evaluate the Q function for all possible instances of (the goal locations) in and obtain ĝ(t + 1), as described by the following expression:
| (12) |
This method involving direct evaluation of the Q function is feasible if all possible goal locations are known a priori and finite. This is not an unusual case in the context of human-intention estimation in practical applications, such as manufacturing assembly, space robotics, and assisted construction. Image processing algorithms, such as the region convolutional NN (R-CNN), faster R-CNN, mask R-CNN, and You Only Look Once, can be used to detect objects in the workspace and extract the 3D locations using camera data.
Online Model Weight Update
In this section, an online learning algorithm is described and used to update the weights of the NN model. The online learning of the NN weights is important to make the inference framework robust to variations in starting arm positions and various motion trajectories taken by different people. The NN weights are iteratively updated as new data become available. A state identifier is developed that computes an estimate of the state derivative based on the current state estimates obtained from the extended Kalman filter (EKF) and the current NN weights. The identifier state error is computed from the state estimate and the measurement. The error in the state identifier is used to update the NN weights for the next time instance. Note that the state identifier can run at a higher sampling rate compared to the EM algorithm, which involves optimizing the Q function that may have a slower convergence rate. Hence, the state identifier is presented in the continuous form. The identifier uses RISE feedback [59] to ensure the asymptotic convergence of the state estimates and their derivatives to the true values. The weight update equations are designed using Lyapunov-based stability analysis.
The state identifier is given by
| (13) |
where , , , is the current estimate of g from the EM algorithm, is the current identifier state, and is the RISE feedback term defined as [where is the state identification error]. Here, is the Filippov generalized solution [58] to the differential equation
| (14) |
where k, α, γ, and are positive constant control gains and sgn(·) denotes a vector signum function.
The weight update equations are given by
| (15) |
where proj(·) is a projection operator defined in [112]; Ûx(t) and Ûg(t) are the submatrices of Û(t) formed by taking the rows corresponding to , respectively; is the first-order derivative of the sigmoid function with respect to its input ÛTŝ(t); and Γw, Γux, and Γug are constant weighting matrices of appropriate dimensions. In the online learning algorithm, ĝ(t) from the EM algorithm is used. Hence, for the online learning step, ĝ(t) is assumed to be a known signal. The derivative of the intention estimate is computed through the finite difference method. It can be shown using Lyapunov analysis that the identifier defined in (13) and the update equations defined in (15) are asymptotically stable and that the state identification error converges to zero.
Gaze Map Computation
This section briefly describes the CNN introduced in [113], which is applied to extract gaze information from an RGB image. To this end, a deep CNN architecture is employed. The input (features) to the CNN is a Dw × Dh RGB image of the subject looking at an object and the relative position of the subject’s head in that image. The output is the gaze map of size Dw × Dh containing the probabilities of each pixel is the gaze point.
Data
The data set used for training the CNN model, as described in [113], is created by concatenating images from six different sources: 1548 images from Scene Understanding; 33,790 images from Microsoft Common Objects in Context; 9135 images from Actions40; 7791 images from Pattern Analysis, Statistical Modeling, and Computational Learning; 508 images from the ImageNet detection challenge; and 198,097 images from the Places data set.
Implementation of Convolutional Neural Network
The five-layered CNN shown in Figure 2 is implemented using Caffe library. Images of size 224 × 224 × 3 are used for training the CNN. These input images are filtered by 96 convolution kernels of size 11 × 11 × 3 and fed into the first convolution layer, of size 55 × 55 × 96. The output of the first layer is filtered with 256 convolution kernels of size 5 × 5 × 48 and fed to the second convolution layer. The subsequent three layers are connected to one another without any pooling layers between them. The third convolution layer has 384 convolution kernels of size 3 × 3 × 256 connected to the normalized and pooled outputs of the second convolution layer. The fourth convolution layer has 384 convolution kernels of size 3 × 3 × 192, and the fifth convolution layer has 256 convolution kernels of size 3 × 3 × 192. The remaining four layers used in the network are fully connected and of sizes 100, 400, 200, and 169. See [113] for a more in-depth description of the CNN framework.
The CNN is used to compute a gaze map, which is an image that assigns a probability value to each pixel to represent the likelihood of the pixel being looked at in the image, based on the head orientation data of a person. To compute the probability of gj being the initial goal location for the optimization problem of the EM-based approach, the average probability of the jth object in the scene is calculated as
| (16) |
where is the probability of the ith pixel being the gaze point, NPj is the number of pixels associated with the jth object, and is the set of all pixel locations associated with the jth object. Of all the objects in the scene, the one that corresponds to the highest average probability is chosen as the initial goal location for the optimization of the EM algorithm of the ANIE method.
Experimental Results for Expectation-Maximization-Based Intention Estimation
In this section, two experiments are presented using real data obtained from a Kinect sensor tracking a human’s movements. In both experiments, the reaching motion data harnessed for training and testing are collected from different human subjects.
Neural Network Training
The starting positions of the human arm and the possible goal locations of the test trajectories are different. In the training phase, some of the trajectories involved reaching for objects that were randomly placed close to each other in a cluttered manner, and some of the recorded arm motions consisted of the subject initially moving his or her hand close to an object and finally reaching another item. Each trajectory contained roughly 40–60 frames of skeletal data. A set of eight trajectories is used for training an NN. The raw position measurements obtained from the RGB-D camera sensor are processed using a KF, such as the one in [114], to obtain the position and velocity estimates. The number of neurons in the hidden layer is empirically chosen to be 50.
Neural Network Testing
Once the NN is trained, the test data from a different subject are used as measurements to infer the underlying intentions of the reaching motion. During the inference, the NN weights are learned online to adapt to the motion performed by the test subject. It should be noted that the total number of frames for each reaching motion is not fixed, and the intended object is reached at varying frame numbers. The Q function is evaluated for all the possible intentions to find the one that leads to the maximum Q value (the direct evaluation method). The initial mean of the state μ(0) is assumed to be a zero vector. The initial state covariance PEM(0), the process noise covariance Q, and the measurement noise covariance Σz are selected to be 0.2I24×24, 0.1I24×24, and 0.2I24×24, respectively, where I denotes the identity matrix. The gains for the online learning algorithm defined in (13) and (15) are selected to be k = 25, α = 5, γ = 25, and β1 = 4, and the adaptation gains are chosen to be ΓW = 0.75I50×50, , and . The state estimates are initialized to the same value as the first measurement z1. The sampling time for discretization is 1/30 s.
When the intention ĝ is modeled as a continuous variable, the GradEM algorithm is used for evaluating the intention estimate. In Figure 3, the convergence of the estimated goal location to the true goal location is shown. The numerical optimization of the Q function is completed for five iterations at every time step. The state and intention estimates are initialized using two methods, namely, random selection and gaze-based selection. In the first set of experiments, the goal location is arbitrarily chosen from the eight possible ones, and the convergence of the estimated intention to the true intention is shown in Figure 3(a). In the second set, the approximated intention is initialized using gaze cues, and the convergence of the forecast intention to the true one is shown in Figure 3(b). It can be observed that the intention estimates ĝx, ĝy, and ĝz for the gaze-based initialization converge faster and show fewer transients, compared to the random initialization. When the intention ĝ is modeled as a discrete variable, a direct evaluation of the Q function is performed to estimate the intention.
FIGURE 3.

The estimated goal location using the adaptive neural intention estimator algorithm by numerically optimizing the Q function. (a) The convergence plot of the estimated intention (X, Y, and Z locations) when the initial estimate is randomly chosen. (b) The convergence plot of the estimated intention (X, Y, and Z locations) when the initial estimate for expectation-maximization optimization is chosen using gaze cues.
In the first method, one of the possible n goal locations is selected randomly as the initial estimate. In Figure 4, the intention estimate progression and trajectory evolution for such a randomly chosen estimate is shown. Error statistics with a large number of experiments are reported in the prior work in [34]. Due to the arbitrary assignment, the initial approximation can be any of the possible goal locations, including the ones far away from the true intention. It is observed that, in some cases, the EM algorithm requires more observations to converge to the true intention when the intention estimate is randomly initialized. To overcome this issue, a gaze-based selection of the goal location for optimizing the EM is tested in the second experiment, which provides better cues of the reaching goal location of the person. A dense gaze probability map is obtained using the method described in [53] and [113], and a probability is assigned to each of the possible goal locations by using the formula in (16). The goal location with the highest probability is chosen as the initial estimate ĝ (0).
FIGURE 4.

An image sequence showing the online inference of intention and the evolution of the human hand (wrist) trajectory (shown in green) through an approximate expectation-maximization algorithm with online model learning. The initial estimate of the reaching goal intention is randomly chosen from the finite number of objects available on the table.
In Figure 5, the dense gaze probability map is presented along with the probabilities computed for each of the possible goal locations. The object with the maximum probability value, 0.2 in this case, is chosen as the initial intention estimate. In Figure 6, the intention estimate progression is shown, along with the trajectory evolution for the initial intention estimate computed from the gaze map. For the case of the sequence in Figure 6, the goal location with the maximum gaze probability is the true goal location. As a result, the algorithm predicts the intention correctly throughout the sequence. In the case of the random initialization of the intention, the ML estimation is equivalent to the MAP approximation with a uniform prior.
FIGURE 5.

(a) Goal location probabilities computed from the gaze map. The numbers shown inside the labels represent the probability of an object being reached by the person, based on the person’s initial head orientation. (b) A dense gaze map showing the gaze probability associated with each pixel.
FIGURE 6.

An image sequence showing the online inference of a person’s reaching intention and the evolution of the human hand (wrist) trajectory for an initial intention chosen using the gaze map. The gaze cue aids in narrowing the human’s reaching motion to the object being selected.
The observations in previous experiments that the ML method gives a faster convergence when initialized with gaze cues compared to the random initialization could be a coincidence. If a human is looking toward the opposite side of the direction of the arm motion, the gaze-based goal initialization will not be very useful. This may lead to a bad initialization for the goal location of the EM algorithm. If the initialization of the goal location is bad, the EM algorithm may require more measurements before it correctly converges to the true goal location because EM is a ML estimation method. As demonstrated in Figure 3, the EM algorithm requires observations of up to 0.6 s to correctly predict the goal location for a 1.5-s sequence.
Intention Estimation as a Multiple-Model Estimation Problem
In this section, a MAP estimation of the human goal-reaching intention is computed using a multiple-model estimation problem. Since the human reaching motion is generated by an inherent stable motion strategy, each movement is modeled as a continuous DS whose solutions converge to the goal location. To ensure that the models reach the goal location, the dynamics that are approximated through NNs are trained subject to convergence to the goal location constraints. The intention estimation strategy is to select the correct model based on the currently observed motion trajectory. A block diagram of the multiple-model-based intention estimation methodology is given in Figure 7. More details of the algorithm can also be found in [53].
FIGURE 7.

A block diagram of the multiple-model-based intention estimation algorithm with gaze priors. The gaze estimator block uses red-green-blue (RGB) images along with the head position to estimate the gaze map of a person, which determines the most probable objects that the person is looking at. The information from the gaze map is used to compute the prior model probability for the interacting multiple-model (IMM) algorithm. CNN: convolutional neural network; ANN: artificial neural network; EKF: extended Kalman filter.
Human Motion Dynamic Model
In this section, a method for learning the nonlinear dynamics of the human arm-reaching motion is presented. Consider a state variable and a set of demonstrations representing reaching motions to various goal locations. Each demonstration would consist of the trajectories of the state and the trajectories of the state derivative from time t = 0 to t = T. All state trajectories of the demonstrations are translated such that they converge to the origin. Let the translated demonstrations be solutions to the underlying DS governed by the first-order differential equation
| (17) |
where fe(x(t)) is a Lipschitz continuous function. Since all the trajectories of the translated demonstrations converge to the origin, the system defined in (17) could be seen as a globally contracting one. The nonlinear function fe(x(t)) is approximated by an NN similar to the one in (2) without the intention variable g. Note that only one NN is used to represent the dynamics of reaching motion trajectories that converge to the origin. Arm motion trajectories pertaining to different goal locations can be obtained by corresponding linear translations of the solutions to the DS in (1).
Learning Contracting Nonlinear Dynamics of Human Reaching Motion
To learn the NN weights from the sample reaching motion data, the following NN weight training algorithm is used. The weights are trained such that the states of the dynamics converge to a given reaching goal location. To achieve this, a constrained optimization problem is solved subject to goal reaching terminal constraints enforced using contraction analysis of nonlinear dynamics. The following optimization problem is set up to learn the weights of the NN:
| (18) |
| (19) |
where , and represent the target and the network’s output of the ith demonstration, EW is the sum of the squares of the NN weights, α, are scalar parameters of regularization, is a strictly positive constant, and represents a constant positive symmetric matrix. The details of the computation of the Jacobian ∂fe(x(t))/∂x for the NN approximation of fe(x(t)) can be found in [62].
Multiple-Model Estimation Algorithm for Intention Estimation
Based on multiple dynamic NN models that reach different goal locations, an IMM algorithm is developed that selects the most probable model, which is a representation of the current trajectory of a human’s reaching motion observed through sensor data. The IMM estimator is first described, and the computation of priors using gaze-based cues is subsequently explained.
Interacting Multiple-Model Estimator
Given the trained network and a trajectory of the reaching hand, the problem involves inferring the goal location in advance. Let a fixed set of candidate goal locations that the human can reach be . The NN weights learned from human demonstrations are used to represent human motion. For each goal location gj, the state vector and the corresponding dynamics are defined as , and . Similarly, for a fixed set of Ng goal locations, a set of Ng dynamic systems is formed. The discretized versions of these systems are
| (20) |
where j = 1, …, Ng, δt is the sampling period, and is a zero-mean Gaussian random process with a covariance matrix . For this section, consider n = 3; that is, only the last joint of the arm’s skeleton is tracked with the Xe, Ye, and Ze positions. The measurement model is
| (21) |
where ze(t) is the measurement vector, is a zero-mean GP with covariance Re, and
is the measurement function.
Let represent the Ng models defined in (20) and (21) for the set of candidate goal locations G. The posterior probability of model j being correct is denoted by P(gj | Z1:t). The expression P(gj | Z1:t) indicates the posterior probability of each gj being the correct goal location, given a set of measurements Z1:t = [ze(1), ze(2), …, ze(t)]. Note that P(gj | Z1:t) = P(Mj | Z1:t) since the models and goal locations have a one-to-one correspondence. Hence, to obtain the posterior probabilities P(gj| Z1:t), j = 1, ‥, Ng, the posterior probabilities of the models P(Mj | Z1:t), j = 1, ‥, Ng are computed. The posterior probability P(Mj | Z1:t) is calculated using the Bayes’ formula as , where p(z(t) | Z1:t−1, Mj) is the likelihood function of mode j at time t and P(Mj) | Z1:t−1 is the prior probability of Mj being correct. In the IMM framework with Ng models, the likelihood function p(z(t) | Z1:t−1, Mj) of mode j at time t is . The innovation and its covariance Sj are computed from the mode-matched filter corresponding to mode j.
The G-MMIE algorithm uses EKFs matched to each mode. Other filters based on the state-dependent coefficient form parameterization of nonlinear systems [115] can also be used for each individual filter. Each iteration of the IMM filter for intention inference is divided into four main steps: the interaction/mixing stage, model-matched filtering, the model probability update, and model switch detection. More details about the static IMM filter can be found in [64]. Details specific to the multiple-model filter in the context of gaze-based intention estimation can be reviewed in [53].
In many real-world applications, a small number of models may not be sufficient to describe all the modes. When the number of models is large, the performance of the IMM filter can degrade, and the computational burden increases. A VS-IMM filter can be used in such cases [65]. A limiting case of the VS-IMM when the mode space is continuous is presented in [66]. In certain HRC applications, a significant number of models is required to represent the application context.
Computation of Prior Distribution the Using Gaze Map for the Interacting Multiple-Model Filter
Using the gaze-based prior computation procedure and the average prior probability computed through (16), the probability of each of the Ng candidate locations being the goal site is , where μj(0) is the prior probability of gj being the goal area for the IMM filter and gj refers to the positioning of the jth object.
Experimental Results for Multiple-Model-Based Intention Estimation
To validate the multiple-model-based intention estimation algorithm, a set of 10 demonstrations collected from a subject is used for training the NN under contraction analysis constraints. For training the NN, each demonstration is labeled based on the ground truth goal location. Note that the ground truth labeling is done only for the training data. All data are collected by a Microsoft Kinect for Windows. The joint position data obtained from the subjects are preprocessed to obtain the velocity and acceleration estimates using a KF (see [114] for details). In all the experiments, the position and velocity of the hand in the 3D Cartesian space are considered to be the elements of the state vector , and the number of possible goal locations Ng = 8.
An IMM filter for computing the intention estimate is implemented using the following parameters. The initial state estimate covariance , j = 1 2, ‥, Ng, the process noise covariance , j = 1 2,,‥,Ng, and the measurement noise covariance Re for the EKF of the IMM filter are selected to be 0.2I6×6, .0 1I6×6, and 0 2I6×6, respectively. The state estimates , j = 1 2, ‥, Ng are initialized using the first two measurements ze(1) and ze(2) (a finite difference method is used for the velocity initialization). The model transition matrix for the IMM is chosen to be an 8 × 8 matrix Πij with diagonal elements of 0.79 and off-diagonal elements of 0.03. For computing the IMM priors, a uniform distribution and gaze-based cues is applied.
In the first experiment, a uniform distribution is used as a prior, which corresponds to all possible goal locations having an equal probability of being the true intention. In Figure 8, the intention estimate is displayed along with the trajectory evolution for the uniform prior. The initial goal location is set to object 5, and, after 0.3 s, the predicted goal location matches the true intention. In the second experiment, the prior distribution is computed from the dense gaze map, as shown in Figure 5, using (16). The gaze-based prior assigns a higher prior probability to the goal locations where the human subject is likely to be looking. In Figure 9, the intention estimate and trajectory evolution for a prior distribution computed by using a dense gaze map are presented. The goal location with the highest prior probability is the true intention. Gaze-based prior computation assigns a higher probability value to the true intention, which results in a better prediction with fewer observations of the human hand trajectory. The interested reader is referred to the prior work in [53], where the error statistics of a large number of experiments are reported.
FIGURE 8.

An image sequence showing the online inference of intention and the evolution of the human hand trajectory for a uniform prior distribution.
FIGURE 9.

An image sequence showing the online inference of intention and the evolution of the human hand trajectory for a prior distribution computed using the gaze map.
SAFE ROBOT CONTROLLER BASED ON HUMAN-INTENTION INFERENCE
This section describes robot control design that takes into consideration the human trajectory generated by the intention estimators to produce motion that is safe around the person. First, a robot’s desired trajectory generation algorithm is presented to determine the desired trajectory based on the human’s predicted course from the intention estimator. A BF formulation is then used to modify the robot’s desired trajectories, which cross the boundary of the safety ellipsoid drawn around the human. In a subsequent section, a torque controller that follows the robot’s desired trajectory is discussed. Note that the dynamics in continuous time are used for the math development in this section because the controller sampling rate is typically much higher compared to the sampling rate of the data obtained from the sensors used for the intention estimation.
Problem Scenario
An object-carrying task is used as a test case, where a person is holding one side of an object and the robot is holding the other side. While carrying the object, the robot’s end effector trajectories may cross the safety ellipsoid around the human, which is when the CBF is used to modify the robot’s desired movements. An example scenario of the task is in Figure 10.
FIGURE 10.

An example scenario showing a human and a robot working closely together to carry an object. The gray sphere around the human operator corresponds to the region the robot should not enter during the task execution. The magenta (dashed) and green (solid) lines show the desired motion trajectory of the object-carrying task. The stars and circles represent the initial and target locations of the motion, respectively.
Generation of the Robot’s Desired Trajectories for the Human–Robot Task
A method is presented that generates the desired robot trajectories as a function of the human’s inferred movement. If the robot’s desired trajectory enters the prescribed safety ellipsoid around the human, then a BF-based formulation is used to modify the robot’s movement such that it always stays outside the danger area. In Figure 11, the details of the reference trajectory generation algorithm are illustrated in the form of a block diagram. Let be the desired position state of the robot’s end effector and be the human’s position trajectory generated by the intention estimation algorithm. The robot trajectory xRd(t) is
| (22) |
where T(·) is a generic transformation between xH(t) and xRd(t). Specifically, an affine transformation with known task-specific parameters and is used in this article. If the robot’s generated desired trajectory xRd(t) enters an unsafe zone represented by an ellipsoid drawn around the human, then the robot’s preferred course is required to be modified such that the robot does not collide with the person.
FIGURE 11.

A block diagram of the reference trajectory generator for the robot, based on observed and inferred human action. The human movement approximation produced by intention estimator is fed into a reference trajectory generator block for the robot. If the robot-reference trajectory enters the unsafe zone around the human, then a control barrier function approach is used to modify the robot’s movement. Otherwise, a control Lyapunov function approach is used to generate a bounded and convergent reference trajectory for the robot.
To modify the robot’s desired trajectory, a dynamic model of xRd(t) is formulated with a control input so that two main objectives are satisfied: 1) the controller is able to modify xRd(t) such that xRd(t) does not enter the safety ellipsoid around the human, and 2) the modified xRd(t) still closely tracks the original desired trajectory xRd(t). To achieve this, consider the robot’s desired end effector motion dynamics , where is a nonlinear continuous function that governs the robot’s end effector motion dynamics. Based on the estimates of the human trajectory intention and its time derivative [and using function approximation methods, such as the extreme learning machine (ELM)], the function can be learned, where is an approximation of fRn(·). See “Extreme Learning Machine” for more details of the ELM.
GMMs can also be used to approximate the nonlinear function for capturing the uncertainty in the data. To ensure that the trajectories generated by the nominal model avoid the working region of the human operator (represented by a 3D ellipsoid), a controller is designed to provide modifications to using CBFs and control Lyapunov function (CLF) theory. To execute this, a robot’s modified desired end effector dynamics can be written as
| (23) |
where is a control input that is designed to avoid unsafe zones. In the next section, the design of using CBFs and CLF theory is discussed.
Control Barrier Function Formulation
A CBF defines a forward invariant region such that solutions of the DS that start in that area remain there permanently. The choice of an invariant region is application specific as long as the CBF conditions specified in “Constructing Control Barrier and Control Lyapunov Functions” are valid choices. In this article, the CBF is used to avoid the obstacle region; that is, the robot trajectories that start outside the unsafe region permanently remain outside that zone. Ellipses, circles, and ellipsoids are commonly used shapes to represent obstacles and regions of operation where it is unsafe for a robot’s end effector to enter. The application uses a 3D ellipsoid shape around a human as an unsafe zone for a robot to enter. The ellipsoidal BF is defined as a continuously differentiable function given by
| (24) |
where x(t) = [x1(t), x2(t), x3(t)]T; , , and are the coordinates of the center of gravity of the ellipse; and a1, a2, and a3 are the semiaxes of an ellipsoid. Using the BF in (24), the CBF candidate is defined as B(x(t)) = 1/b(x(t)), which satisfies the properties given in “Constructing Control Barrier and Control Lyapunov Functions.”
Control Lyapunov Function Formulation
The CLF is used to ensure that the desired trajectories of the robot remain stable and converge to a point in the 3D space where the object is being placed. Consider a quadratic Lyapunov function candidate of the form , . Using the time derivative of the CLF and (23),
| (25) |
where is a positive constant.
Control Design Using Control Barrier Function and Control Lyapunov Function Constraints
Given the approximated nonlinear function , an online controller learning problem with safety and stability constraints is now discussed. Two cases are considered: 1) when the robot’s desired trajectories are not crossing the safety ellipsoid and 2) when the robot’s desired trajectories are crossing the safety ellipsoid.
Case 1
When the robot’s desired trajectories are not crossing the safety ellipsoid, then a controller that uses CLF constraints is synthesized. The controller with CLF constraints will ensure that the trajectories generated by (23) remain stable with respect to the equilibrium point . To this end, the following quadratic program (QP) is solved to synthesize the controller:
| (26) |
where is a positive constant added as a relaxation variable that ensures the solvability of the QP as penalized by p > 0 and is a positive definite weight matrix.
Case 2
When the robot’s desired trajectories are crossing the safety ellipsoid, then the optimal controller is synthesized using both CLF and CBF constraints. The CBF constraint will ensure that the trajectories generated by (23) will not enter the unsafe ellipsoidal zone, and the CLF constraint will guarantee that the trajectories remain stable with respect to . The optimization problem subject to the CBF and CLF is
| (27) |
| (28) |
where , γl > 0, and γB > 0 are positive constants.
Numerical Simulation Results
Given the position estimates of the human-demonstrated motion trajectory, the robot’s desired movement can be obtained from (23). However, the position trajectory generated from the affine function does not guarantee that the human workspace will be avoided. As shown in Figure 12, the robot’s desired trajectory (in solid green lines) intersects with the human’s workspace. An ellipsoid is used to represent the unsafe region for the robot’s end effector. In that area, the human completes his or her own tasks. Based on the robot’s desired end effector states (position and velocity), an ELM parameterization is employed to approximate the nonlinear function that governs the underlying dynamics. A controller is synthesized by optimizing the objective function, subject to the constraints derived using the CBF and the CLF, to ensure the position trajectory generated from the learned model avoids the human workspace.
FIGURE 12.

Human-demonstrated motion trajectory is enclosed by an ellipsoid to represent the unsafe region for a robot’s end effector to enter. The robot’s desired end effector position trajectory is modified by a controller that avoids the ellipsoid zone while reaching the target location. CBF: control barrier function; CLF: control Lyapunov function.
In Figure 12, the simulated results are shown in which the controller applies corrections to the trajectory generated by the model to avoid the human workspace while still reaching the target location. In Figure 13, an example is illustrated in which the robot’s desired end effector trajectory is far from the human workspace. Thus, the robot poses no harm to the human operator. Therefore, a CLF is used to learn a controller, which ensures that the robot’s end effector reaches the target location without interfering with the person’s workspace.
FIGURE 13.

A scenario where the robot’s desired end effector position trajectory is far from the unsafe region dictated by the human operator. A controller is learned using the control Lyapunov function (CLF) that avoids the region where a human is present (represented by an ellipsoid) while reaching the target location.
ADAPTIVE ROBOT CONTROLLER
In this section, an adaptive controller for the robot arm is designed that follows the preferred movement created by the desired trajectory generator block. First, a robot dynamic model represented by EL dynamics is discussed. An adaptive controller for the robot manipulator that tracks the desired trajectory is presented with a Lyapunov stability analysis.
Robot Dynamic Model
Consider the following robot manipulator dynamics (represented by EL dynamics) given by
| (29) |
In (29), q(t), , and denote the generalized states; represents a generalized inertia matrix; designates a generalized centripetal Coriolis matrix; indicates a generalized gravity vector; and signifies the generalized input control vector. It is well known that, for the robot manipulator dynamics in (29), the following properties hold:
Property 1: The matrix M(q) is positive definite, and λminI ≤ M(q) # λmaxI, where λmin and λmax are positive constants.
Property 2: For any differentiable vector , the dynamics in (29) are linearly parameterizable as , where is a set of robot-specific parameters and is a matrix of known functions of the generalized coordinates and their higher directives.
Property 3: The inertia and centripetal Coriolis matrices satisfied the property , ; that is, is a skew-symmetric matrix. See [112] and [116] for more details.
Let be the robot’s end effector state and be the joint angles of the robot. The forward and velocity kinematic relationships are given by , , where is the mapping between the joint space QJ and the task space and is the known Jacobian.
Controller Design
The desired trajectory of the robot end effector state is produced based on the robot’s desired trajectory generator block. To design the robot controller, consider the signals a(t), v(t), and s(t) for nonredundant robots given by , , and , where is the tracking error, , , and is a positive definite diagonal matrix. Also consider the filtered error in the task space, r(t) ≜ J(q) s(t). Using (29), r(t) can be expressed as
| (30) |
The control input in (29) is designed to be
| (31) |
where is a regressor matrix; consists of estimates of the parameters M, C, and g, respectively; is the estimated tracking error; and are diagonal matrices; , ; and . The parameter update rule is
| (32) |
where is a positive definite matrix and proj(·), is a standard projection operator, which ensures that the parameter estimates are bounded (see [117] for details).
Remark 1
The parameter estimation error is uniformly continuous since evolves according to (32). Substituting (31) in (29) yields the closed-loop system as given by
| (33) |
Assumption 1
The signals , , and are uniformly continuous, and , , as t → ∞.
Remark 2
Based on Assumption 1 and formulas for , , and , it can be seen that , , , as t → ∞.
Stability Analysis
A Lyapunov stability theorem is provided for the closed-loop system defined in (33).
Theorem 1
The closed-loop system in (33) is stable, and the tracking error is globally and uniformly bounded in the sense that ∥e(t)∥ → ϵ0exp(−ϵ1t)+ϵ2, where ϵ0, ϵ1, denote positive bounding constants if the following conditions for the gains of the controller defined in (31) are satisfied: Kt > 0 and KJ > 0. See [9] for the proof and more details. Extensions of the adaptive controller design for redundant manipulators can be found in [118]. The stability analysis can be extended to the modified controller, and the redundant manipulator can be shown to asymptotically track the desired trajectory.
Numerical Simulation Results
The results of the tracking controller (shown in Figures 14 and 15) were generated using a simulation of Rethink Robotics’ seven-degrees-of-freedom Baxter Robot in the Gazebo simulation platform. The reference trajectories and their derivatives (which are obtained by observing the motion of a human hand and then modified according to the CBF and CLF constraints) are used as and for the controller defined in (31). During execution, the true robot task space end effector positions xR(t) and velocities are read from virtual sensors provided in Gazebo.
FIGURE 14.

(a) The desired trajectories of the robot end effector being tracked by the robot controller. (b) The position tracking errors.
FIGURE 15.

(a) The desired trajectories of the robot end effector being tracked by the robot controller. (b) The position tracking errors.
Experimental Results
A collaborative task involving a human and a robot moving a heavy object is designed for experimental evaluation. In it, the robot must infer the motion intention of the human online and adjust its trajectory to coordinate with the human to move a rectangular box from a workbench to an elevated surface. While doing so, the robot must follow a trajectory that does not enter the human workspace to ensure the safety of the person. The human is assumed to move between two different dynamical models (one for moving the box parallel to the surface of the workbench and the other for lifting the object and placing it on the elevated surface). The experiment is demonstrated through a sequence of images, as shown in Figure 16.
FIGURE 16.

Experimental results for collaboratively moving an object using a Baxter research robot and a person.
Four demonstrations of the human hand motion training data for each movement are collected using Microsoft Kinect skeletal tracking to train two NN models (corresponding to the two dynamical models) subject to contraction constraints. The single-layer NN model for the translational motion consists of 12 neurons, and the model for the vertical motion consists of 15 neurons. The trained NNs are then used as motion models for EKFs, which are incorporated into the IMM framework. A Microsoft Kinect sensor placed on top of Baxter’s head is employed for acquiring the human hand position data online during the experiment. The IMM estimator is used to approximate the 3D position of the human hand and the likelihood of each model being the true representation at each time instance.
A controller is designed by the solving the optimization problem in (26)–(28), which generates the robot’s desired trajectory . The controller ensures that the robot’s desired trajectory stays outside the human workspace, ensuring the safety of the person. The matrix H is chosen to be identity for the optimization. The adaptive tracking controller in (31) is used to monitor the safe trajectory generated by solving the optimization problem. The experiments are completed with the Baxter research robot and the Baxter application programming interface along with the Robot Operating System. The gains chosen for the adaptive controller are KJ = diag {27, 30, 37.5, 0, 0, 0}, Kt = diag {3, 6, 3, 3.25, 3, 3, 1.5}, and Ʌ = diag {1, 1, 1, 0, 0, 0}. The orientation of the robot end effector was constant throughout the experiment. The results of the experiment are summarized in Figure 17(a) and (b). The robot end effector trajectories converge and track the desired trajectories, as shown in Figure 17(a). The error between the actual and desired trajectories appears in Figure 17(b).
FIGURE 17.

Experimental results from a Baxter research robot. (a) The desired trajectories of the Baxter robot end effector being tracked by the robot controller. (b) The position tracking errors.
CONCLUSION
For assistive robots to integrate seamlessly into human environments, they should be able to understand the intentions of human agents and adapt to those people’s motions. Many HRC tasks in advanced manufacturing operations require humans and robots to perform joint tasks, such as carrying heavy loads and large, flexible materials and completing welding operations, for which the robot can act as an assistant to learn people’s motion behaviors, based on RGB-D sensor data. A robot can then adapt its motions to complete a cooperative task. In this article, a survey of the topic was first provided by presenting existing literature in a tutorial manner, and approaches that focus and expand this article were discussed for inferring human motion trajectories and estimating reaching goal intentions.
The first method is based on an ML estimation technique called approximate EM that uses online model learning to accommodate uncertainties in human motions. The second is based on a multiple-model estimator that switches between multiple nonlinear human motion models. Short tutorials on approximate EM and multiple-model estimation methods were presented that use nonlinear models of human motion learned through NN function estimation from labeled sensor data. For the ML-based estimator, the optimizer is initialized using human gaze cues and random initialization. For the MAP-based estimator, a gaze-based prior is employed that is computed by analyzing RGB image data using a deep NN. The results were compared with the uniform prior.
A human-in-the-loop control strategy was discussed that uses the estimated human motion trajectory and intended reaching goal location to determine a safe robot-reference trajectory, which is tracked via an adaptive controller for the robot. For the desired trajectory generation, a CBF formulation is used to produce movements that are safe around humans. The desired motions are created with two objectives: 1) the robot end effector trajectories do not enter a safety region around a human, and 2) the robot end effector trajectories are able to synchronize with human motion as closely as possible. The adaptive controller converges to the desired trajectories produced by the generation block.
A case study of humans and robots carrying an object together was discussed. Future challenges, such as ensuring the safety of humans in the presence of actuator and sensor failures, can be a potential topic to explore, for example, via human supervision of automation to achieve resilient control [119]. Estimating intention by using sensors, such as ultrasound imaging, can be another avenue for potential research [120].
Supplementary Material
ACKNOWLEDGMENTS
This work was supported, in part, by a Space Technology Research Institutes grant (80NSSC19K1076) from the NASA Space Technology Research Grants Program, by Office of Naval Research Award N00014-20-1-2040, and by the U.S. Department of Energy’s Office of Energy Efficiency and Renewable Energy (EERE) under the Advanced Manufacturing Office Award DE-EE0007613. The authors would like to thank Tanisha Mitra for her help with figure formatting.
AUTHOR INFORMATION
Ashwin P. Dani (ashwin.dani@uconn.edu) received the M.S. and Ph.D. degrees from the University of Florida, Gainesville. He was a postdoctoral research associate at the University of Illinois at Urbana–Champaign. He is currently an associate professor at the University of Connecticut, Storrs. He has authored more than 50 technical articles and four book chapters. His current research interests include nonlinear estimation and control, machine learning for control, human–robot collaboration, and autonomous navigation. He is a Senior Member of IEEE.
Iman Salehi received the M.S. degree from the Department of Electrical and Computer Engineering, University of Hartford, Connecticut, in 2015. He is currently working toward the Ph.D. degree in electrical and computer engineering at the University of Connecticut, Storrs. His research interests include learning for control, human–robot interaction, and system identification.
Ghananeel Rotithor received the M.S. degree in biomedical engineering from the University of Florida in 2017. He is currently pursuing the Ph.D. degree in electrical and computer engineering at the University of Connecticut, Storrs. His research interests include deep learning for control, computer vision, robotics, and vision-based estimation and control.
Daniel Trombetta received the B.S. and M.S. degrees in electrical and computer engineering from the University of Connecticut, Storrs, in 2018. His research interests include machine learning, human–robot collaboration, and estimator design.
Harish Ravichandar received the M.S. degree in electrical and computer engineering from the University of Florida, Gainesville, in 2014 and the Ph.D. degree in electrical and computer engineering from the University of Connecticut, Storrs, in 2018. He is a research scientist in the School of Interactive Computing, Georgia Institute of Technology. His current research interests include robot learning, human–robot interaction, and multiagent systems. He is a Member of IEEE.
Contributor Information
ASHWIN P. DANI, University of Connecticut, Storrs.
IMAN SALEHI, University of Connecticut, Storrs.
GHANANEEL ROTITHOR, University of Connecticut, Storrs.
DANIEL TROMBETTA, University of Connecticut, Storrs.
HARISH RAVICHANDAR, School of Interactive Computing, Georgia Institute of Technology.
REFERENCES
- [1].Berger S, Making in America: From Innovation to Market. Cambridge, MA: MIT Press, 2013. [Google Scholar]
- [2].Modares H, Ranatunga I, AlQaudi B, Lewis FL, and Popa DO, “Intelligent human–robot interaction systems using reinforcement learning and neural networks,” in Trends in Control and Decision-Making for Human–Robot Collaboration Systems, Wang Y and Zhang F, Eds. Switzerland: Springer, 2017, pp. 153–176. [Google Scholar]
- [3].Mumm J and Mutlu B, “Human–robot proxemics: Physical and psychological distancing in human-robot interaction,” in Proc. 6th Int. Conf. Human-Robot Interaction, 2011, pp. 331–338. doi: 10.1145/1957656.1957786. [DOI] [Google Scholar]
- [4].Nobile C, “Robotics business review perspectives 2013: Outlook for nextgen, new-gen industrial co-worker robotics,” Robot. Business Rev 2013. [Online]. Available: https://www.roboticsbusinessreview.com/manufacturing/outlook_for_next_gen_new_gen_industrial_co_worker_robotics/ [Google Scholar]
- [5].Sampath M and Khargonekar PP, “Socially responsible automation: A framework for shaping the future,” Nat. Acad. Eng. Bridge, vol. 48, no. 4, pp. 45–52, 2018. [Google Scholar]
- [6].Peshkin MA, Colgate JE, Wannasuphoprasit W, Moore CA, Gillespie RB, and Akella P, “Robot architecture,” IEEE Trans. Robot. Autom, vol. 17, no. 4, pp. 377–390, 2001. doi: 10.1109/70.954751. [DOI] [Google Scholar]
- [7].Goodrich MA and Schultz AC, “Human–robot interaction: A survey,” Found. Trends Hum.-Comput. Interaction, vol. 1, no. 3, pp. 203–275, 2007. doi: 10.1561/1100000005. [DOI] [Google Scholar]
- [8].Martinez E and del Pobil AP, “Safety for human–robot interaction in dynamic environments,” in Proc. IEEE Int. Symp. Assembly and Manufacturing, 2009, pp. 327–332. doi: 10.1109/ISAM.2009.5376949. [DOI] [Google Scholar]
- [9].Ravichandar HC, Trombetta D, and Dani AP, “Human intention-driven learning control for trajectory synchronization in human-robot collaborative tasks,” IFAC Cyber-Phys. Hum. Syst, vol. 51, no. 34, pp. 1–7, 2019. doi: 10.1016/j.ifacol.2019.01.001. [DOI] [Google Scholar]
- [10].Sodhi M, Reimer B, and Llamazares I, “Glance analysis of driver eye movements to evaluate distraction,” Behav. Res. Methods, Instrum., Comput, vol. 34, no. 4, pp. 529–538, 2002. doi: 10.3758/BF03195482. [DOI] [PubMed] [Google Scholar]
- [11].Yepes JL, Hwang I, and Rotea M, “New algorithms for aircraft intent inference and trajectory prediction,” J. Guid., Control, Dyn, vol. 30, no. 2, pp. 370–382, 2007. doi: 10.2514/1.26750. [DOI] [Google Scholar]
- [12].Alqaudi B, Modares H, Ranatunga I, Tousif SM, Lewis FL, and Popa DO, “Model reference adaptive impedance control for physical human-robot interaction,” Control Theory Technol, vol. 14, no. 1, pp. 68–82, 2016. doi: 10.1007/s11768-016-5138-2. [DOI] [Google Scholar]
- [13].Tsai C-S, Hu J-S, and Tomizuka M, “Ensuring safety in human-robot coexistence environment,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems (IROS), 2014, pp. 4191–4196. doi: 10.1109/IROS.2014.6943153. [DOI] [Google Scholar]
- [14].Kim D-J, Wang Z, Paperno N, and Behal A, “System design and implementation of UCF-MANUS: An intelligent assistive robotic manipulator,” IEEE/ASME Trans. Mechatron, vol. 19, no. 1, pp. 225–237, 2012. doi: 10.1109/TMECH.2012.2226597. [DOI] [Google Scholar]
- [15].Zhang F and Huang H, “Source selection for real-time user intent recognition toward volitional control of artificial legs,” IEEE J. Biomed. Health Inform, vol. 17, no. 5, pp. 907–914, 2012. doi: 10.1109/JBHI.2012.2236563. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Sugar TG et al. , “Design and control of RUPERT: A device for robotic upper extremity repetitive therapy,” IEEE Trans. Neural Syst. Rehabil. Eng, vol. 15, no. 3, pp. 336–346, 2007. doi: 10.1109/TNSRE.2007.903903. [DOI] [PubMed] [Google Scholar]
- [17].Baldwin DA and Baird JA, “Discerning intentions in dynamic human action,” Trends Cogn. Sci, vol. 5, no. 4, pp. 171–178, 2001. doi: 10.1016/S1364-6613(00)01615-6. [DOI] [PubMed] [Google Scholar]
- [18].Simon MA, Understanding Human Action: Social Explanation and the Vision of Social Science. New York: SUNY Press, 1982. [Google Scholar]
- [19].Kleinman DL, Baron S, and Levison W, “An optimal control model of human response part I: Theory and validation,” Automatica, vol. 6, no. 3, pp. 357–369, 1970. doi: 10.1016/0005-1098(70)90051-8. [DOI] [Google Scholar]
- [20].Baron S, Kleinman D, and Levison W, “An optimal control model of human response part II: Prediction of human performance in a complex task,” Automatica, vol. 6, no. 3, pp. 371–383, 1970. doi: 10.1016/0005-1098(70)90052-X. [DOI] [Google Scholar]
- [21].Warrier RB and Devasia S, “Inferring intent for novice human-in-the-loop iterative learning control,” IEEE Trans. Control Syst. Technol, vol. 25, no. 5, pp. 1698–1710, 2017. doi: 10.1109/TCST.2016.2628769. [DOI] [Google Scholar]
- [22].Liu C et al. , “Goal inference improves objective and perceived performance in human–robot collaboration,” in Proc. Int. Conf. Autonomous Agents Multiagent Systems, 2016, pp. 940–948. [Google Scholar]
- [23].Li Y and Ge S, “Human-robot collaboration based on motion intention estimation,” IEEE/ASME Trans. Mechatron, vol. 19, no. 3, pp. 1007–1014, 2014. doi: 10.1109/TMECH.2013.2264533. [DOI] [Google Scholar]
- [24].Kulic D and Croft EA, “Affective state estimation for human–robot interaction,” IEEE Trans. Robot, vol. 23, no. 5, pp. 991–1000, 2007. doi: 10.1109/TRO.2007.904899. [DOI] [Google Scholar]
- [25].Meisner E, Isler V, and Trinkle J, “Controller design for human–robot interaction,” Autonom. Robot, vol. 24, no. 2, pp. 123–134, 2008. doi: 10.1007/s10514-007-9054-7. [DOI] [Google Scholar]
- [26].Razin YS, Pluckter K, Ueda J, and Feigh K, “Predicting task intent from surface electromyography using layered hidden Markov models,” IEEE Robot. Autom. Lett, vol. 2, no. 2, pp. 1180–1185, 2017. doi: 10.1109/LRA.2017.2662741. [DOI] [Google Scholar]
- [27].Bartlett MS, Littlewort G, Fasel I, and Movellan JR, “Real time face detection and facial expression recognition: Development and applications to human computer interaction,” in Proc. IEEE Conf. Computer Vision Pattern Recognition Workshop (CVPRW), 2003, vol. 5, p. 53. doi: 10.1109/CVPRW.2003.10057. [DOI] [Google Scholar]
- [28].Strabala K, Lee MK, Dragan A, Forlizzi J, and Srinivasa S, “Learning the communication of intent prior to physical collaboration,” in Proc. IEEE Int. Symp. Robot Human Interactive Communication, Sept. 2012, pp. 968–973. doi: 10.1109/ROMAN.2012.6343875. [DOI] [Google Scholar]
- [29].Matsumoto Y, Heinzmann J, and Zelinsky A, “The essential components of human-friendly robot systems,” in Proc. Int. Conf. Field Service Robotics, 1999, pp. 43–51. [Google Scholar]
- [30].Traver VJ, del Pobil AP, and Perez-Francisco M, “Making service robots human-safe,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems, 2000, pp. 696–701. doi: 10.1109/IROS.2000.894685. [DOI] [Google Scholar]
- [31].Mutlu B, Forlizzi J, and Hodgins J, “A storytelling robot: Modeling and evaluation of human-like gaze behavior,” in Proc. IEEE-RAS Int. Conf. Humanoid Robots, 2006, pp. 518–523. doi: 10.1109/ICHR.2006.321322. [DOI] [Google Scholar]
- [32].Fong T, Nourbakhsh I, and Dautenhahn K, “A survey of socially interactive robots,” Robot. Autonom. Syst, vol. 42, nos. 3–4, pp. 143–166, 2003. doi: 10.1016/S0921-8890(02)00372-X. [DOI] [Google Scholar]
- [33].Wang Z et al. , “Probabilistic movement modeling for intention inference in human–robot interaction,” Int. J. Robot. Res, vol. 32, no. 7, pp. 841–858, 2013. doi: 10.1177/0278364913478447. [DOI] [Google Scholar]
- [34].Ravichandar H and Dani AP, “Human intention inference using expectation-maximization algorithm with online model learning,” IEEE Trans. Autom. Sci. Eng, vol. 14, no. 2, pp. 855–868, 2017. doi: 10.1109/TASE.2016.2624279. [DOI] [Google Scholar]
- [35].Strabala KW et al. , “Towards seamless human–robot handovers,” J. Hum.–Robot Interact, vol. 2, no. 1, pp. 112–132, 2013. doi: 10.5898/JHRI.2.1.Strabala. [DOI] [Google Scholar]
- [36].De Carli D et al. , “Measuring intent in human–robot cooperative manipulation,” in Proc. IEEE Int. Workshop Haptic Audio Visual Environments Games, 2009, pp. 159–163. doi: 10.1109/HAVE.2009.5356124. [DOI] [Google Scholar]
- [37].Ding H, Reißig G, Wijaya K, Bortot D, Bengler K, and Stursberg O, “Human arm motion modeling and long-term prediction for safe and efficient human–robot interaction,” in Proc. IEEE Int. Conf. Robotics Automation, 2011, pp. 5875–5880. doi: 10.1109/ICRA.2011.5980248. [DOI] [Google Scholar]
- [38].Gehrig D et al. , “Combined intention, activity, and motion recognition for a humanoid household robot,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems (IROS), 2011, pp. 4819–4825. doi: 10.1109/IROS.2011.6095118. [DOI] [Google Scholar]
- [39].Schrempf OC and Hanebeck UD, “A generic model for estimating user intentions in human–robot cooperation,” in Proc. Int. Conf. Informatics Control, Automation Robotics, 2005, pp. 251–256. [Google Scholar]
- [40].Elfring J, Van De Molengraft R, and Steinbuch M, “Learning intentions for improved human motion prediction,” Robot. Autonom. Syst, vol. 62, no. 4, pp. 591–602, 2014. doi: 10.1016/j.robot.2014.01.003. [DOI] [Google Scholar]
- [41].Koppula HS, Gupta R, and Saxena A, “Learning human activities and object affordances from RGB-D videos,” Int. J. Robot. Res, vol. 32, no. 8, pp. 951–970, 2013. doi: 10.1177/0278364913478446. [DOI] [Google Scholar]
- [42].Jiang Y and Saxena A, “Modeling high-dimensional humans for activity anticipation using Gaussian process latent CRFs,” in Proc. Robotics: Science Systems (RSS), 2014. [Google Scholar]
- [43].Hu N, Lou Z, Englebienne G, and Kröse B, “Learning to recognize human activities from soft labeled data,” in Proc. Robotics: Science Systems, Berkeley, CA, 2014. [Google Scholar]
- [44].Kelley R, Nicolescu M, Tavakkoli A, Nicolescu M, King C, and Bebis G, “Understanding human intentions via Hidden Markov Models in autonomous mobile robots,” in Proc. ACM/IEEE Int. Conf. Human Robot Interaction, 2008, pp. 367–374. doi: 10.1145/1349822.1349870. [DOI] [Google Scholar]
- [45].Croft DKE, “Estimating intent for human–robot interaction,” in Proc. IEEE Int. Conf. Advanced Robotics, 2003, pp. 810–815. [Google Scholar]
- [46].Ferguson S, Luders B, Grande RC, and How JP, “Real-time predictive modeling and robust avoidance of pedestrians with uncertain, changing intentions,” in Algorithmic Foundations of Robotics XI, Akin H, Amato N, Isler V, and van der Stappen A, Eds. Berlin: Springer-Verlag, 2015, pp. 161–177. [Google Scholar]
- [47].Luo R and Berenson D, “A framework for unsupervised online human reaching motion recognition and early prediction,” in Proc. Int. Conf. Intelligent Robots Systems (IROS), 2015, pp. 2426–2433. doi: 10.1109/IROS.2015.7353706. [DOI] [Google Scholar]
- [48].Monfort M, Liu A, and Ziebart BD, “Intent prediction and trajectory forecasting via predictive inverse linear-quadratic regulation,” in Proc. AAAI Conf. Artificial Intelligence, 2015, pp. 3672–3678. [Google Scholar]
- [49].Ramadan A, Choi J, Radcliffe CJ, Popovich JM, and Reeves NP, “Inferring control intent during seated balance using inverse model predictive control,” IEEE Robot. Automat. Lett, vol. 4, no. 2, pp. 224–230, 2018. doi: 10.1109/LRA.2018.2886407. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [50].Yildiz Y, Agogino A, and Brat G, “Predicting pilot behavior in medium scale scenarios using game theory and reinforcement learning,” in Proc. AIAA Modeling Simulation Technologies (MST) Conf, 2013, p. 4908. doi: 10.2514/6.2013-4908. [DOI] [Google Scholar]
- [51].Mainprice J, Hayne R, and Berenson D, “Predicting human reaching motion in collaborative tasks using inverse optimal control and iterative re-planning,” in Proc. IEEE Int. Conf. Robotics Automation (ICRA), 2015, pp. 885–892. doi: 10.1109/ICRA.2015.7139282. [DOI] [Google Scholar]
- [52].Mainprice J and Berenson D, “Human–robot collaborative manipulation planning using early prediction of human motion,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems, 2013, pp. 299–306. doi: 10.1109/IROS.2013.6696368. [DOI] [Google Scholar]
- [53].Ravichandar HC, Kumar A, and Dani AP, “Gaze and motion information fusion for human intention inference,” Int. J. Intell. Robot. Appl, vol. 2, no. 2, pp. 136–148, 2018. doi: 10.1007/s41315-018-0051-0. [DOI] [Google Scholar]
- [54].Koppula HS and Saxena A, “Anticipating human activities using object affordances for reactive robotic response,” in Proc. Robotics: Science Systems, 2013. [DOI] [PubMed] [Google Scholar]
- [55].Dinh HT, Kamalapurkar R, Bhasin S, and Dixon WE, “Dynamic neural network-based robust observers for uncertain nonlinear systems,” Neural Netw, vol. 60, pp. 44–52, Dec. 2014. doi: 10.1016/j.neunet.2014.07.009. [DOI] [PubMed] [Google Scholar]
- [56].Goodwin GC and Aguero J, “Approximate EM algorithms for parameter and state estimation in nonlinear stochastic models,” in Proc. IEEE Conf. Decision Control, European Control Conf, 2005, pp. 368–373. doi: 10.1109/CDC.2005.1582183. [DOI] [Google Scholar]
- [57].Goodwin G and Feuer A, “Estimation with missing data,” Math. Comput. Model. Dyn. Syst, vol. 5, no. 3, pp. 220–244, 1999. doi: 10.1076/mcmd.5.3.220.3681. [DOI] [Google Scholar]
- [58].Bhasin S, Kamalapurkar R, Dinh HT, and Dixon WE, “Robust identification-based state derivative estimation for nonlinear systems,” IEEE Trans. Autom. Control, vol. 58, no. 1, pp. 187–192, 2013. doi: 10.1109/TAC.2012.2203452. [DOI] [Google Scholar]
- [59].Patre PM, MacKunis W, Kaiser K, and Dixon WE, “Asymptotic tracking for uncertain dynamic systems via a multilayer neural network feedforward and rise feedback control structure,” IEEE Trans. Autom. Control, vol. 53, no. 9, pp. 2180–2185, 2008. doi: 10.1109/TAC.2008.930200. [DOI] [Google Scholar]
- [60].Hogan N, “An organizing principle for a class of voluntary movements,” J. Neurosci, vol. 4, no. 11, pp. 2745–2754, 1984. doi: 10.1523/JNEUROSCI.04-11-02745.1984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [61].Lohmiller W and Slotine J-JE, “On contraction analysis for nonlinear systems,” Automatica, vol. 34, no. 6, pp. 683–696, 1998. doi: 10.1016/S0005-1098(98)00019-3. [DOI] [Google Scholar]
- [62].Ravichandar H, Salehi I, and Dani A, “Learning partially contracting dynamical systems from demonstrations,” in Proc. 1st Annu. Conf. Robot Learning, 2017, vol. 78, pp. 369–378. [Google Scholar]
- [63].Ravichandar H and Dani AP, “Human intention inference through interacting multiple model filtering,” in Proc. IEEE Conf. Multisensor Fusion Integration (MFI), 2015, pp. 220–225. doi: 10.1109/MFI.2015.7295812. [DOI] [Google Scholar]
- [64].Bar-Shalom Y, Li XR, and Kirubarajan T, Estimation with Applications to Tracking and Navigation. New York: Wiley, 2001. [Google Scholar]
- [65].Li X-R and Bar-Shalom Y, “Multiple-model estimation with variable structure,” IEEE Trans. Autom. Control, vol. 41, no. 4, pp. 478–493, 1996. doi: 10.1109/9.489270. [DOI] [Google Scholar]
- [66].Li XR, Jilkov VP, Ru J, and Bashi A, “Expected-mode augmentation algorithms for variable-structure multiple-model estimation,” IFAC Proc. Vol, vol. 35, no. 1, pp. 175–180, 2002. doi: 10.3182/20020721-6-ES-1901.00440. [DOI] [Google Scholar]
- [67].Ranatunga I, Lewis FL, Popa DO, and Tousif SM, “Adaptive admittance control for human–robot interaction using model reference design and adaptive inverse filtering,” IEEE Trans. Control Syst. Technol, vol. 25, no. 1, pp. 278–285, 2016. doi: 10.1109/TCST.2016.2523901. [DOI] [Google Scholar]
- [68].Hogan N, “Impedance control: An approach to manipulation: Part II-Implementation,” Trans. ASME, J. Dyn. Syst. Meas. Control, vol. 107, no. 1, pp. 8–16, 1985. doi: 10.1115/1.3140713. [DOI] [Google Scholar]
- [69].Calinon S, Sardellitti I, and Caldwell DG, “Learning-based control strategy for safe human-robot interaction exploiting task and robot redundancies,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems, 2010, pp. 249–254. doi: 10.1109/IROS.2010.5648931. [DOI] [Google Scholar]
- [70].Hoffman G and Breazeal C, “Effects of anticipatory perceptual simulation on practiced human-robot tasks,” Autonom. Robot, vol. 28, no. 4, pp. 403–423, 2010. doi: 10.1007/s10514-009-9166-3. [DOI] [Google Scholar]
- [71].Geravand M, Flacco F, and De Luca A, “Human-robot physical interaction and collaboration using an industrial robot with a closed control architecture,” in Proc. IEEE Int. Conf. Robotics Automation, 2013, pp. 4000–4007. doi: 10.1109/ICRA.2013.6631141. [DOI] [Google Scholar]
- [72].Wilcox R, Nikolaidis S, and Shah J, “Optimization of temporal dynamics for adaptive human-robot interaction in assembly manufacturing,” in Proc. Robot. Sci. Syst, 2012. [Google Scholar]
- [73].Shivashankar V, Kaipa KN, Nau DS, and Gupta SK, “Towards integrating hierarchical goal networks and motion planners to support planning for human-robot teams,” in Proc. AAAI Fall Symp., Artificial Intelligence Human-Robot Interaction, 2014, pp. 13–15. [Google Scholar]
- [74].Shah J and Breazeal C, “An empirical analysis of team coordination behaviors and action planning with application to human–robot teaming,” Hum. Factors, J. Hum. Factors Ergonom. Soc, vol. 52, no. 2, pp. 234–245, 2010. doi: 10.1177/0018720809350882. [DOI] [PubMed] [Google Scholar]
- [75].Yucelen T, Yildiz Y, Sipahi R, Yousefi E, and Nguyen N, “Stability limit of human-in-the-loop model reference adaptive control architectures,” Int. J. Control, vol. 91, no. 10, pp. 2314–2331, 2018. doi: 10.1080/00207179.2017.1342274. [DOI] [Google Scholar]
- [76].Yousefi E, Yildiz Y, Sipahi R, and Yucelen T, “Stability analysis of a human-in-the-loop telerobotics system with two independent time-delays,” IFAC-PapersOnLine, vol. 50, no. 1, pp. 6519–6524, 2017. doi: 10.1016/j.ifacol.2017.08.596. [DOI] [Google Scholar]
- [77].Bemporad A, “Reference governor for constrained nonlinear systems,” IEEE Trans. Autom. Control, vol. 43, no. 3, pp. 415–419, 1998. doi: 10.1109/9.661611. [DOI] [Google Scholar]
- [78].Kolmanovsky I, Garone E, and Di Cairano S, “Reference and command governors: A tutorial on their theory and automotive applications,” in Proc. American Control Conf, 2014, pp. 226–241. doi: 10.1109/ACC.2014.6859176. [DOI] [Google Scholar]
- [79].Kalabić UV, Kolmanovsky IV, and Gilbert EG, “Reduced order extended command governor,” Automatica, vol. 50, no. 5, pp. 1466–1472, 2014. doi: 10.1016/j.automatica.2014.03.012. [DOI] [Google Scholar]
- [80].Berkenkamp F, Turchetta M, Schoellig A, and Krause A, “Safe model-based reinforcement learning with stability guarantees,” in Proc. Advances Neural Information Processing Systems, 2017, pp. 908–919. [Google Scholar]
- [81].Kontoudis GP and Vamvoudakis KG, “Kinodynamic motion planning with continuous-time q-learning: An online, model-free, and safe navigation framework,” IEEE Trans. Neural Netw. Learn. Syst, vol. 30, no. 12, pp. 3803–3817, 2019. doi: 10.1109/TNNLS.2019.2899311. [DOI] [PubMed] [Google Scholar]
- [82].Fisac JF, Akametalu AK, Zeilinger MN, Kaynama S, Gillula J, and Tomlin CJ, “A general safety framework for learning-based control in uncertain robotic systems,” IEEE Trans. Autom. Control, vol. 64, no. 7, pp. 2737–2752, 2018. doi: 10.1109/TAC.2018.2876389. [DOI] [Google Scholar]
- [83].Schouwenaars T, How J, and Feron E, “Receding horizon path planning with implicit safety guarantees,” in Proc. American Control Conf, 2004, pp. 5576–5581. doi: 10.23919/ACC.2004.1384742. [DOI] [Google Scholar]
- [84].Breger LS and How JP, “Safe trajectories for autonomous rendezvous of spacecraft,” J. Guid., Control, Dyn, vol. 31, no. 5, pp. 1478–1489, 2008. doi: 10.2514/1.29590. [DOI] [Google Scholar]
- [85].Mao Y, Dueri D, Szmuk M, and Açıkmeşe B, “Successive convexification of non-convex optimal control problems with state constraints,” IFAC-PapersOnLine, vol. 50, no. 1, pp. 4063–4069, 2017. doi: 10.1016/j.ifacol.2017.08.789. [DOI] [Google Scholar]
- [86].Hovakimyan N, Cao C, Kharisov E, Xargay E, and Gregory IM, “l1 adaptive control for safety-critical systems,” IEEE Control Syst. Mag, vol. 31, no. 5, pp. 54–104, 2011. doi: 10.1109/MCS.2011.941961. [DOI] [Google Scholar]
- [87].Prajna S and Jadbabaie A, “Safety verification of hybrid systems using barrier certificates,” in Proc. Int. Workshop Hybrid Systems, Computation Control, 2004, pp. 477–492. doi: 10.1007/978-3-540-24743-2_32. [DOI] [Google Scholar]
- [88].Ames AD, Xu X, Grizzle JW, and Tabuada P, “Control barrier function based quadratic programs for safety critical systems,” IEEE Trans. Autom. Control, vol. 62, no. 8, pp. 3861–3876, 2017. doi: 10.1109/TAC.2016.2638961. [DOI] [Google Scholar]
- [89].Berkenkamp F, Moriconi R, Schoellig AP, and Krause A, “Safe learning of regions of attraction for uncertain, nonlinear systems with gaussian processes,” in Proc. IEEE 55th Conf. Decision Control, 2016, pp. 4661–4666. doi: 10.1109/CDC.2016.7798979. [DOI] [Google Scholar]
- [90].Wang L, Theodorou EA, and Egerstedt M, “Safe learning of quadrotor dynamics using barrier certificates,” in Proc. IEEE Int. Conf. Robotics Automation, 2018, pp. 2460–2465. doi: 10.1109/ICRA.2018.8460471. [DOI] [Google Scholar]
- [91].Nguyen Q, Hereid A, Grizzle JW, Ames AD, and Sreenath K, “3d dynamic walking on stepping stones with control barrier functions,” in Proc. IEEE 55th Conf. Decision Control (CDC), 2016, pp. 827–834. doi: 10.1109/CDC.2016.7798370. [DOI] [Google Scholar]
- [92].Wu G and Sreenath K, “Safety-critical control of a planar quadrotor,” in Proc. American Control Conf, 2016, pp. 2252–2258. doi: 10.1109/ACC.2016.7525253. [DOI] [Google Scholar]
- [93].Xu X, Tabuada P, Grizzle JW, and Ames AD, “Robustness of control barrier functions for safety critical control,” IFAC-PapersOnLine, vol. 48, no. 27, pp. 54–61, 2015. doi: 10.1016/j.ifacol.2015.11.152. [DOI] [Google Scholar]
- [94].Nguyen Q and Sreenath K, “Exponential control barrier functions for enforcing high relative-degree safety-critical constraints,” in Proc. 2016 American Control Conf. (ACC), pp. 322–328. doi: 10.1109/ACC.2016.7524935. [DOI] [Google Scholar]
- [95].Nguyen Q and Sreenath K, “Safety-critical control for dynamical bipedal walking with precise footstep placement,” IFAC-PapersOnLine, vol. 48, no. 27, pp. 147–154, 2015. doi: 10.1016/j.ifacol.2015.11.167. [DOI] [Google Scholar]
- [96].Yang Y, Yin Y, He W, Vamvoudakis KG, Modares H, and Wunsch DC, “Safety-aware reinforcement learning framework with an actor-criticbarrier structure,” in Proc. American Control Conf, 2019, pp. 2352–2358. doi: 10.23919/ACC.2019.8815335. [DOI] [Google Scholar]
- [97].Salehi I, Yao G, and Dani AP, “Active sampling based safe identification of dynamical systems using extreme learning machines and barrier certificates,” in Proc. IEEE Int. Conf. Robotics Automation, 2019, pp. 22–28. doi: 10.1109/ICRA.2019.8793891. [DOI] [Google Scholar]
- [98].Akash K, Polson K, Reid T, and Jain N, “Improving human-machine collaboration through transparency-based feedback–part I: Human trust and workload model,” IFAC-PapersOnLine, vol. 51, no. 34, pp. 315–321, 2019. doi: 10.1016/j.ifacol.2019.01.028. [DOI] [Google Scholar]
- [99].Nakanishi J, Morimoto J, Endo G, Cheng G, Schaal S, and Kawato M, “Learning from demonstration and adaptation of biped locomotion,” Robot. Autonom. Syst, vol. 47, nos. 2–3, pp. 79–91, 2004. doi: 10.1016/j.robot.2004.03.003. [DOI] [Google Scholar]
- [100].Rahman S, Sadrfaridpour B, and Wang Y, “Trust-based optimal subtask allocation and model predictive control for human-robot collaborative assembly in manufacturing,” in Proc. ASME 2015 Dynamic Systems Control Conf, pp. 1–10. doi: 10.1115/DSCC2015-9850. [DOI] [Google Scholar]
- [101].Rahman SM and Wang Y, “Mutual trust-based subtask allocation for human–robot collaboration in flexible lightweight assembly in manufacturing,” Mechatronics, vol. 54, pp. 94–109, Oct. 2018. doi: 10.1016/j.mechatronics.2018.07.007. [DOI] [Google Scholar]
- [102].Jayaraman SK et al. , “Trust in AV: An uncertainty reduction model of AV-pedestrian interactions,” in Proc. Companion 2018 ACM/IEEE Int. Conf. Human-Robot Interaction, pp. 133–134. doi: 10.1145/3173386.3177073. [DOI] [Google Scholar]
- [103].Srivastava V, Carli R, Langbort C, and Bullo F, “Attention allocation for decision making queues,” Automatica, vol. 50, no. 2, pp. 378–388, 2014. doi: 10.1016/j.automatica.2013.11.028. [DOI] [Google Scholar]
- [104].Peters JR, Srivastava V, Taylor GS, Surana A, Eckstein MP, and Bullo F, “Human supervisory control of robotic teams: Integrating cognitive modeling with engineering design,” IEEE Control Syst. Mag, vol. 35, no. 6, pp. 57–80, 2015. doi: 10.1109/MCS.2015.2471056. [DOI] [Google Scholar]
- [105].Bestick A, Pandya R, Bajcsy R, and Dragan AD, “Learning human ergonomic preferences for handovers,” in Proc. 2018 IEEE Int. Conf. Robotics Automation (ICRA), pp. 3257–3264. doi: 10.1109/ICRA.2018.8461216. [DOI] [Google Scholar]
- [106].Bestick AM, Burden SA, Willits G, Naikal N, Sastry SS, and Bajcsy R, “Personalized kinematics for human-robot collaborative manipulation,” in Proc. 2015 IEEE/RSJ Int. Conf. Intelligent Robots Systems (IROS), 2015, pp. 1037–1044. doi: 10.1109/IROS.2015.7353498. [DOI] [Google Scholar]
- [107].Wang Y and Zhang F, Trends in Control and Decision-Making for Human-Robot Collaboration Systems. New York: Springer-Verlag, 2017. [Google Scholar]
- [108].Yamauchi J, Atman MWS, Hatanaka T, Chopra N, and Fujita M, “Passivity-based control of human-robotic networks with inter-robot communication delays and experimental verification,” in Proc. 2017 IEEE Int. Conf. Advanced Intelligent Mechatronics (AIM), pp. 628–633. doi: 10.1109/AIM.2017.8014087. [DOI] [Google Scholar]
- [109].Lu Y and Song D, “Robustness to lighting variations: An RGB-D indoor visual odometry using line segments,” in Proc. IEEE/RSJ Int. Conf. Intelligent Robots Systems, 2015, pp. 688–694. doi: 10.1109/IROS.2015.7353447. [DOI] [Google Scholar]
- [110].Khoshelham K and Elberink SO, “Accuracy and resolution of kinect depth data for indoor mapping applications,” Sensors, vol. 12, no. 2, pp. 1437–1454, 2012. doi: 10.3390/s120201437. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [111].Lange K, “A gradient algorithm locally equivalent to the EM algorithm,” J. Roy. Statist. Soc. Ser. B, vol. 57, no. 2, pp. 425–437, 1995. doi: 10.1111/j.2517-6161.1995.tb02037.x. [DOI] [Google Scholar]
- [112].Dixon WE, Behal A, Dawson DM, and Nagarkatti S, Nonlinear Control of Engineering Systems: A Lyapunov-Based Approach. Boston: Birkhäuser, 2003. [Google Scholar]
- [113].Recasens A, Khosla A, Vondrick C, and Torralba A, “Where are they looking?” in Proc. Advances Neural Information Processing Systems (NIPS), 2015, pp. 199–207. [Google Scholar]
- [114].Morato C, Kaipa KN, Zhao B, and Gupta SK, “Toward safe human robot collaboration by using multiple kinects based real-time human tracking,” ASME J. Comput. Inform. Sci. Eng, vol. 14, no. 1, pp. 011006–011009, 2014. doi: 10.1115/1.4025810. [DOI] [Google Scholar]
- [115].Dani AP, Chung S-J, and Hutchinson S, “Observer design for stochastic nonlinear systems via contraction-based incremental stability,” IEEE Trans. Autom. Control, vol. 60, no. 3, pp. 700–714, 2015. doi: 10.1109/TAC.2014.2357671. [DOI] [Google Scholar]
- [116].Spong MW, Hutchinson S, and Vidyasagar M, Robot Modeling and Control, vol. 3. Hoboken, NJ: Wiley, 2006. [Google Scholar]
- [117].Dixon WE, “Adaptive regulation of amplitude limited robot manipulators with uncertain kinematics and dynamics,” IEEE Trans. Autom. Control, vol. 52, no. 3, pp. 488–493, 2007. doi: 10.1109/TAC.2006.890321. [DOI] [Google Scholar]
- [118].Liu Y-C and Chopra N, “Controlled synchronization of heterogeneous robotic manipulators in the task space,” IEEE Trans. Robot, vol. 28, no. 1, pp. 268–275, 2012. doi: 10.1109/TRO.2011.2168690. [DOI] [Google Scholar]
- [119].Farjadian AB, Thomsen B, Annaswamy AM, and Woods DD, “Resilient flight control: An architecture for human supervision of automation,” IEEE Trans. Control Syst. Technol, early access, 2020. [Google Scholar]
- [120].Zhang Q, Kim K, and Sharma N. “Prediction of ankle dorsiflexion moment by combined ultrasound sonography and electromyography,” IEEE Trans. Neural Syst. Rehabil. Eng, vol. 28, no. 1, pp. 318–327, 2019. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
