Abstract
This paper proposes a novel predefined-time adaptive neural tracking control method for uncertain manipulator systems based on Actor-Critic reinforcement learning framework. The proposed control scheme integrates the advantages of predefined-time stability theory and reinforcement learning to achieve fast convergence with guaranteed settling time bounds while handling unknown system dynamics. An Actor neural network is designed to approximate the unknown nonlinear functions and generate control inputs, while a Critic neural network evaluates the cost-to-go function to guide the learning process. The predefined-time convergence is ensured by incorporating specially designed terms into both the control law and the neural network weight update laws. The upper bound of the settling time can be explicitly preset by a single design parameter, independent of initial conditions and system parameters. Rigorous stability analysis based on Lyapunov theory proves that all closed-loop signals are bounded and the tracking error converges to a small neighborhood of the origin within the predefined time. Simulation results on a single-link manipulator system demonstrate the effectiveness and superiority of the proposed control scheme compared with conventional PID control.
Keywords: predefined-time control, actor-critic reinforcement learning, adaptive neural network control, backstepping control
1. Introduction
Robotic manipulators have been extensively deployed in industrial manufacturing, medical surgery, space exploration, and military applications due to their high flexibility, precision, repeatability, and efficiency [1,2]. In these applications, the control system must achieve accurate trajectory tracking while adapting to varying operating conditions and task requirements. However, the control of robotic manipulators remains challenging due to their inherent nonlinearities arising from trigonometric functions in the dynamic equations, strong coupling effects between joints, and inevitable uncertainties stemming from unmodeled dynamics, parameter variations, friction, and external disturbances [3]. Therefore, developing advanced control strategies that simultaneously guarantee tracking performance, fast convergence, and robustness against uncertainties has become a critical research topic in the field of robotics and control engineering.
To address the challenges posed by system uncertainties, numerous advanced control strategies have been developed over the past decades. Adaptive control provides an effective approach to handle parametric uncertainties through online parameter estimation, enabling the controller to adjust its parameters in real-time based on system behavior [4,5]. Neural network (NN)-based control has gained significant attention for its universal approximation capability, which allows it to deal with unknown nonlinear functions without requiring explicit mathematical models [6]. The combination of adaptive control and neural networks, known as adaptive neural network control, has demonstrated excellent performance in handling both parametric and functional uncertainties, and has been successfully applied to various robotic systems [7,8]. Despite these advances, most existing adaptive neural control methods only guarantee asymptotic or exponential convergence, where the settling time depends on initial conditions and system parameters, which may not satisfy the strict timing requirements in practical applications.
In practical robotic applications, fast convergence is often a critical requirement, particularly in time-critical tasks such as assembly operations, surgical procedures, and emergency response scenarios. To achieve convergence in finite time, finite-time control and fixed-time control have been developed based on nonsmooth Lyapunov analysis [9,10]. Finite-time control ensures that the system states converge to the equilibrium within a finite settling time, but this settling time depends on initial conditions, making it difficult to predict or prescribe in advance. Fixed-time control addresses this limitation by ensuring that the settling time is bounded regardless of initial conditions [3]. However, the relationship between the settling time bound and control parameters in fixed-time control is implicit and complex, typically involving multiple design parameters in a nonlinear manner, which complicates the controller tuning process for achieving desired convergence speed.
Recently, predefined-time control has emerged as a promising approach that allows designers to explicitly preset the upper bound of the settling time through a single design parameter [11,12,13]. This feature is particularly attractive for applications with strict timing requirements, as the maximum convergence time can be directly specified according to task demands without complex parameter calculations. Several predefined-time control schemes have been proposed for various systems including rigid spacecraft attitude stabilization [14] and robotic manipulators [15]. However, most existing predefined-time control methods require accurate system models or assume that the system uncertainties are bounded with known bounds, which significantly limits their practical applicability to real-world robotic systems where model parameters are often unknown or time-varying.
On the other hand, reinforcement learning (RL) has shown great potential in control applications due to its ability to learn optimal control policies through interaction with the environment without requiring explicit system models [16,17]. Among various RL architectures, the Actor-Critic (AC) framework is particularly well-suited for continuous control problems, where the Actor network generates control actions and the Critic network evaluates the performance by estimating the value function or cost-to-go [18,19]. The combination of Actor-Critic reinforcement learning and neural network approximation has been successfully applied to various robotic control problems, demonstrating improved adaptability and optimality compared to conventional methods [20,21]. The Actor-Critic structure offers several advantages: the Critic provides a global performance metric for guiding the Actor’s learning, the dual-network architecture separates policy evaluation from policy improvement for enhanced learning efficiency, and the framework naturally accommodates online learning in real-time control scenarios.
Despite the significant progress in each individual area, there remains a gap in the literature regarding the unified treatment of predefined-time convergence, adaptive learning capability, and optimal control for uncertain robotic systems. Most existing predefined-time control methods lack the ability to handle unknown nonlinearities adaptively, while conventional adaptive neural control schemes cannot guarantee predefined-time convergence. The integration of predefined-time stability with Actor-Critic reinforcement learning presents unique theoretical challenges: the predefined-time convergence mechanism must be incorporated into both the control law and the neural network weight update laws in a compatible manner, and the stability analysis must account for the coupled dynamics of tracking errors and weight estimation errors within the predefined-time framework. To the best of the authors’ knowledge, the problem of predefined-time adaptive neural control using Actor-Critic reinforcement learning for robotic manipulators has not been adequately addressed in the existing literature.
Motivated by the above observations, this paper proposes a novel predefined-time adaptive neural tracking control scheme for uncertain single-link manipulator systems based on the Actor-Critic reinforcement learning framework. The main contributions of this paper are summarized as follows:
A novel control framework that synergistically integrates predefined-time stability theory with Actor-Critic reinforcement learning is proposed. The Actor neural network approximates unknown system dynamics and generates control inputs, while the Critic neural network evaluates the cost-to-go function to guide the learning process, achieving both guaranteed convergence time and online learning capability.
Predefined-time neural network weight update laws are designed with specially constructed terms that incorporate the predefined-time convergence mechanism. These update laws ensure the convergence of both tracking errors and weight estimation errors within the predefined time while maintaining the learning and approximation capabilities of the neural networks.
The upper bound of the settling time can be explicitly preset by a single design parameter , satisfying , which is independent of initial conditions and system parameters. This explicit relationship between the design parameter and settling time bound greatly simplifies the controller design process for applications with specific timing requirements.
The remainder of this paper is organized as follows. Section 2 presents the single-link manipulator system model, introduces the Actor-Critic neural network framework, and provides necessary mathematical preliminaries including the predefined-time stability lemma. Section 3 details the controller design procedure, including the predefined-time virtual controller, the Actor-Critic reinforcement learning controller, and the predefined-time weight update laws. Section 4 provides the rigorous stability analysis based on Lyapunov theory. Section 5 presents comprehensive simulation results to validate the effectiveness and superiority of the proposed control scheme. Finally, Section 6 concludes the paper and discusses future research directions.
2. Preliminaries and Problem Formulation
2.1. System Model
Consider a single-link robotic manipulator system described by the following dynamic equation:
| (1) |
where denotes the joint angle, is the angular velocity, is the angular acceleration, represents the control torque, is the moment of inertia, m is the link mass, l is the link length, is the viscous friction coefficient, g is the gravitational acceleration, and d represents the bounded external disturbance satisfying with being a known positive constant.
Define the state variables and . The system (1) can be rewritten in the following state-space form:
| (2) |
where is the control input, is the system output, is a known positive constant, and
| (3) |
is an uncertain nonlinear function.
Assumption 1.
The desired reference trajectory and its derivatives , are continuous and bounded, i.e., there exist positive constants , , such that , , .
2.2. Control Objective
The control objective is to design an adaptive neural tracking controller based on Actor-Critic reinforcement learning such that:
-
(i)
The joint angle tracks the desired trajectory with the tracking error converging to a small neighborhood of the origin within a predefined time , where is a preset design parameter.
-
(ii)
All signals in the closed-loop system remain bounded within the predefined time.
-
(iii)
The Actor-Critic neural networks learn to compensate for the unknown system dynamics online.
2.3. Actor-Critic Neural Network Framework
To handle the unknown nonlinear functions in the system and achieve adaptive optimal control, this paper employs an Actor-Critic reinforcement learning framework. This framework consists of two cooperatively working neural networks: the Actor network is responsible for approximating unknown dynamics and generating control policies, while the Critic network evaluates the control performance and guides the Actor’s learning process.
2.3.1. RBF Basis Function
Both neural networks adopt Radial Basis Functions (RBFs) as basis functions due to their universal approximation capability. For a continuous function defined on a compact set , it can be approximated by an RBF neural network as:
| (4) |
where is the input vector, is the ideal weight vector, l is the number of neural network nodes, is the basis function vector, and is the approximation error satisfying .
The Gaussian function is employed as the basis function:
| (5) |
where is the center of the i-th basis function, and is the width of the Gaussian function.
2.3.2. Critic Network Structure
The Critic network is designed to evaluate the long-term performance of the current control policy. The long-term cost function is defined as:
| (6) |
where is the discount factor, and the instantaneous cost function is defined as:
| (7) |
where and are positive definite weight matrices that penalize the tracking error and control effort, respectively.
Using the RBF neural network to approximate the cost function:
| (8) |
| (9) |
where is the Critic network input, is the ideal weight vector, is the estimated weight vector, is the basis function vector, and satisfies .
When , based on the Bellman equation, the temporal difference (TD) error can be expressed as:
| (10) |
where . The learning objective of the Critic network is to minimize the TD error.
2.3.3. Actor Network Structure
The Actor network is designed to approximate the unknown nonlinear functions in the system and assist in generating control inputs. Define the composite unknown function:
| (11) |
where is the unknown nonlinear term of the system, and is the derivative of the virtual control.
Using the RBF neural network to approximate :
| (12) |
| (13) |
where is the Actor network input, is the ideal weight vector, is the estimated weight vector, is the basis function vector, and satisfies .
2.3.4. Actor-Critic Cooperative Learning Mechanism
The cooperative learning mechanism of the Actor-Critic framework operates as follows:
-
(1)
Critic evaluates policy performance: The Critic network computes the estimated cost function based on the current state and control input, evaluating the quality of the Actor’s current policy. A larger indicates poorer policy performance that requires improvement.
-
(2)
Actor improves control policy: The Actor network utilizes the evaluation information provided by the Critic as feedback to adjust its weights , thereby improving the control policy to minimize the long-term cost.
-
(3)
Online cooperative update: The weights of both networks are updated in real-time during the control process. Through continuous “evaluation-improvement” cycles, the control performance is progressively optimized.
Define the weight estimation errors as:
| (14) |
The specific weight update laws for the Actor-Critic networks will be designed in Section 3, incorporating the predefined-time stability requirements.
Remark 1.
Compared with traditional single neural network adaptive control, the Actor-Critic framework possesses the following advantages: (i) The value function evaluation provided by the Critic offers a global performance metric for the Actor, rather than relying solely on local error information; (ii) The dual-network structure separates policy evaluation from policy improvement, enhancing learning efficiency and stability; (iii) This framework is naturally suited for integration with predefined-time control, allowing the predefined-time convergence mechanism to be incorporated into the weight update laws of both networks.
2.4. Technical Lemmas
Lemma 1
([22]). For any and , the following inequality holds:
(15) where .
Lemma 2
([23]). For () and , the following inequalities hold:
(16)
(17)
Lemma 3.
For and , the following inequality holds:
(18)
Lemma 4
([24]). (Predefined-Time Stability) Consider the system . If there exists a continuous positive definite function and parameters , , such that
(19) then the system is practically predefined-time stable (PPTS), and the convergence region is
(20) where is the settling time satisfying .
Remark 2.
Lemma 4 is fundamental to predefined-time stability theory. The key feature is that the upper bound of the settling time can be explicitly preset through the parameter , independent of the initial conditions and system parameters. This is in contrast to finite-time control where the settling time depends on initial conditions, and fixed-time control where the settling time bound is implicitly determined by multiple parameters.
Lemma 5
([25]). For any and :
(21)
Lemma 6.
(Power Function Inequality) For any and , the following inequality holds:
(22) where is a positive constant depending only on β.
Proof.
Define for . Taking the derivative:
Setting yields the critical point . Since , this is a minimum point. The minimum value is:
Therefore, , which completes the proof. □
3. Actor-Critic Predefined-Time Controller Design
In this section, we present the design of the predefined-time adaptive neural tracking controller based on the Actor-Critic reinforcement learning framework. The control system architecture is illustrated in Figure 1. The Actor network receives system states and reference signals, outputs the control signal to compensate for unknown dynamics. The Critic network evaluates the cost-to-go and provides feedback to guide the Actor’s learning process. Both networks are updated using predefined-time weight update laws.
Figure 1.
Block diagram of the Actor-Critic predefined-time control system.
3.1. Predefined-Time Virtual Controller Design
Define the tracking error variables as:
| (23) |
| (24) |
where is the virtual control law to be designed.
The time derivative of is:
| (25) |
Design the predefined-time virtual controller as:
| (26) |
where is a small positive constant, and
| (27) |
with , being the predefined time parameter, and .
Remark 3.
The virtual controller (26) is specifically designed to achieve predefined-time convergence. The structure ensures that the derivative remains bounded even when approaches zero, thus avoiding the singularity issue that commonly arises in traditional finite-time control designs where terms like with would cause unbounded derivatives.
Consider the Lyapunov function candidate:
| (28) |
Taking the time derivative of and substituting (25) and (26):
| (29) |
Applying Lemma 5:
| (30) |
Therefore:
| (31) |
3.2. Actor-Critic Reinforcement Learning Controller Design
The time derivative of is:
| (32) |
Define the unknown nonlinear function:
| (33) |
3.2.1. Critic Network Design
The Critic network is designed to approximate the cost-to-go function and evaluate the control performance. Define the long-term cost function:
| (34) |
where is a discount factor, and the instantaneous cost function is defined as:
| (35) |
with and being positive definite weight matrices.
The cost-to-go function is approximated by the Critic neural network:
| (36) |
| (37) |
where is the Critic network input, is the ideal weight vector, is the estimated weight vector, is the basis function vector, and is the approximation error.
When , the temporal difference error can be expressed as:
| (38) |
The predefined-time Critic network weight update law is designed as:
| (39) |
where , is the learning rate, , and is a design parameter.
3.2.2. Actor Network Design
The Actor network is designed to approximate the unknown function and generate control inputs. Using RBFNN approximation:
| (40) |
where is the Actor network input, is the ideal weight vector, is the basis function vector, and .
The predefined-time Actor network weight update law is designed as:
| (41) |
where is the learning rate, , , is the Critic feedback gain and is a design parameter.
Remark 4.
The weight update laws (39) and (41) ensure both learning capability and predefined-time convergence by incorporating three essential terms: the first is the standard gradient descent term, which minimizes the approximation or temporal difference error; the second term, , introduces a damping effect to prevent weight drift; and the third term, , acts as the predefined-time convergence component, guaranteeing that the weights converge within the specified time frame.
3.2.3. Predefined-Time Actual Controller
The actual control law is designed as:
| (42) |
where
| (43) |
with and
| (44) |
| (45) |
where is a feedback gain.
Remark 5.
The control law defined in Equations (42)–(45) comprises three key components: a predefined-time convergence term that ensures the tracking error converges within the specified time ; a neural network compensation term, , which provides online compensation for unknown system dynamics; and stabilizing feedback terms, , designed to enhance closed-loop stability.
Remark 6.
The proposed Actor-Critic framework is rooted in the Adaptive Dynamic Programming (ADP) paradigm [16,17,18]. Specifically, the Critic network approximates the value function associated with the Hamilton–Jacobi–Bellman equation:
(46) where is the instantaneous cost that penalizes both tracking error and control effort. The TD error measures the discrepancy between the current value estimate and the Bellman optimality condition. Minimizing drives the Critic toward the true value function.
The term in the Actor update law (41) can be interpreted as an approximate policy gradient step: it adjusts the Actor weights in a direction that reduces the estimated long-term cost , analogous to the policy improvement step in policy iteration methods. Together with the error-driven gradient term , the Actor update simultaneously ensures Lyapunov stability (via error reduction) and approximate optimality (via cost minimization).
It should be noted that due to the integration with predefined-time stability requirements, the damping term and the predefined-time term modify the pure policy gradient direction. Therefore, the optimality guarantee is approximate rather than exact, representing a meaningful design trade-off between guaranteed predefined-time convergence and strict optimality. This is consistent with the ADP literature where stability-constrained policy optimization yields near-optimal rather than globally optimal policies.
4. Stability Analysis
Theorem 1.
Consider the single-link manipulator system (2) satisfying Assumption 1. Under the virtual controller (26), the actual controller (42), and the Actor-Critic neural network weight update laws (39) and (41), if the design parameters satisfy:
,
then the closed-loop system is practically predefined-time stable (PPTS). Specifically:
- (i)
The error signals converge to a compact set within the predefined time .
- (ii)
All signals in the closed-loop system remain bounded.
- (iii)
The convergence region is given by:
(47)
Proof.
Consider the following Lyapunov function candidate:
(48) where and are the weight estimation errors.
From (31), we have:
(49) Taking the derivative of :
(50) Using the neural network approximation (40):
(51) Applying Lemma 1:
(52) Define . Substituting the control law (42):
(53) By Lemma 5:
(54) Expanding and combining terms:
(55) where .
Taking the derivative of :
(56) Substituting the Actor weight update law (41):
(57) Using Young’s inequality for :
(58) Using Lemma 3 for :
(59) Therefore:
(60) From the definition , we have:
(61) Substituting into the third term of (60):
(62) To achieve the target form , we require:
(63) Solving for :
(64) With this choice of , we obtain:
(65) From the definition , we have:
(66) Applying Lemma 6 with :
(67) Multiplying both sides by :
(68) Substituting (66):
(69) where is a bounded positive constant.
Therefore:
(70) To achieve the target form , we require:
(71) Solving for :
(72) With this choice of , we obtain:
(73) Therefore, substituting (73) and (65) into (60):
(74) where is a bounded positive constant.
By the definitions of and , and applying Lemma 6:
(75) From , the cross term involving the Actor network is , which arises because the control law yields and the neural network compensation term in contains . From , substituting the Actor update law (41), the first term is , these two terms cancel exactly for any .
Similarly, for the Critic network:
(76) where is a bounded positive constant.
Combining all terms:
(77) Note that the cross terms cancel:
from cancels with from .
from cancels exactly with from , since the Actor weight update law (41) explicitly includes the factor in the gradient term, and the control law ensures that . This exact cancellation holds for any without requiring any approximation.
Therefore:
(78) where is a positive constant.
Applying Lemma 2, for :
(79) for :
(80) Therefore:
(81) By Lemma 4, the system is practically predefined-time stable with settling time .
From the predefined-time stability, , , , are all bounded.
This completes the proof. □
Remark 7.
By adjusting the predefined time parameter , the upper bound of the settling time can be explicitly preset: . A smaller leads to faster convergence but may require larger control efforts.
Remark 8.
The predefined-time parameters for both Actor and Critic networks are derived from the requirement that the weight estimation error dynamics satisfy the predefined-time stability condition in Lemma 4. The key insight is:
The damping term generates the component through Lemma 6, which dominates when V is large.
The predefined-time term directly generates the component through algebraic substitution, which dominates when V is small.
The combination of both terms ensures predefined-time convergence for all values of .
5. Simulation Results
In this section, numerical simulations are conducted to verify the effectiveness of the proposed Actor-Critic predefined-time control scheme. The simulations are performed on a single-link manipulator system using MATLAB R2025a with Runge–Kutta 4th order integration.
5.1. Simulation Setup
The initial conditions are set as rad and rad/s. The simulation runs for 20 s with a step size of s. For the Actor network with 100 nodes () processing 5-dimensional input , the basis centers are uniformly sampled from the hypercube with width parameter . All weights are initialized to zero, , and bounded by via saturation clipping. The Critic network uses 64 nodes () with 2-dimensional input . Centers are placed on an uniform grid over with width . Weights are similarly initialized as and bounded by . Regarding the discount factor, we implement rather than the theoretical limit . This is a standard simplification in ADP literature [18], where using a sufficiently large makes the term negligible compared to , effectively approximating the infinite discount case while maintaining numerical stability. The system and control parameters are given in Table 1.
Table 1.
System and Control Parameters.
| Parameter | Description | Value | Unit |
|---|---|---|---|
| System Parameters | |||
| m | Link mass | 1.0 | kg |
| l | Link length | 0.5 | m |
| Friction coefficient | 1.0 | N·m·s/rad | |
| g | Gravitational acceleration | 9.8 | m/s2 |
| External disturbance | N·m | ||
| Reference trajectory | rad | ||
| Predefined-Time Parameters | |||
| Predefined time parameter | 2.0 | s | |
| Maximum settling time | s | ||
| Convergence parameter | 0.6 | - | |
| Controller Parameters | |||
| Feedback gain | 100 | - | |
| Small constants | - | ||
| Smoothing parameter | 0.05 | - | |
| Neural Network Parameters | |||
| Actor network nodes | 100 | - | |
| Critic network nodes | 64 | - | |
| Learning rates | 100, 50 | - | |
| Critic feedback gain | 2.0 | - | |
| , | RBF widths | 1.2, 1.0 | - |
| Discount factor | 10 | - | |
| , | Weight bounds | 200, 100 | - |
| PID Controller | |||
| PID gains | 25, 12, 5 | - | |
To verify that the performance is not an artifact of a specific initial condition, we additionally conducted 20 independent Monte Carlo simulations with randomized initial conditions uniformly drawn from rad and rad/s. The statistical results are reported in Table 2.
Table 2.
Statistical Performance over 20 Monte Carlo Runs (Mean ± Std).
| Performance Metric | AC-PT | PID | Improvement |
|---|---|---|---|
| Total RMSE (rad) | 67.0% | ||
| SS RMSE (rad) | 96.9% | ||
| Max SS Error (rad) | 97.5% | ||
| Settling Time (s) | 98.5% | ||
| Satisfied | 20/20 (100%) | N/A | — |
N/A: Since the PID controller’s response is slow and does not meet the basic premise for evaluating this time constraint, this metric is not applicable.
5.2. Tracking Performance Analysis
Figure 2 compares the tracking performance of the proposed AC-PT controller and the conventional PID controller. As shown in Figure 2a, both controllers track the reference trajectory , but the AC-PT controller achieves stabilization within approximately 0.23 s, well within the theoretical upper bound s. In contrast, the PID controller requires approximately 13.84 s to reach the rad tolerance band (Figure 2b). The zoomed steady-state view in Figure 2c confirms that the AC-PT controller maintains the tracking error consistently within the specified tolerance, whereas the PID controller exhibits noticeable residual oscillations. The logarithmic-scale convergence plot in Figure 2d further illustrates the characteristic rapid error decay before , corroborating the predefined-time stability guarantee of Theorem 1. The quantitative comparison is summarized in Table 3: the AC-PT controller achieves 96.9% reduction in steady-state RMSE and 98.3% reduction in settling time compared to PID control.
Figure 2.
Tracking performance comparison: (a) Position tracking showing both controllers following the reference trajectory; (b) Tracking error with rad tolerance band; (c) Steady-state error detail (zoomed view after s); (d) Error convergence in logarithmic scale showing the convergence rate.
Table 3.
Performance Comparison: AC-PT vs PID Control ( s). Single-run results with baseline initial condition rad, rad/s.
| Performance Metric | AC-PT | PID | Improvement |
|---|---|---|---|
| Total RMSE (rad) | 0.0467 | 0.1333 | 65.0% |
| Steady-State RMSE (rad) | 0.0014 | 0.0465 | 96.9% |
| Max Steady-State Error (rad) | 0.0037 | 0.1259 | 97.1% |
| Settling Time to rad (s) | 0.229 | 13.841 | 98.3% |
| Time Within rad (%) | 100.0 | 12.2 | - |
5.3. Neural Network Learning Process
The online learning behavior of the Actor-Critic neural networks is shown in Figure 3. Both the Actor and Critic weight norms (Figure 3a,b) converge to bounded values and remain stable throughout the simulation, confirming that the predefined-time weight update laws incorporating the terms successfully prevent weight divergence. The adaptive parameter (Figure 3c) increases during the transient phase to compensate for system uncertainties and subsequently stabilizes as the tracking error diminishes. Figure 3d shows that both the cost-to-go estimation and the instantaneous cost decrease rapidly during the initial phase, indicating that the Actor-Critic framework effectively optimizes the control policy while compensating for unknown system dynamics.
Figure 3.
Neural network learning process: (a) Actor network weight norm ; (b) Critic network weight norm ; (c) Adaptive parameter ; (d) Cost-to-go estimation and instantaneous cost .
5.4. Effect of Predefined Time Parameter
The influence of the predefined time parameter on control performance is investigated through simulations with s, as shown in Figure 4. Smaller values lead to faster error convergence (Figure 4a), with the system converging before s for s. However, this faster convergence comes at the cost of larger initial control effort (Figure 4b), presenting a trade-off between convergence speed and actuator requirements. Figure 4c demonstrates that all tested values achieve comparable steady-state accuracy, indicating that primarily governs the transient response rather than the ultimate tracking precision. The Lyapunov function evolution in Figure 4d confirms that decreases below its corresponding bound in all cases, thereby validating the predefined-time stability theory of Theorem 1 across different parameter settings.
Figure 4.
Comparison of different predefined time parameters: (a) Tracking error for different ; (b) Control input comparison; (c) Steady-state error comparison; (d) Lyapunov function evolution.
5.5. Comparison with State-of-the-Art Methods
To further substantiate the contributions, the proposed AC-PT controller is compared with two representative methods from the literature: a disturbance-observer-based fixed-time sliding mode controller (FxT-SMC) based on [26], and a predefined-time robust controller without neural networks (PT-Robust) based on [24]. The tracking error comparison in Figure 5a shows that all three advanced controllers significantly outperform PID, with AC-PT and PT-Robust achieving comparable transient performance. The steady-state error detail in Figure 5c reveals that AC-PT achieves the smallest residual error among all methods. The quantitative results demonstrate that the proposed AC-PT method provides competitive convergence speed while offering two key advantages: online learning capability for unknown dynamics compensation (absent in PT-Robust) and an explicit, user-tunable settling time bound (which FxT-SMC cannot directly prescribe).
Figure 5.
Comparison of multiple methods: (a) Tracking error comparison; (b) Control input comparison; (c) Steady-state error comparison; (d) SS RMSE Comparison.
5.6. Robustness Evaluation
To evaluate the robustness of the proposed AC-PT controller under model uncertainties, we conduct simulations under three categories of perturbations: (i) mass uncertainty (m varied by ), (ii) friction coefficient uncertainty ( varied by ), and (iii) increased external disturbance (amplitude scaled to , , and nominal). All tests use the baseline initial condition rad, rad/s with all controller parameters unchanged from Table 1.
The results are summarized in Table 4. The AC-PT controller satisfies the predefined-time guarantee s in all tested scenarios without any parameter re-tuning. The settling time remains within the narrow range s, and the steady-state RMSE is maintained at approximately rad across all cases.
Table 4.
Robustness Evaluation under Parameter Perturbations ( s, s).
| Scenario | Total RMSE | SS RMSE | Settling Time | |
|---|---|---|---|---|
| (rad) | (rad) | (s) | Satisfied | |
| Nominal (, ) | 0.0450 | 0.0015 | 0.209 | Yes |
| Mass Uncertainty | ||||
| kg () | 0.0429 | 0.0015 | 0.207 | Yes |
| kg () | 0.0469 | 0.0016 | 0.212 | Yes |
| Friction Uncertainty | ||||
| () | 0.0448 | 0.0015 | 0.208 | Yes |
| () | 0.0453 | 0.0015 | 0.210 | Yes |
| Increased Disturbance | ||||
| disturbance | 0.0450 | 0.0015 | 0.209 | Yes |
| disturbance | 0.0450 | 0.0015 | 0.209 | Yes |
| disturbance | 0.0450 | 0.0015 | 0.209 | Yes |
This strong invariance is theoretically grounded: the predefined-time convergence rate in Lemma 4 depends on the control gains , and the design parameter , which are independent of the physical parameters. The adaptive parameter and the Actor neural network compensate for the parametric variations and disturbance changes online, as predicted by Theorem 1. The representative tracking error trajectories in Figure 6 confirm that the convergence behavior is qualitatively preserved under all perturbation conditions.
Figure 6.
Robustness evaluation: tracking error under parameter perturbations. (a) Nominal parameters; (b) Mass increased by 30% ( kg); (c) Disturbance amplitude tripled (); (d) Disturbance amplitude quintupled (). The dashed vertical line indicates s. The green band denotes the rad tolerance. All scenarios satisfy .
6. Conclusions
This paper has presented a predefined-time adaptive neural tracking control framework for uncertain single-link manipulator systems, integrating predefined-time stability theory with an Actor-Critic reinforcement learning architecture. The main contribution lies in the synergistic design where the predefined-time convergence mechanism is incorporated into both the control law and the neural network weight update laws, enabling a single parameter to explicitly prescribe the settling time upper bound as , independent of initial conditions and system parameters.
The current work has several limitations that motivate future research. First, the single-link manipulator setting does not capture the coupling effects present in multi-DOF systems; extending the framework to multi-link and redundant manipulators with inter-joint coupling is a natural next step. Second, the current validation is simulation-based; experimental validation on physical robot platforms is essential to assess real-world applicability. Additional future directions include: incorporating input saturation constraints and actuator dynamics; developing event-triggered implementations to reduce computational and communication overhead.
Author Contributions
Conceptualization, Y.Q. and Y.S.; methodology, Y.S.; software, Y.S.; validation, Y.Q., Y.S. and Y.L.; formal analysis, Y.Q. and Y.S.; investigation, J.H.; resources, Y.S.; data curation, Y.L.; writing—original draft preparation, Y.Q. and Y.S.; writing—review and editing, J.H. and Y.S.; visualization, J.H.; supervision, Y.S. and Y.L.; project administration, Y.S.; funding acquisition, Y.Q. and Y.S. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available from the corresponding author upon reasonable request.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This work was supported by the Natural Science Foundation of Jiangsu Province, China, No. BK20240771 and the Key Laboratory of AI and Information Processing, Education Department of Guangxi Zhuang Autonomous Region (Hechi University), No. 2024GXZDSY008.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Gao H., Yang Y., Liu J., Sun C. Reinforcement Learning-Based Admittance Control for Physical Human–Robot Interaction with Output Constraints. IEEE Trans. Autom. Sci. Eng. 2025;22:16334–16345. doi: 10.1109/TASE.2025.3576586. [DOI] [Google Scholar]
- 2.Vyas Y.J., van der Wijk V., Cocuzza S. A Review of Mechanical Design Approaches for Balanced Robotic Manipulation. Robotics. 2025;14:151. doi: 10.3390/robotics14110151. [DOI] [Google Scholar]
- 3.Zhang D., Hu J., Cheng J., Wu Z.G., Yan H. A Novel Disturbance Observer Based Fixed-Time Sliding Mode Control for Robotic Manipulators with Global Fast Convergence. IEEE/CAA J. Autom. Sin. 2024;11:661–672. doi: 10.1109/JAS.2023.123948. [DOI] [Google Scholar]
- 4.Sun Y., Yan B., Shi P., Lim C.C. Consensus for Multiagent Systems Under Output Constraints and Unknown Control Directions. IEEE Syst. J. 2024;17:1035–1044. doi: 10.1109/JSYST.2022.3192573. [DOI] [Google Scholar]
- 5.Liu J., Wang Q.G., Yu J. Event-Triggered Adaptive Neural Network Tracking Control for Uncertain Systems with Unknown Input Saturation Based on Command Filters. IEEE Trans. Neural Netw. Learn. Syst. 2024;35:8702–8707. doi: 10.1109/TNNLS.2022.3224065. [DOI] [PubMed] [Google Scholar]
- 6.Li W., Zhang Z., Ge S.S. Dynamic Gain Reduced-Order Observer-Based Global Adaptive Neural-Network Tracking Control for Nonlinear Time-Delay Systems. IEEE Trans. Cybern. 2023;53:7105–7114. doi: 10.1109/TCYB.2022.3178385. [DOI] [PubMed] [Google Scholar]
- 7.Xie X., Chen W., Xia C., Xing J., Chang L. An RBFNN-Based Prescribed Performance Controller for Spacecraft Proximity Operations with Collision Avoidance. Sensors. 2026;26:108. doi: 10.3390/s26010108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Zhang X., Li H., Zhu G., Zhang Y., Wang C., Wang Y., Su C.Y. Finite-Time Adaptive Quantized Control for Quadrotor Aerial Vehicle with Full States Constraints and Validation on QDrone Experimental Platform. Drones. 2024;8:264. doi: 10.3390/drones8060264. [DOI] [Google Scholar]
- 9.Zhang S., Yang P., Kong L., Li G., He W. A Single Parameter-Based Adaptive Approach to Robotic Manipulators with Finite Time Convergence and Actuator Fault. IEEE Access. 2020;8:15123–15131. doi: 10.1109/ACCESS.2020.2966639. [DOI] [Google Scholar]
- 10.Li G., Chen X., Yu J., Liu J. Adaptive Neural Network-Based Finite-Time Impedance Control of Constrained Robotic Manipulators with Disturbance Observer. IEEE Trans. Circuits Syst. II Express Briefs. 2022;69:1412–1416. doi: 10.1109/TCSII.2021.3109257. [DOI] [Google Scholar]
- 11.Jiménez-Rodríguez E., Muñoz-Vázquez A.J., Sánchez-Torres J.D., Defoort M., Loukianov A.G. A Lyapunov-Like Characterization of Predefined-Time Stability. IEEE Trans. Autom. Control. 2020;65:4922–4927. doi: 10.1109/TAC.2020.2967555. [DOI] [Google Scholar]
- 12.Zhang T., Bai R., Li Y. Practically Predefined-Time Adaptive Fuzzy Quantized Control for Nonlinear Stochastic Systems with Actuator Dead Zone. IEEE Trans. Fuzzy Syst. 2023;31:1240–1253. doi: 10.1109/TFUZZ.2022.3197970. [DOI] [Google Scholar]
- 13.Liu B., Wang W., Li Y., Yi Y., Xie G. Adaptive Quantized Predefined-Time Backstepping Control for Nonlinear Strict-Feedback Systems. IEEE Trans. Circuits Syst. II Express Briefs. 2022;69:3859–3863. doi: 10.1109/TCSII.2022.3175739. [DOI] [Google Scholar]
- 14.Xie S., Chen Q. Adaptive Nonsingular Predefined-Time Control for Attitude Stabilization of Rigid Spacecrafts. IEEE Trans. Circuits Syst. II Express Briefs. 2022;69:189–193. doi: 10.1109/TCSII.2021.3078708. [DOI] [Google Scholar]
- 15.Fan Y., Yang C., Zhan H., Li Y. Neuro-Adaptive-Based Predefined-Time Smooth Control for Manipulators with Disturbance. IEEE Trans. Syst. Man Cybern. Syst. 2024;54:4605–4616. doi: 10.1109/TSMC.2024.3382748. [DOI] [Google Scholar]
- 16.Lewis F.L., Vrabie D., Vamvoudakis K.G. Reinforcement Learning and Feedback Control: Using Natural Decision Methods to Design Optimal Adaptive Controllers. IEEE Control Syst. Mag. 2012;32:76–105. [Google Scholar]
- 17.Ouyang Y., He W., Li X. Reinforcement learning control of a single-link flexible robotic manipulator. IET Control Theory Appl. 2017;11:1426–1433. doi: 10.1049/iet-cta.2016.1540. [DOI] [Google Scholar]
- 18.Vamvoudakis K.G., Lewis F.L. Online actor–critic algorithm to solve the continuous-time infinite horizon optimal control problem. Automatica. 2010;46:878–888. doi: 10.1016/j.automatica.2010.02.018. [DOI] [Google Scholar]
- 19.Guan X., Li Y.X., Hou Z., Ahn C.K. Reinforcement Learning-Based Event-Triggered Adaptive Fixed-Time Optimal Formation Control of Multiple QAAVs. IEEE Trans. Aerosp. Electron. Syst. 2025;61:11849–11864. doi: 10.1109/TAES.2025.3569643. [DOI] [Google Scholar]
- 20.Liu Y.J., Li S., Tong S., Chen C.L.P. Adaptive Reinforcement Learning Control Based on Neural Approximation for Nonlinear Discrete-Time Systems with Unknown Nonaffine Dead-Zone Input. IEEE Trans. Neural Netw. Learn. Syst. 2019;30:295–305. doi: 10.1109/TNNLS.2018.2844165. [DOI] [PubMed] [Google Scholar]
- 21.Zhang Y., Liang X., Li D., Ge S.S., Gao B., Chen H., Lee T.H. Reinforcement Learning-Based Time-Synchronized Optimized Control for Affine Systems. IEEE Trans. Artif. Intell. 2024;5:5216–5231. doi: 10.1109/TAI.2024.3420261. [DOI] [Google Scholar]
- 22.Sun Y., Shi P., Lim C.C. Event-triggered adaptive leaderless consensus control for nonlinear multi-agent systems with unknown backlash-like hysteresis. Int. J. Robust Nonlinear Control. 2021;31:7409–7424. doi: 10.1002/rnc.5692. [DOI] [Google Scholar]
- 23.Hu G., Xu D., Hua W., Jiang B., Shi P., Rudas I.J. Fixed-Time Cooperative Sliding Mode Control for Synchronization of Multilinear Motor Systems. IEEE/ASME Trans. Mechatronics. 2025;31:173–184. doi: 10.1109/TMECH.2025.3585574. [DOI] [Google Scholar]
- 24.Muñoz-Vázquez A.J., Sánchez-Torres J.D., Jiménez-Rodríguez E., Loukianov A.G. Predefined-time robust stabilization of robotic manipulators. IEEE/ASME Trans. Mechatronics. 2019;24:1033–1040. doi: 10.1109/TMECH.2019.2906289. [DOI] [Google Scholar]
- 25.Sun Y., Shi P., Lim C.C. Adaptive consensus control for output-constrained nonlinear multi-agent systems with actuator faults. J. Frankl. Inst. 2022;359:4216–4232. doi: 10.1016/j.jfranklin.2022.03.025. [DOI] [Google Scholar]
- 26.Zhang L., Su Y., Wang Z., Wang H. Fixed-time terminal sliding mode control for uncertain robot manipulators. ISA Trans. 2024;144:364–373. doi: 10.1016/j.isatra.2023.10.011. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data presented in this study are available from the corresponding author upon reasonable request.






