Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 7 de 7
Filter
Add more filters










Database
Language
Publication year range
1.
Article in English | MEDLINE | ID: mdl-37991914

ABSTRACT

Drones are set to penetrate society across transport and smart living sectors. While many are amateur drones that pose no malicious intentions, some may carry deadly capability. It is crucial to infer the drone's objective to prevent risk and guarantee safety. In this article, a policy error inverse reinforcement learning (PEIRL) algorithm is proposed to uncover the hidden objective of drones from online data trajectories obtained from cooperative sensors. A set of error-based polynomial features are used to approximate both the value and policy functions. This set of features is consistent with current onboard storage memories in flight controllers. The real objective function is inferred using an objective constraint and an integral inverse reinforcement learning (IRL) batch least-squares (LS) rule. The convergence of the proposed method is assessed using Lyapunov recursions. Simulation studies using a quadcopter model are provided to demonstrate the benefits of the proposed approach.

2.
ISA Trans ; 137: 646-655, 2023 Jun.
Article in English | MEDLINE | ID: mdl-36543735

ABSTRACT

Risk mitigation is usually addressed in simulated environments for safety critical control. The migration of the final controller requires further adjustments due to the simulation assumptions and constraints. This paper presents the design of an experience inference algorithm for safety critical control of unknown multi-agent linear systems. The approach is inspired in the close relationship between three main areas of the brain cortex that enables transfer learning and decision making: the hippocampus, the neocortex, and the striatum. The hippocampus is modelled as a stable linear model that communicates to the striatum how the real-world system is expected to behave. The hippocampus model is controlled by an adaptive dynamic programming (ADP) algorithm to achieve an optimal desired performance. The neocortex and the striatum are designed simultaneously by an actor control policy algorithm that ensures experience inference to the real-world system. Experimental and simulations studies are carried out to verify the proposed approach.


Subject(s)
Algorithms , Learning , Computer Simulation , Brain , Hippocampus
3.
IEEE Trans Cybern ; 53(3): 1379-1391, 2023 Mar.
Article in English | MEDLINE | ID: mdl-36129867

ABSTRACT

While autonomous systems can be used for a variety of beneficial applications, they can also be used for malicious intentions and it is mandatory to disrupt them before they act. So, an accurate trajectory inference algorithm is required for monitoring purposes that allows to take appropriate countermeasures. This article presents a closed-loop output error approach for trajectory inference of a class of linear systems. The approach combines the main advantages of state estimation and parameter identification algorithms in a complementary fashion using online data and an estimated model, which is constructed by the state and parameter estimates, that inform about the physics of the system to infer the followed noise-free trajectory. Exact model matching and estimation error cases are analyzed. A composite update rule based on a least-squares rule is also proposed to improve robustness and parameter and state convergence. The stability and convergence of the proposed approaches are assessed via the Lyapunov stability theory under the fulfilment of a persistent excitation condition. Simulation studies are carried out to validate the proposed approaches.

4.
IEEE Trans Cybern ; 52(6): 4485-4494, 2022 Jun.
Article in English | MEDLINE | ID: mdl-33232250

ABSTRACT

In this article, we discuss continuous-time H2 control for the unknown nonlinear system. We use differential neural networks to model the system, then apply the H2 tracking control based on the neural model. Since the neural H2 control is very sensitive to the neural modeling error, we use reinforcement learning to improve the control performance. The stabilities of the neural modeling and the H2 tracking control are proven. The convergence of the approach is also given. The proposed method is validated with two benchmark control problems.


Subject(s)
Algorithms , Nonlinear Dynamics , Computer Simulation , Neural Networks, Computer , Reinforcement, Psychology
5.
ISA Trans ; 122: 88-95, 2022 Mar.
Article in English | MEDLINE | ID: mdl-33941378

ABSTRACT

A solution of the constant cutting velocity problem of quick-return mechanisms is the main concern of this paper. An optimal sliding mode control in the task space is used to achieve uniform and accurate cuts throughout the workpiece. The switching hyperplane is designed to minimize the position error of the slider-dynamics in an infinite horizon. A Jacobian compensator is used to exploit the mechanical advantage and ensure controllability. The velocity profile is constructed in terms of the mechanism and workpiece geometric properties. Stability of the closed-loop dynamics is verified with the Lyapunov stability theory. Experiments are carried out in a quick-return mechanism prototype to validate the proposal.

6.
Neural Netw ; 145: 33-41, 2022 Jan.
Article in English | MEDLINE | ID: mdl-34715533

ABSTRACT

In this paper, a complementary learning scheme for experience transference of unknown continuous-time linear systems is proposed. The algorithm is inspired in the complementary learning properties that exhibit the hippocampus and neocortex learning systems via the striatum. The hippocampus is modelled as pattern-separated data of a human optimized controller. The neocortex is modelled as a Q-reinforcement learning algorithm which improves the hippocampus control policy. The complementary learning (striatum) is designed as an inverse reinforcement learning algorithm which relates the hippocampus and neocortex learning models to seek and transfer the weights of the hidden expert's utility function. Convergence of the proposed approach is analysed using Lyapunov recursions. Simulations are given to verify the proposed approach.


Subject(s)
Algorithms , Neocortex , Hippocampus , Humans , Reinforcement, Psychology
7.
IEEE Trans Neural Netw Learn Syst ; 32(11): 4879-4889, 2021 11.
Article in English | MEDLINE | ID: mdl-33017294

ABSTRACT

In this article, we discuss H2 control for unknown nonlinear systems in discrete time. A discrete-time recurrent neural network is used to model the nonlinear system, and then, the H2 tracking control is applied based on the neural model. Since this neural H2 control is very sensitive to the neural modeling error, we use reinforcement learning and another neural approximator to improve tracking accuracy and robustness of the controller. The stabilities of the neural identifier and the H2 tracking control are proven. The convergence of the approach is also given. The proposed method is validated with the control of the pan and tilt robot and the surge tank.


Subject(s)
Neural Networks, Computer , Nonlinear Dynamics , Reinforcement, Psychology , Humans , Time Factors
SELECTION OF CITATIONS
SEARCH DETAIL
...