In the theory of ordinary differential equations (ODEs), Lyapunov functions, named after Aleksandr Lyapunov, are scalar functions that may be used to prove the stability of an equilibrium of an ODE. Lyapunov functions (also called Lyapunov’s second method for stability) are important to stability theory of dynamical systems and control theory. A similar concept appears in the theory of general state space Markov chains, usually under the name Foster–Lyapunov functions.
For certain classes of ODEs, the existence of Lyapunov functions is a necessary and sufficient condition for stability. There is no general technique for constructing Lyapunov functions for ODEs, however, depending on formulation type, a systematic method to construct Lyapunov functions for ordinary differential equations using their most general form in autonomous cases was given by Prof. Cem Civelek. Though, in many specific cases the construction of Lyapunov functions is known. For instance, according to a lot of applied mathematicians, for a dissipative gyroscopic system a Lyapunov function could not be constructed. However, using the method expressed in the publication above, even for such a system a Lyapunov function could be constructed as per related article by C. Civelek and Ö. Cihanbegendi. In addition, quadratic functions suffice for systems with one state; the solution of a particular linear matrix inequality provides Lyapunov functions for linear systems, and conservation laws can often be used to construct Lyapunov functions for physical systems.
A Lyapunov function for an autonomous dynamical system
with an equilibrium point at is a scalar function that is continuous, has continuous first derivatives, is strictly positive for , and for which the time derivative is non positive (these conditions are required on some region containing the origin). The (stronger) condition that is strictly positive for is sometimes stated as is locally positive definite, or is locally negative definite.
Lyapunov functions arise in the study of equilibrium points of dynamical systems.
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Linear and nonlinear dynamical systems are found in all fields of science and engineering. After a short review of linear system theory, the class will explain and develop the main tools for the quali
To cope with constant and unexpected changes in their environment, robots need to adapt their paths rapidly and appropriately without endangering humans. this course presents method to react within mi
This course offers an introduction to control systems using communication networks for interfacing sensors, actuators, controllers, and processes. Challenges due to network non-idealities and opportun
Various types of stability may be discussed for the solutions of differential equations or difference equations describing dynamical systems. The most important type is that concerning the stability of solutions near to a point of equilibrium. This may be discussed by the theory of Aleksandr Lyapunov. In simple terms, if the solutions that start out near an equilibrium point stay near forever, then is Lyapunov stable. More strongly, if is Lyapunov stable and all solutions that start out near converge to , then is said to be asymptotically stable (see asymptotic analysis).
In mathematics, a dynamical system is a system in which a function describes the time dependence of a point in an ambient space, such as in a parametric curve. Examples include the mathematical models that describe the swinging of a clock pendulum, the flow of water in a pipe, the random motion of particles in the air, and the number of fish each springtime in a lake. The most general definition unifies several concepts in mathematics such as ordinary differential equations and ergodic theory by allowing different choices of the space and how time is measured.
Control theory is a field of control engineering and applied mathematics that deals with the control of dynamical systems in engineered processes and machines. The objective is to develop a model or algorithm governing the application of system inputs to drive the system to a desired state, while minimizing any delay, overshoot, or steady-state error and ensuring a level of control stability; often with the aim to achieve a degree of optimality. To do this, a controller with the requisite corrective behavior is required.
This paper deals with the initial value problem for a semilinear wave equation on a bounded domain and solutions are required to vanish on the boundary of this domain. The essential feature of the situation considered here is that the ellipticity of the sp ...
Path-following control is a critical technology for autonomous vehicles. However, time-varying parameters, parametric uncertainties, external disturbances, and complicated environments significantly challenge autonomous driving. We propose an iterative rob ...
In this thesis we study stability from several viewpoints. After covering the practical importance, the rich history and the ever-growing list of manifestations of stability, we study the following. (i) (Statistical identification of stable dynamical syste ...