Worst-case error bounds for the super-twisting differentiator in presence of measurement noise
Abstract: The super-twisting differentiator, also known as the first-order robust exact differentiator, is a well known sliding mode differentiator. In the absence of measurement noise, it achieves exact reconstruction of the time derivative of a function with bounded second derivative. This note proposes an upper bound for its worst-case differentiation error in the presence of bounded measurement noise, based on a novel Lipschitz continuous Lyapunov function. It is shown that the bound can be made arbitrarily tight and never exceeds the true worst-case differentiation error by more than a factor of two. A numerical simulation illustrates the results and also demonstrates the non-conservativeness of the proposed bound.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.