Introduction to Dynamic Systems: Theory, Models, and Applications
Integrates the traditional approach to differential equations with the modern systems and control theoretic approach to dynamic systems, emphasizing theoretical principles and classic models in a wide variety of areas. Provides a particularly comprehensive theoretical development that includes chapters on positive dynamic systems and optimal control theory. Contains numerous problems.
What people are saying - Write a review
We haven't found any reviews in the usual places.
DIFFERENCE AND DIFFERENTIAL EQUATIONS
11 other sections not shown
algebra analysis applied arbitrary assume asymptotically stable Ax(k basis behavior canonical forms chapter characteristic equation characteristic polynomial closed class coefficients column completely controllable components Consider constant continuous-time systems converges corresponding defined definition denote derived determined diagonal difference equation differential equations discrete-time discrete-time system dominant eigenvalue dynamic system eigenvalue elements equal equilibrium point example expressed feedback finite follows Frobenius-Perron Gambler's Ruin geneotype geometric sequence given growth homogeneous equation initial conditions input left eigenvector Liapunov function linear combination linear systems linearly independent Markov chain mathematical multiple nonhomogeneous nonnegative nonzero nth-order nxn matrix optimal control optimal control problem original system output parameters population positive systems possible probability probability vector represents result roots satisfy Sect Show simple solution specific state-transition matrix structure Suppose system equation system matrix system x(k Theorem theory tion trajectory transfer function transform variables vector yields z-transform zero