Continuous Time System

Continuous-Time Systems

Luis F. Chaparro , in Signals and Systems using MATLAB, 2011

Publisher Summary

A continuous-time system is a system in which the signals at input and output are continuous-time signals. This chapter connects signals with systems, especially the study of linear time-invariant dynamic systems. Simple examples of systems, ranging from the vocal system to simple circuits, illustrate the use of the linear time-invariant (LTI) model and point to its practical application. At the same time, modulators also show that more complicated systems need to be explored to be able to communicate wirelessly. Although a system is viewed as a mathematical transformation of an input signal (or signals) into an output signal (or signals), it is important to understand that such transformation results from an idealized model of the physical device. A system's approach to the theory of differential equations and some features of transforms are also discussed. The general characteristics attributed to systems are classified such as static or dynamic systems, lumped- or distributed-parameter systems, and passive or active systems are also discussed in the chapter. The Laplace transform allows transient as well as steady-state analysis and converts the solution of differential equations into an algebraic problem and is very significant in the area of classic control. The chapter also provides the concept of transfer function that connects with the impulse response and the convolution integral. The analysis of systems with continuous-time signals by means of transforms is presented. When developing a mathematical model for a continuous-time system it is important to contrast the accuracy of the model with its simplicity and practicality.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123747167000053

REVIEWS

Wing-Kuen Ling , in Nonlinear Digital Filters, 2007

Controllability and observability

For a continuous time system, assume that x(0) = 0. ∀x 1, if ∃t 1 > 0 and u(t) such that x(t 1) = x 1, then the continuous time system is said to be reachable. Similarly, for a discrete time system, assume that x(0) = 0. ∀x 1, if ∃n 1 > 0 and u(n) such that x(n 1) = x 1, then the discrete time system is said to be reachable. For a continuous time system, if ∀x 0, x 1, ∃t 1 > 0 and u(t) such that x(0) = x 0 and x(t 1) = x 1, then the continuous time system is said to be controllable. Similarly, for a discrete time system, if ∀x 0, x 1, ∃n 1 > 0 and u(n) such that x(0) = x 0 and x(n 1 = x 1, then the discrete time system is said to be controllable. For LTI systems, the set of reachable state is R(|BABAnB|), where R(A) is defined as the range of A, that is R(A) = {y : y = A x}. Also, the LTI systems are controllable if and only if R(A) = R n Or in other words, rank(|BABAn B |) = n.

For a continuous time system, ∀x 1, if ∃t 1 > 0 and u(t) such that x 1 can be determined from y(t) for t > t 1, then the continuous time system is said to be observable. Similarly, for a discrete time system, ∀x 1, if ∃n 1 > 0 and u(n) such that x 1 can be determined from y(n) for n > n 1, then the discrete time system is said to be observable. For LTI systems, the set of unobservable state is

N ( [ C CA CA n 1 ] ) ,

where N(A) is defined as the null space of the kernel of A, that is N(A) ≡ {x: Ax = 0}. Also, the LTI systems are observable if and only if

N ( [ C CA CA n 1 ] ) = { 0 } . Or in other words , r a n k ( [ C CA CA n 1 ] ) = n .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123725363500028

Continuous-Time Systems

Luis F. Chaparro , Aydin Akan , in Signals and Systems Using MATLAB (Third Edition), 2019

2.3.2 Time-Invariance

A continuous-time system S is time-invariant if whenever for an input x ( t ) , with a corresponding output y ( t ) = S [ x ( t ) ] , the output corresponding to a shifted input x ( t τ ) (delayed or advanced) is the original output equally shifted in time, y ( t τ ) = S [ x ( t τ ) ] (delayed or advanced). Thus

(2.9) x ( t ) y ( t ) = S [ x ( t ) ] , x ( t τ ) y ( t τ ) = S [ x ( t ± τ ) ] .

That is, the system does not age—its parameters are constant.

A system that satisfies both the linearity and the time-invariance is called Linear Time-Invariant or LTI.

Remarks

1.

It should be clear that linearity and time-invariance are independent of each other. Thus, one can have linear time-varying, or non-linear time-invariant systems.

2.

Although most actual systems are, according to the above definitions, non-linear and time-varying, linear models are used to approximate (around an operating point) the non-linear behavior, and time-invariant models are used to approximate (in short segments) the system's time-varying behavior. For instance, in speech synthesis the vocal system is typically modeled as a linear time-invariant system for intervals of about 20 ms attempting to approximate the continuous variation in shape in the different parts of the vocal system (mouth, cheeks, nose, etc.). A better model for such a system is clearly a linear time-varying model.

3.

In many cases time invariance can be determined by identifying—if possible—the input, the output and letting the rest represent the parameters of the system. If these parameters change with time, the system is time-varying. For instance, if the input x ( t ) and the output y ( t ) of a system are related by the equation y ( t ) = f ( t ) x ( t ) , the parameter of the system is the function f ( t ) , and if it is not constant, the system is time-varying. Thus, the system y ( t ) = A x ( t ) , where A is a constant, is time invariant as can be easily verified. But the AM modulation system given by y ( t ) = cos ( Ω 0 t ) x ( t ) is time-varying as the function f ( t ) = cos ( Ω o t ) is not constant.

The Beginnings of Radio

The names of Nikola Tesla (1856–1943) and Reginald Fessenden (1866–1932) are linked to the invention of amplitude modulation and radio [65,78,7]. Radio was initially called "wireless telegraphy" and then "wireless" [26].

Tesla was a mechanical as well as an electrical engineer, but mostly an inventor. He has been credited with significant contributions to electricity and magnetism in the late 19th and early 20th century. His work is the basis of the alternating current (AC) power system, and the induction motor. His work on wireless communications using the "Tesla coils" was capable of transmitting and receiving radio signals. Although Tesla submitted a patent application for his basic radio before Guglielmo Marconi, it was Marconi who was initially given the patent for the invention of radio (1904). The Supreme Court in 1943 reversed the decision in favor of Tesla [6].

Fessenden has been called the "father of radio broadcasting." His early work on radio led to demonstrations on December 1906, of the capability of point-to-point wireless telephony, and what appears to be the first audio radio broadcasts of entertainment and music ever made to an audience (in this case shipboard radio operators in the Atlantic). Fessenden was a professor of electrical engineering at Purdue University and the first chairman of the electrical engineering department of the University of Pittsburgh in 1893.

AM Communication System

Amplitude modulation (AM) communication systems arose from the need to send an acoustic signal, a "message", over the airwaves using a reasonably sized antenna to radiate it. The size of the antenna depends inversely on the highest frequency present in the message, and voice and music have relatively low frequencies. A voice signal typically has frequencies in the range of 100 Hz to about 5 kHz (the frequencies needed to make a telephone conversation intelligible) while music typically displays frequencies up to about 22 kHz. The transmission of such signals with a practical antenna is impossible. To make the transmission possible, modulation was introduced, i.e., multiplying the message m ( t ) by a periodic signal such as a cosine cos ( Ω 0 t ) , the carrier, with a frequency Ω 0 much larger than those in the acoustic signal. Amplitude modulation provided the larger frequencies needed to reduce the size of the antenna. Thus y ( t ) = m ( t ) cos ( Ω 0 t ) is the signal to be transmitted, and the effect of this multiplication is to change the frequency content of the input.

The AM system is clearly linear, but time-varying. Indeed, if the input is m ( t τ ) , the message delayed τ seconds, the output would be m ( t τ ) cos ( Ω 0 t ) which is not y ( t τ ) = cos ( Ω 0 ( t τ ) ) m ( t τ ) as a time-invariant system would give. Fig. 2.4 illustrates the AM transmitter and receiver. The carrier continuously changes independent of the input and as such the system is time-varying.

Figure 2.4

Figure 2.4. AM modulation: transmitter and receiver.

FM Communication System

In comparison with an AM system, a frequency modulation (FM) system is non-linear and time-varying. An FM modulated signal z ( t ) is given by

z ( t ) = cos ( Ω c t + t m ( τ ) d τ ) ,

where m ( t ) is the input message.

To show the FM system is non-linear, assume we scale the message to γ m ( t ) , for some constant γ, the corresponding output is given by

cos ( Ω c t + γ t m ( τ ) d τ )

which is not the previous output scaled, i.e., γ z ( t ) , thus FM is a non-linear system. Likewise, if the message is delayed or advanced, the output will not be equally delayed or advanced—thus the FM system is not time-invariant.

Vocal System

A remarkable system that we all have is the vocal system (see Fig. 2.5). The air pushed out from the lungs in this system is directed by the trachea through the vocal cords making them vibrate and create resonances similar to those from a musical wind instrument. The generated sounds are then muffled by the mouth and the nasal cavities resulting in an acoustic signal carrying a message. Given the length of the typical vocal system, on average for adult males about 17 cm and 14 cm for adult females, it is modeled as a distributed system and represented by partial differential equations. Due to the complexity of this model, it is the speech signal along with the understanding of the speech production that is used to obtain models of the vocal system. Speech processing is one of the most fascinating areas of electrical engineering.

Figure 2.5

Figure 2.5. Vocal system: principal organs of articulation. Model for speech production.

A typical linear time-invariant model for speech production considers segments of speech of about 20 ms, and for each a low-order LTI system is developed. The input is either a periodic pulse for the generation of voiced sounds (e.g., vowels) or a noise-like signal for unvoiced sounds (e.g., the /sh/ sound). Processing these inputs gives speech-like signals. A linear time-varying model would take into consideration the variations of the vocal system with time and it would be thus more appropriate.

Example 2.4

Consider constant linear capacitors and inductors, represented by ordinary differential equations

d v c ( t ) d t = 1 C i ( t ) , d i L ( t ) d t = 1 L v ( t ) ,

with initial conditions v c ( 0 ) = 0 and i L ( 0 ) = 0 , under what conditions are these time-invariant systems?

Solution: Given the duality of the capacitor and the inductor, we only need to consider one of these. Solving the ordinary differential equation for the capacitor we obtain (the initial voltage v c ( 0 ) = 0 ):

v c ( t ) = 1 C 0 t i ( τ ) d τ .

Suppose then we delay the input current i ( t ) by λ s. The corresponding output is given by

(2.10) 1 C 0 t i ( τ λ ) d τ = 1 C λ 0 i ( ρ ) d ρ + 1 C 0 t λ i ( ρ ) d ρ

by changing the integration variable to ρ = τ λ . For the above equation to equal the voltage across the capacitor delayed λ s, given by

v c ( t λ ) = 1 C 0 t λ i ( ρ ) d ρ ,

we need that i ( t ) = 0 for t < 0 , so that the first integral in the right expression in Equation (2.10) is zero. Thus, the system is time-invariant if the input current i ( t ) = 0 for t < 0 . Putting this together with the condition on the initial voltage v c ( 0 ) = 0 , we can say that for the capacitor to be a linear and time-invariant system it should not be initially energized—i.e., the input i ( t ) = 0 for t < 0 and no initial voltage across the capacitor, v ( 0 ) = 0 . Similarly, using duality, for the inductor to be linear and time-invariant, the input voltage across the inductor v ( t ) = 0 for t < 0 (to guarantee time-invariance) and the initial current in the inductor i L ( 0 ) = 0 (to guarantee linearity). □

Example 2.5

Consider the RLC circuit in Fig. 2.6 consisting of a series connection of a resistor R, an inductor L and a capacitor C. The switch has been open for a very long time and it is closed at t = 0 , so that there is no initial energy stored in either the inductor or the capacitor and the voltage applied to the elements is zero for t < 0 . Obtain an equation connecting the input v ( t ) and output the current i ( t ) .

Figure 2.6

Figure 2.6. RLC circuit with a switch for setting the initial conditions to zero.

Solution: Because of the presence of the capacitor and the inductor, both capable of storing energy, this circuit is represented by a second order ordinary differential equation with constant coefficients. According to Kirchhoff's voltage law:

v ( t ) = R i ( t ) + 1 C 0 t i ( τ ) d τ + L d i ( t ) d t .

To get rid of the integral we find the derivative of v ( t ) with respect to t:

d v ( t ) d t = R d i ( t ) d t + 1 C i ( t ) + L d 2 i ( t ) d t 2 ,

a second order ordinary differential equation, with input the voltage source v ( t ) , and output the current i ( t ) . Because the circuit is not energized for t < 0 , the circuit represented by the above ordinary differential equation is linear and time-invariant. □

Remark

An RLC circuit is represented by an ordinary differential equation of order equal to the number of independent inductors and capacitors in the circuit. If two or more capacitors (two or more inductors) are connected in parallel (in series), sharing the same initial voltage across them (same initial current) we can convert them into an equivalent capacitor with the same initial voltage (equivalent inductor with the same current).

A system represented by a linear ordinary differential equation, of any order N, having constant coefficients, and with input x ( t ) and output y ( t ) :

(2.11) a 0 y ( t ) + a 1 d y ( t ) d t + + d N y ( t ) d t N = b 0 x ( t ) + b 1 d x ( t ) d t + + b M d M x ( t ) d t M t > 0

is linear time-invariant if the system is not initially energized (i.e., the initial conditions are zero, and the input x ( t ) is zero for t < 0 ). If one or more of the coefficients { a i , b i } are functions of time, the system is time-varying.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128142042000120

Techniques in Discrete and Continuous Robust Systems

Ahmad A. Mohammad , J.A. De Abreu-Garcia , in Control and Dynamic Systems, 1996

3 Theoretical Aspects of the LBCT

This is best summarized in the following theorem:

Theorem: For any given stable, minimal balanced DTS (Adb, Bdb, Cdb, Dd, Σ, T), with distinct second order modes, there exists a unique, minimal, stable, and balanced CTS model (Acb, Bcb, Ccb, Dc  =   Dd, Σ), where

B c b = B d b / T , C c b = C d b / T ,

and Acb is the unique solution of the CT Lyapunov equations

A c b + A c b T = B c b B c b T = T B d b B d b T A c b T + A c b = C c b T C c b = T C d b T C d b .

Moreover, the CTS approximation of the DTS model has the following properties:

1)

The CTS model is minimal and stable if the DTS model is minimal and stable.

2)

There exists a solution for Eqs.(84) – (85), for a fixed choice of T, B, and C.

3)

The solution of 2) is unique.

4)

The |H|2 norm of the DTS is equal to the |H|2 norm of the CTS (notice that we only deal with the strictly proper part).

5)

The initial error in the step response is identically zero.

Proof:

(1)

Guaranteed by Lyapunov equations.

(2, 3)

Proved in section 2.

(4)

A direct consequence of matching the Hankel Singular values of the DT and CT systems.

(5)

First, it is noted that both systems share the same feedthrough term D. Moreover, from the initial value theorem, the initial response of the strictly proper part is identically zero.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0090526796800673

12th International Symposium on Process Systems Engineering and 25th European Symposium on Computer Aided Process Engineering

Muxin Sun , ... Efstratios N. Pistikopoulos , in Computer Aided Chemical Engineering, 2015

5.2 Comparison with multiparametric programming

By discretising the continuous time system, the optimization problem can be formulated as conventional multiparametric programming problem ( Pistikopoulos, 2012). Critical regions can be obtained by solving mpQP, where input is assumed constant for each time interval, are shown in Figure 4. Critical regions for mpQP are approximations of the regions for mpDO, the smaller the sample time, the more closely the critical regions approach to the exact solutions in Figure 2. A special region in both cases is the middle one, namely CR08 (a) or CR15 (b), which represents the case that no constraint is active. The others are all with the first couple of time intervals containing active constrains, with same solution structure as CR02 or CR03 in Figure 2. This unifies the definition of critical regions for different branches of multiparametric programing and also provides a way to explore the solution structures of mpDO by solving approximate mpQP.

Figure 4. Critical Regions with sampling time (a) Ts   =   0.1 and (b) Ts   =   0.05.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780444635778500978

The Z-transform

Luis F. Chaparro , Aydin Akan , in Signals and Systems Using MATLAB (Third Edition), 2019

10.6 State Variable Representation

Modern control theory uses the state-variable representation of systems, whether continuous- or discrete-time. In this section we introduce the discrete-time state-variable representation of systems. In many respects, it is very similar to the state-variable representation of continuous-time systems.

State variables are the memory of a system. In the discrete-time, just like in the continuous-time, knowing the state of a system at a present index n provides the necessary information from the past that together with present and future inputs allows us to calculate the present and future outputs of the system. The advantage of a state-variable representation over a transfer function is the inclusion of initial conditions in the analysis and the ability to use it in multiple-input and multiple-output systems.

Assume a discrete-time system is represented by a difference equation (which could have been obtained from an ordinary differential equation representing a continuous-time system) where x [ n ] is the input and y [ n ] the output:

(10.42) y [ n ] + a 1 y [ n 1 ] + + a N y [ n N ] = b 0 x [ n ] + b 1 x [ n 1 ] + + b M x [ n M ] , n 0 ,

where M N and the initial conditions of the system are { y [ i ] , N i 1 } . As in the continuous-time, the state-variable representation of a discrete-time system is not unique. To obtain a state-variable representation from the difference equation in (10.42) let

(10.43) v 1 [ n ] = y [ n 1 ] , v 2 [ n ] = y [ n 2 ] , v N [ n ] = y [ n N ] .

We then obtain from the difference equation the state-variable equations:

(10.44) v 1 [ n + 1 ] = y [ n ] = a 1 v 1 [ n ] a N v N [ n ] + b 0 x [ n ] + + b M x [ n M ] , v 2 [ n + 1 ] = v 1 [ n ] , v N [ n + 1 ] = v N 1 [ n ] ,

and the output equation

y [ n ] = a 1 v 1 [ n ] a N v N [ n ] + b 0 x [ n ] + b 1 x [ n 1 ] + + b M x [ n M ] .

These state and output equations can then be written in a matrix form

(10.45) v [ n + 1 ] = A v [ n ] + B x [ n ] , y [ n ] = c T v [ n ] + d T x [ n ] , n 0 ,

by appropriately defining the matrices A and B, the vectors c and d as well as the state vector v [ n ] and the input vector x [ n ] .

Fig. 10.13 displays the block diagrams for a delay, a constant multiplier and an adder that are used in obtaining block diagrams for discrete-time system. Different from the continuous-time system representation, the representation of discrete-time systems does not require the equivalent of integrators, instead it uses delays.

Figure 10.13

Figure 10.13. Block diagrams of different components used to represent discrete-time systems (top to bottom): delay, constant multiplier and adder.

Example 10.27

A continuous-time system is represented by the ordinary differential equation

d 2 y ( t ) d t 2 + d y ( t ) d t + y ( t ) = x ( t ) t 0 .

To convert this ordinary differential equation into a difference equation we approximate the derivatives by

d y ( t ) d t y ( t ) y ( t T s ) T s , d 2 y ( t ) d t 2 y ( t ) 2 y ( t T s ) + y ( t 2 T s ) T s 2 .

Find the difference equation when T s = 1 and t = n T s , and find the corresponding state and output equations.

Solution: Replacing the approximations of the derivatives for T s = 1 and letting t = n T s = n we obtain the difference equation

( y [ n ] 2 y [ n 1 ] + y [ n 2 ] ) + ( y [ n ] y [ n 1 ] ) + y [ n ] = x [ n ] or y [ n ] y [ n 1 ] + 1 3 y [ n 2 ] = 1 3 x [ n ] ,

which can be realized as indicated by the block diagram in Fig. 10.14. Because the second-order difference equation is realized using two unit delays, this realization is called minimal.

Figure 10.14

Figure 10.14. Block diagram of y[n]−y[n − 1]+(1/3)y[n − 2]=(1/3)x[n] displaying the state variables v 1[n]=y[n − 2] and v 2[n]=y[n − 1].

Letting the outputs of the delays be the state variables

v 1 [ n ] = y [ n 2 ] , v 2 [ n ] = y [ n 1 ] ,

we then have the following matrix equation for the state:

[ v 1 [ n + 1 ] v 2 [ n + 1 ] ] = [ 0 1 1 / 3 1 ] A [ v 1 [ n ] v 2 [ n ] ] + [ 0 1 / 3 ] b x [ n ] .

The output equation is then

y [ n ] = 1 3 v 1 [ n ] + v 2 [ n ] + 1 3 x [ n ] or in matrix form y [ n ] = [ 1 3 1 ] c T [ v 1 [ n ] v 2 [ n ] ] + [ 1 3 ] d x [ n ] .

The state variables are not unique. Indeed we can use an invertible transformation matrix F to define a new set of state variables

w [ n ] = F v [ n ]

with a matrix representation for the new state variables and the output given by

w [ n + 1 ] = F v [ n + 1 ] = F A v [ n ] + F b x [ n ] = F A F 1 w [ n ] + F b x [ n ] , y [ n ] = c T v [ n ] + d x [ n ] = c T F 1 w [ n ] + d x [ n ] .

Example 10.28

A discrete-time system is represented by the difference equation

y [ n ] y [ n 1 ] y [ n 2 ] = x [ n ] + x [ n 1 ]

where x [ n ] is the input and y [ n ] the output. Obtain a state-variable representation for it.

Solution: Notice that this difference equation has x [ n 1 ] as well as x [ n ] in the input. The transfer function corresponding to the difference equation is

H ( z ) = Y ( z ) X ( z ) = 1 + z 1 1 z 1 z 2 ,

i.e., it is not a "constant-numerator" transfer function. A direct realization of the system in this case will not be minimal. Indeed, a block diagram obtained from the difference equation (see Fig. 10.15) shows that three delays are needed to represent this second-order system. It is important to realize that this representation despite being non-minimal is a valid representation. This is different from an analogous situation in the continuous-time representation where the input and its derivatives are present. Differentiators in the continuous-time representation are deemed not acceptable, while delays in the discrete-time representations are.  

Figure 10.15

Figure 10.15. Block diagram of y[n]−y[n − 1]−y[n − 2]=x[n]+x[n − 1] displaying the state variables v 1[n] and v 2[n], and corresponding to a non-minimal realization.

Solution of the State and Output Equations

The solution of the state-variable equations

v [ n + 1 ] = A v [ n ] + B x [ n ] , n 0 ,

can be obtained recursively:

v [ 1 ] = A v [ 0 ] + B x [ 0 ] , v [ 2 ] = A v [ 1 ] + B x [ 1 ] = A 2 v [ 0 ] + AB x [ 0 ] + B x [ 1 ] , v [ n ] = A n v [ 0 ] + k = 0 n 1 A n 1 k B x [ k ] ,

where A 0 = I , the identity matrix. The complete solution is then obtained,

(10.46) y [ n ] = c T A n v [ 0 ] y z i [ n ] zero-input response + k = 0 n 1 c T A n 1 k B x [ k ] + d x [ n ] y z s [ n ] zero-state response .

By definition of the state variables in Equation (10.43) the initial conditions of the state variables coincide with the initial conditions of the system,

(10.47) v 1 [ 0 ] = y [ 1 ] , v 2 [ 0 ] = y [ 2 ] , v N [ 0 ] = y [ N ] .

Using the Z-transform we can obtain a close solution to the state and output equations. Indeed, if the state and output equations are in their matrix form

(10.48) v [ n + 1 ] = A v [ n ] + B x [ n ] , y [ n ] = c T v [ n ] + d T x [ n ] , n 0 ,

calling the Z-transforms of the state variables V i ( z ) = Z ( v i [ n ] ) , for i = 1 , , N ; of the input X m ( z ) = Z ( x [ n m ] ) , for m = 0 , , M , and of the output Y ( z ) = Z ( y [ n ] ) we have the following matrix expression for the Z-transforms of the state equations in (10.48):

z V ( z ) z v [ 0 ] = A V ( z ) + B X ( z ) or ( z I A ) V ( z ) = z v [ 0 ] + B X ( z )

using the Z-transform of v i [ n + 1 ] , and v [ 0 ] is the vector of initial conditions of the state variables and I the unit matrix. Assuming that the inverse of ( z I A ) exists, i.e., det ( z I A ) 0 , we can solve for V ( z ) :

(10.49) V ( z ) = ( z I A ) 1 z v [ 0 ] + ( z I A ) 1 B X ( z ) = Adj ( z I A ) det ( z I A ) z v [ 0 ] + Adj ( z I A ) det ( z I A ) B X ( z )

by expressing the inverse of a matrix in terms of its adjoint and determinant. We can then obtain the Z-transform of the output as

(10.50) Y ( z ) = c T Adj ( z I A ) det ( z I A ) z v [ 0 ] + [ c T Adj ( z I A ) det ( z I A ) B + d ] X ( z ) .

If the initial conditions v [ 0 ] are zero, and the input is x [ n ] , then we find that the transfer function is given by

(10.51) H ( z ) = Y ( z ) X ( z ) = c T Adj ( z I A ) det ( z I A ) b + d .

Example 10.29

Consider the state-variable representation of the system in Example 10.27 with matrices:

A = [ 0 1 1 / 3 1 ] , b = [ 0 1 / 3 ] , c T = [ 1 3 1 ] , d = [ 1 3 ] .

Determine the transfer function of the system.

Solution: Instead of finding the inverse ( z I A ) 1 we can use Cramer's rule to find Y ( z ) . Indeed, writing the state equations with zero initial conditions

[ z 1 1 / 3 z 1 ] ( z I A ) [ V 1 ( z ) V 2 ( z ) ] V ( z ) = [ 0 X ( z ) / 3 ] b X ( z )

according to Cramer's rule we have

V 1 ( z ) = det [ 0 1 X ( z ) / 3 z 1 ] Δ ( z ) = X ( z ) / 3 Δ ( z ) , V 2 ( z ) = det [ z 0 1 / 3 X ( z ) / 3 ] Δ ( z ) = z X ( z ) / 3 Δ ( z ) , Δ ( z ) = z 2 z + 1 / 3 ,

and the Z-transform of the output is obtained after replacing the Z-transforms of the state variables:

Y ( z ) = [ 1 / 3 1 ] [ V 1 ( z ) V 2 ( z ) ] + X ( z ) 3 = 1 / 9 + z / 3 + ( z 2 / 3 z / 3 + 1 / 9 ) z 2 z + 1 / 3 X ( z ) = z 2 / 3 z 2 z + 1 / 3 X ( z ) , so that H ( z ) = Y ( z ) X ( z ) = z 2 / 3 z 2 z + 1 / 3 = 1 / 3 1 z 1 + z 2 / 3 .

Example 10.30

The transfer function of a LTI discrete-time system is

H ( z ) = z 2 1 0.5 z 1 + 0.5 z 2 .

Obtain a minimal state-variable realization (i.e., a realization that only uses two delays, corresponding to the second-order system). Determine the initial conditions of the state variables in terms of initial values of the output.

Solution: The given transfer function is not "constant-numerator" transfer function and the numerator indicates the system has delayed inputs. Letting x [ n ] be the input and y [ n ] the output with Z-transforms X ( z ) and Y ( z ) , respectively, we factor H ( z ) as follows:

H ( z ) = Y ( z ) X ( z ) = z 2 Y ( z ) / W ( z ) × 1 1 0.5 z 1 + 0.5 z 2 W ( z ) / X ( z )

so that we have the following equations:

w [ n ] = 0.5 w [ n 1 ] 0.5 w [ n 2 ] + x [ n ] , y [ n ] = w [ n 2 ] .

The corresponding block diagram in Fig. 10.16 is a minimal realization as it only uses two delays. As shown in the figure, the state variables are defined as

Figure 10.16

Figure 10.16. Minimal realization of a system with transfer function H(z)=z −2/(1 − 0.5z −1 + 0.5z −2).

v 1 [ n ] = w [ n 1 ] , v 2 [ n ] = w [ n 2 ] .

The state and output equations are in matrix form

v [ n + 1 ] = [ 1 / 2 1 / 2 1 0 ] v [ n ] + [ 1 0 ] x [ n ] , y [ n ] = [ 0 1 ] v [ n ] .

Expressing the state variables in terms of the output:

v 1 [ n ] = w [ n 1 ] = y [ n + 1 ] , v 2 [ n ] = y [ n ] ,

we then have v 2 [ 0 ] = y [ 0 ] and v 1 [ 0 ] = y [ 1 ] . The reason for these initial conditions is that in the difference equation

y [ n ] 0.5 y [ n 1 ] + 0.5 y [ n 2 ] = x [ n 2 ]

the input is delayed two samples and so the initial conditions should be y [ 0 ] and y [ 1 ] .  

Example 10.31

Consider the state-variable representation of a system having the matrices

A = [ 0 1 1 / 8 1 / 4 ] , b = [ 0 1 ] , c T = [ 1 0 ] , d = [ 0 ] .

Determine the zero-input response of the system for arbitrary initial conditions v [ 0 ] .

Solution: The zero-input response in the time-domain is

y [ n ] = c T A n v [ 0 ]

requiring one to compute the matrix A n . To do so, consider

Z [ A n ] = n = 0 A n z n = ( I A z 1 ) 1 , det ( I A z 1 ) 0 ,

which can be verified by pre-multiplying by I A z 1 the above equation, giving

[ I A z 1 ] n = 0 A n z n = n = 0 A n z n n = 0 A n + 1 z ( n + 1 ) = I

or that the infinite summation is the inverse of ( I A z 1 ) . In positive powers of z the Z-transform of A n is

Z [ A n ] = ( z 1 [ I z A ] ) 1 = z ( I z A ) 1 .

Now, using the fact that for a 2 × 2 matrix

F = [ a b c d ] F 1 = 1 a d b c [ d b c a ]

provided that the determinant of the matrix, a d b c 0 , we then have

I A z 1 = [ 1 z 1 z 1 / 8 1 z 1 / 4 ] ( I A z 1 ) 1 = 1 1 z 1 / 4 z 2 / 8 [ 1 z 1 / 4 z 1 z 1 / 8 1 ] .

We then need to determine the inverse Z-transform of the four entries to find A n . If we let

P ( z ) = 1 1 z 1 / 4 z 2 / 8 = 2 / 3 1 0.5 z 1 + 1 / 3 1 + 0.25 z 1 with inverse Z-transform p [ n ] = ( 2 3 0.5 n + 1 3 ( 0.25 ) n ) u [ n ]

then

A n = [ p [ n ] 0.25 p [ n 1 ] p [ n 1 ] 0.125 p [ n 1 ] p [ n ] ] ,

which is used in finding the zero-input response

y [ n ] = [ 1 0 ] [ p [ n ] 0.25 p [ n 1 ] p [ n 1 ] 0.125 p [ n 1 ] p [ n ] ] v [ 0 ] = ( p [ n ] 0.25 p [ n 1 ] ) v 1 [ 0 ] + p [ n 1 ] v 2 [ 0 ] .

Canonical Realizations

Just like in the continuous case, discrete-time state-variable realizations have different canonical forms. In this section we will illustrate the process for obtaining a parallel realization to demonstrate the similarity.

The state-variable parallel realization is obtained by realizing each of the partial fraction expansion components of a transfer function. Consider the case of simple poles of the system, so that the transfer function is

H ( z ) = Y ( z ) X ( z ) = i = 1 N A i 1 α i z 1 H i ( z )

so that

Y ( z ) = i = 1 N H i ( z ) X ( z ) Y i ( z ) .

The state variable for Y i ( z ) ( 1 α i z 1 ) = X ( z ) or y i [ n ] α i y i [ n 1 ] = x [ n ] would be v i [ n ] = y i [ n 1 ] so that

v i [ n + 1 ] = y i [ n ] = α i v i [ n ] + x [ n ] , i = 1 , , N .

For the whole system we then have

v 1 [ n + 1 ] = α 1 v 1 [ n ] + x [ n ] , v 2 [ n + 1 ] = α 2 v 2 [ n ] + x [ n ] , v N [ n + 1 ] = α N v N [ n ] + x [ n ] ,

and the output is

y [ n ] = i = 1 N y i [ n ] = i = 1 N α i v i [ n ] + N x [ n ]

or in matrix form

v [ n + 1 ] = A v [ n ] + b x [ n ] , y [ n ] = c T v [ n ] + d x [ n ] , A = [ α 1 0 0 0 α 2 0 0 0 α N ] , b = [ 1 1 1 ] , c T = [ 1 1 1 ] , d = N .

In general, whenever the transfer function is factored in terms of first- and second-order systems, with real coefficients, a parallel realization is obtained by doing a partial fraction expansion in terms of first- and second-order components and realizing each of these components. The following example illustrates the procedure.

Example 10.32

Obtain a parallel state-variable realization from the transfer function

H ( z ) = z 3 ( z + 0.5 ) [ ( z 0.5 ) 2 + 0.25 ] .

Show a block realization of the corresponding state and output equations.

Solution: If we obtain a partial fraction expansion of the form

H ( z ) = 1 ( 1 + 0.5 z 1 ) [ ( 1 0.5 z 1 ) 2 + 0.25 z 2 ] = A 1 1 + 0.5 z 1 + A 2 + A 3 z 1 1 z 1 + 0.5 z 2

where

A 1 = H ( z ) ( 1 + 0.5 z 1 ) | z 1 = 2 = 1 1 z 1 + 0.5 z 2 | z 1 = 2 = 1 5

and then

A 2 + A 3 z 1 1 z 1 + 0.5 z 2 = 1 ( 1 + 0.5 z 1 ) ( 1 z 1 + 0.5 z 2 ) 1 / 5 1 + 0.5 z 1 = 1 0.2 ( 1 z 1 + 0.5 z 2 ) ( 1 + 0.5 z 1 ) ( 1 z 1 + 0.5 z 2 ) = 0.8 0.2 z 1 1 z 1 + 0.5 z 2

so that A 2 = 0.8 and A 3 = 0.2 . This allows us to write the output as

Y ( z ) = 0.2 X ( z ) 1 + 0.5 z 1 Y 1 ( z ) + ( 0.8 0.2 z 1 ) X ( z ) 1 z 1 + 0.5 z 2 Y 2 ( z )

or for two difference equations

y 1 [ n ] + 0.5 y 1 [ n 1 ] = 0.2 x [ n ] , y 2 [ n ] y 2 [ n 1 ] + 0.5 y [ n 2 ] = 0.8 x [ n ] 0.2 x [ n 1 ] ,

for which we need to obtain state variables. Minimal realizations can be obtained as follows. The first system is a "constant-numerator" system, in terms of negative powers of z, and a minimal realization is obtained directly. The transfer function of the second difference equation can be written

Y 2 ( z ) = X ( z ) 1 z 1 + 0.5 z 2 W ( z ) × ( 0.8 0.2 z 1 ) Y 2 ( z ) / W ( z ) ,

from which we obtain the equations

w [ n ] w [ n 1 ] + 0.5 w [ n 2 ] = x [ n ] , y 2 [ n ] = 0.8 w [ n ] 0.2 w [ n 1 ] .

Notice that the realization of these equations only require two delays. If we let

v 1 [ n ] = w [ n 1 ] , v 2 [ n ] = w [ n 2 ] ,

we get

v 1 [ n + 1 ] = w [ n ] = v 1 [ n ] 0.5 v 2 [ n ] + x [ n ] , v 2 [ n + 1 ] = v 1 [ n ] , y 2 [ n ] = 0.8 ( v 1 [ n ] 0.5 v 2 [ n ] + x [ n ] ) 0.2 v 1 [ n ] = 0.6 v 1 [ n ] 0.4 v 2 [ n ] + 0.8 x [ n ] ,

so that

[ v 1 [ n + 1 ] v 2 [ n + 1 ] ] = [ 1 0.5 1 0 ] [ v 1 [ n ] v 2 [ n ] ] + [ 1 0 ] x [ n ] , y 2 [ n ] = [ 0.6 0.4 ] [ v 1 [ n ] v 2 [ n ] ] + 0.8 x [ n ] .

The following script shows how to obtain the state-variable representation for the two components of H ( z ) using the function tf2ss. The second part of the script shows that by using the function ss2tf we can get back the two components of H ( z ) . Notice that these are the same functions used in the continuous case. Fig. 10.17 shows the block diagram of the parallel realization.  

Figure 10.17

Figure 10.17. Block diagram of a parallel state-variable realization of the system with H(z)=z 3/(z + 0.5)[(z−0.5)2 + 0.25]. Notice this is a minimum realization with three delays corresponding to the third-order system.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128142042000211

Mathematical Preliminaries

Erdal Kayacan , Mojtaba Ahmadieh Khanesar , in Fuzzy Neural Networks for Real Time Control Applications, 2016

1.4 Stability Analysis

In order to prove the stability of a nonlinear ordinary differential equation, the Lyapunov stability theory is the most powerful and applicable method. In order to investigate the stability of a system using this method, a positive energy-like function of the states on the trajectory of the differential equation is taken into account, which is called the Lyapunov function. A Lyapunov function decreases along the trajectory of the ODE.

Let the ordinary differential equation of the system be of the following form:

(1.19) x ˙ = F ( x )

where F ( x ) : R n R n and x R n be the state vector of the system. The following theorem holds:

Theorem 1.1

(Stability of Continuous Time Systems [ 4]) Let x = 0 be an equilibrium point and D R n be a domain containing x = 0. Let V : D R be a continually differentiable function such that:

(1.20) V ( 0 ) = 0 , a n d V ( x ) > 0 i n D { 0 }

and

(1.21) V ˙ 0 i n D

then x = 0 is stable. Moreover, if:

(1.22) V ˙ < 0 i n D { 0 }

then x = 0 is asymptotically stable.

It is also possible to investigate the stability of a discrete time difference equation using Lyapunov theory. Let the discrete time difference equation of a system be of the following form:

(1.23) x k + 1 = F ( x k )

where F ( x ) : R n R n and x R n be the state vector of the system. The following theorem holds:

Theorem 1.2

(Stability of Discrete Time Systems [5]) Let x = 0 be an equilibrium point and D R n be a domain containing x = 0. Let V : D R be a continually differentiable function such that

(1.24) V ( 0 ) = 0 , a n d V ( x ) > 0 i n D { 0 }

then x = 0 is stable. Moreover, if:

(1.25) Δ V = V ( x k + 1 ) V ( x k ) < 0 i n D { 0 }

then x = 0 is asymptotically stable.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128026878000013

22nd European Symposium on Computer Aided Process Engineering

Nina P.G. Salau , ... Argimiro R. Secchi , in Computer Aided Chemical Engineering, 2012

2 Formulation and Solution of the Estimation Problem

Consider the following nonlinear dynamic and continuous-time system with discrete-time measurements to be used in the state estimation formulations:

(1) x ˙ = f ( x , u , t , p ) + ω ( t ) ; y k = h ( x k , t ) + v k

Where u denotes the deterministic inputs, x denotes the states, and y denotes the measurements. The process-noise vector, rn(t), and the measurement-noise vector, v k , are assumed to be a white Gaussian random process with zero mean and covariance Q and R k , respectively. The Hybrid Extended Kalman Filter formulation uses a continuous and nonlinear model for state estimation, linearized models of the nonlinear system for state covariance estimation, and discrete measurements. This is often referred to as continuous-discrete extended Kalman filter [5]. Here, the system is linearized at each time step to obtain the local state-space matrices as below:

(2) F = ( f / x ) x , u , t , p ; H = ( h / x ) x , u , t , p

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780444595201501135

Techniques in Discrete-Time Stochastic Control Systems

Gianni Ferretti , ... Riccardo Scattolini , in Control and Dynamic Systems, 1995

C OTHER ALGORITHMS

The identification of the parameters of the continuous-time system is directly dealt with by Zhao et al. [29], considering the following model in the continuoustime domain:

(15) d n y t d t n + i = 1 n α i d n i y t d t n i = i = 1 n β i d n i u t τ d t n i .

Performing a multiple integration on both sides of (15) and approximating the continuous-time integration as a function of the sampled measurements through the linear integral filter defined in [29, 30], yields:

Ι n y k + i = 1 n α i Ι n i y k = i = 1 n β i Ι n i u k τ *

where

Ι j x k = i = 0 nl p i j x k i , j 0 , 1 , , n ,

l is the length factor of the linear integral filter, the coefficients p j i are defined [30] according to the adopted formula of numerical integration (trapezoidal, Simpson, etc.) and τ*   = τ/h  = d  + ε. The estimation of the time delay τ  = τ* h and the parameters αi , βi is then performed by minimizing, via Newton's algorithm, the following LS performance criterion:

J NTDIF ϑ = 1 2 E [ e k ϑ 2 = 1 2 E Ι n y k φ k T ϑ 2 , ϑ = α 1 α n β 1 β n τ * T , φ k = [ Ι n 1 y k Ι 0 y k Ι n 1 u k τ * Ι 0 u k τ * 0 ] T .

Some bias-eliminating techniques are also proposed by Zhao et al. in [29], in order to deal with noisy measurements.

Finally, a rather complex algorithm, based on a Bayesian approach and on the estimation of a set of different models, each one related to a different value of the delay, has been proposed by Juricic [31].

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0090526705800076

From the Ground Up!

Luis F. Chaparro , Aydin Akan , in Signals and Systems Using MATLAB (Third Edition), 2019

0.5.4 The Phasor Connection

The fundamental property of a circuit made up of constant-valued resistors, capacitors and inductors is that its steady-state response to a sinusoid is also a sinusoid of the same frequency [23,21,14]. The effect of the circuit upon the input sinusoid is on its magnitude and phase and it depends on the frequency of the input sinusoid. This is due to the linear and time-invariant nature of the circuit. As we will see in Chapters 3, 4, 5, 10 and 11, this behavior can be generalized to more complex continuous-time as well as discrete-time systems.

To illustrate the connection of phasors with dynamic systems consider the RC circuit ( R = 1 Ω and C = 1 F) in Fig. 0.7. If the input to the circuit is a sinusoidal voltage source v i ( t ) = A cos ( Ω 0 t ) and the voltage across the capacitor v c ( t ) is the output of interest, the circuit can easily be represented by the first-order ordinary differential equation

v i ( t ) = d v c ( t ) d t + v c ( t ) .

Assume that the steady-state response of this circuit (i.e., v c ( t ) as t ) is also a sinusoid

v c ( t ) = C cos ( Ω 0 t + ψ )

of the same frequency as the input, but with amplitude C and phase ψ to be determined. Since this response must satisfy the ordinary differential equation, we have

v i ( t ) = d v c ( t ) d t + v c ( t ) , A cos ( Ω 0 t ) = C Ω 0 sin ( Ω 0 t + ψ ) + C cos ( Ω 0 t + ψ ) = C Ω 0 cos ( Ω 0 t + ψ + π / 2 ) + C cos ( Ω 0 t + ψ ) = C 1 + Ω 0 2 cos ( Ω 0 t + ψ + tan 1 Ω 0 ) .

Comparing the two sides of the above equation gives

C = A 1 + Ω 0 2 , ψ = tan 1 ( Ω 0 ) ,

and for a steady-state response

v c ( t ) = A 1 + Ω 0 2 cos ( Ω 0 t tan 1 ( Ω 0 ) ) .

Comparing the steady-state response v c ( t ) with the input sinusoid v i ( t ) , we see that they both have the same frequency Ω 0 , but the amplitude and phase of the input are changed by the circuit depending on the frequency Ω 0 . Since at each frequency the circuit responds differently, obtaining the frequency response of the circuit is useful not only in analysis but in design of circuits.

The above sinusoidal steady-state response can also be obtained using phasors. Expressing the steady-state response of the circuit as

v c ( t ) = R e [ V c e j Ω 0 t ]

where V c = C e j ψ is the corresponding phasor for v c ( t ) we find that

d v c ( t ) d t = d R e [ V c e j Ω 0 t ] d t = R e [ V c d e j Ω 0 t d t ] = R e [ j Ω 0 V c e j Ω 0 t ] .

Replacing v c ( t ) , d v c ( t ) / d t , obtained above, and

v i ( t ) = R e [ V i e j Ω 0 t ] where V i = A e j 0

in the ordinary differential equation we obtain

R e [ V c ( 1 + j Ω 0 ) e j Ω 0 t ] = R e [ A e j Ω 0 t ] ,

so that

V c = A 1 + j Ω 0 = A 1 + Ω 0 2 e j tan 1 ( Ω 0 ) = C e j ψ

and the sinusoidal steady-state response is

v c ( t ) = R e [ V c e j Ω 0 t ] = A 1 + Ω 0 2 cos ( Ω 0 t tan 1 ( Ω 0 ) ) ,

which coincides with the response obtained above. The ratio of the output phasor V c to the input phasor V i

V c V i = 1 1 + j Ω 0

gives the response of the circuit at frequency Ω 0 . If the frequency of the input is a generic Ω, changing Ω 0 above for Ω gives the frequency response for all possible frequencies.

The concepts of linearity and time-invariance will be used in both continuous as well as discrete-time systems, along with the Fourier representation of signals in terms of sinusoids or complex exponentials, to simplify the analysis and to allow the design of systems. Thus, transform methods such as Laplace and the Z-transform will be used to solve differential and difference equations in an algebraic setup. Fourier representations will provide the frequency perspective. This is a general approach for both continuous and discrete-time signals and systems. The introduction of the concept of transfer function will provide tools for the analysis as well as the design of linear time-invariant systems. The design of analog and discrete filters is the most important application of these concepts. We will look into this topic in Chapters 5, 7 and 12.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128142042000090