LTI system theory

From Wikipedia, the free encyclopedia

LTI system theory or linear time-invariant system theory is a theory in the field of electrical engineering, specifically in circuits, signal processing, and control theory, that investigates the response of a linear, time-invariant system to an arbitrary input signal. Though the standard independent variable is time, it could just as easily be space (as in image processing and field theory) or some other coordinate. Thus an alternately used term is linear translation-invariant. The term linear shift-invariant is the corresponding concept for a discrete-time (sampled) system.

Contents

[edit] Overview

The defining properties of any linear time-invariant system are, of course, linearity and time invariance:

  • Linearity means that the relationship between the input and the output of the system satisfies the superposition property. If the input to the system is the sum of two component signals:
x(t) = (a)x_1(t) + (b)x_2(t) \,
then the output of the system will be
y(t) = (a)y_1(t) + (b)y_2(t) \,
where a and b are constants, and yn(t) is the output resulting from the sole input xn(t).
It can be shown that, given this superposition property, the scaling property follows for any rational scalar. If the output due to input x(t) is y(t), then the output due to input cx(t) is cy(t).
Then, formally, a linear system is a system which exhibits the following property: if the input of the system is
x(t) = \sum_n c_n x_n(t) \,
then the output of the system will be
y(t) = \sum_n c_n y_n(t) \,
for any constants cn and where each yn(t) is the output resulting from the sole input xn(t).
  • Time invariance means that whether we apply an input to the system now or T seconds from now, the output will be identical, except for a time delay of the T seconds. If the output due to input x(t) is y(t), then the output due to input x(tT) is y(tT). More specifically, an input affected by a time delay should effect a corresponding time delay in the output, hence time-invariant.

The fundamental result in LTI system theory is that any LTI system can be characterized entirely by a single function called the system's impulse response. The output of the system is simply the convolution of the input to the system with the system's impulse response. This method of analysis is often called the time domain point-of-view. The same result is true of discrete-time linear shift-invariant systems, in which signals are discrete-time samples, and convolution is defined on sequences.

Relationship between the time domain and the frequency domain
Relationship between the time domain and the frequency domain

Equivalently, any LTI system can be characterized in the frequency domain by the system's transfer function, which is the Laplace transform of the system's impulse response (or Z transform in the case of discrete-time systems). As a result of the properties of these transforms, the output of the system in the frequency domain is the product of the transfer function and the transform of the input. In other words, convolution in the time domain is equivalent to multiplication in the frequency domain.

For all LTI systems, the eigenfunctions, and the basis functions of the transforms, are complex exponentials. This is, if the input to a system is the complex waveform Aexp(st) for some complex amplitude A and complex frequency s, the output will be some complex constant times the input, say Bexp(st) for some new complex amplitude B. The ratio B / A is the transfer function at frequency s.

Because sinusoids are a sum of complex exponentials with complex-conjugate frequencies, if the input to the system is a sinusoid, then the output of the system will also be a sinusoid, perhaps with a different amplitude and a different phase, but always with the same frequency.

LTI system theory is good at describing many important systems. Most LTI systems are considered "easy" to analyze, at least compared to the time-varying and/or nonlinear case. Any system that can be modeled as a linear homogeneous differential equation with constant coefficients is an LTI system. Examples of such systems are electrical circuits made up of resistors, inductors, and capacitors (RLC circuits). Ideal spring–mass–damper systems are also LTI systems, and are mathematically equivalent to RLC circuits.

Most LTI system concepts are similar between the continuous-time and discrete-time (linear shift-invariant) cases. In image processing, the time variable is replaced with 2 space variables, and the notion of time invariance is replaced by two-dimensional shift invariance. When analyzing filter banks and MIMO systems, it is often useful to consider vectors of signals.

A linear system that is not time-invariant can be solved using other approaches such as the Green function method.

[edit] Continuous-time systems

[edit] Time invariance and linear transformation

Let us start with a time-varying system whose impulse response is a 2-dimensional function and see how the condition of time invariance helps us reduce it to one dimension. For example, suppose the input signal is x(t) where its index set is the real line, i.e., t \in \mathbb{R}. The linear operator \mathcal{H} represents the system operating on the input signal. The appropriate operator for this index set is a 2-dimensional function

h(t_1, t_2) \mbox{ where } t_1, t_2 \in \mathbb{R}.

Since \mathcal{H} is a linear operator, the action of the system on the input signal x(t) is a linear transformation represented by the following superposition integral

y(t_1) = \int_{-\infty}^{\infty} h(t_1, t_2) \, x(t_2) \, d t_2.

If the linear operator \mathcal{H} is also time-invariant, then

 h(t_1, t_2) = h(t_1 + \tau, t_2 + \tau)  \qquad  \forall \, \tau \in \mathbb{R}.

For the choice

 \tau = -t_2, \,

it follows that

h(t_1, t_2) = h(t_1 - t_2, 0).  \,

For brevity of notation, we usually drop the zero second argument, and the superposition integral becomes the familiar convolution integral used in filtering:

y(t_1)\, = \int_{-\infty}^{\infty} h(t_1 - t_2) \, x(t_2) \, d t_2 \,
= (h * x) (t_1)\,
= (x * h) (t_1) \,       (commutativity)

Thus, the convolution integral represents the response of a linear, time-invariant system to any input function. At the moment t1, this is a weighted-average of the input function, where the weighting function is h( − t2), and it has been shifted by the amount t1.  As t1 changes, this is called a moving (or sliding) average. For most systems (i.e. most impulse responses), the output is a constantly-updated average of the most recent portion of the input function.

The term impulse response for h(t) refers to the case x(t) = δ(t), a Dirac delta function. The response to this "impulse" is:

 (h * \delta) (t)  = \int_{-\infty}^{\infty} h(t - \tau) \, \delta (\tau) \, d \tau = h(t),

because of the sifting property of δ(τ)).

[edit] Exponentials as eigenfunctions

An eigenfunction is a function for which the output of the operator is the same function, just scaled by some amount. In symbols,

\mathcal{H}f = \lambda f,

where f is the eigenfunction and λ is the eigenvalue, a constant.

The exponential functions est, where s \in \mathbb{C}, are eigenfunctions of a linear, time-invariant operator. A simple proof illustrates this concept.

Suppose the input is x(t) = est. The output of the system with impulse response h(t) is then

\int_{-\infty}^{\infty} h(t - \tau)  e^{s \tau}  d \tau

which is equivalent to the following by the commutative property of convolution

\int_{-\infty}^{\infty} h(\tau) \, e^{s (t - \tau)} \, d \tau
 \quad = e^{s t} \int_{-\infty}^{\infty} h(\tau) \, e^{-s \tau} \, d \tau
 \quad = e^{s t} H(s),

where

H(s) = \int_{-\infty}^\infty h(t) e^{-s t} d t

is dependent only on the parameter s.

So, est is an eigenfunction of an LTI system because the system response is the same as the input times the constant H(s).

[edit] Fourier and Laplace transforms

The eigenfunction property of exponentials is very useful for both analysis and insight into LTI systems. The Laplace transform

H(s) = \mathcal{L}\{h(t)\} = \int_{-\infty}^\infty h(t) e^{-s t} d t

is exactly the way to get the eigenvalues from the impulse response. Of particular interest are pure sinusoids, i.e. exponentials of the form exp(jωt) where \omega \in \mathbb{R} and j = \sqrt{-1}. These are generally called complex exponentials even though the argument is purely imaginary. The Fourier transform H(j \omega) = \mathcal{F}\{h(t)\} gives the eigenvalues for pure complex sinusoids. Both of H(s) and H(jω) are called the system function, system response, or transfer function.

The Laplace transform is usually used in the context of one-sided signals, i.e. signals that are zero for all values of t less than some value. Usually, this "start time" is set to zero, for convenience and without loss of generality, with the transform integral being taken from zero to infinity (the transform shown with lower limit of integration of negative infinity is formally known as the bilateral Laplace transform).

The Fourier transform is used for analyzing systems that process signals that are infinite in extent, such as modulated sinusoids, even though it can not be directly applied to input and output signals that are not square integrable. The Laplace transform actually works directly for these signals if they are zero before a start time, even if they are not square integrable, for stable systems. The Fourier transform is often applied to spectra of infinite signals via the Wiener–Khinchin theorem even when Fourier transforms of the signals do not exist.

Due to the convolution property of both of these transforms, the convolution that gives the output of the system can be transformed to a multiplication in the transform domain, given signals for which the transforms exist

y(t) = (h*x)(t) = \int_{-\infty}^\infty h(t - \tau) x(\tau) d \tau
\quad = \mathcal{L}^{-1}\{H(s)X(s)\}

Not only is it often easier to do the transforms, multiplication, and inverse transform than the original convolution, but one can also gain insight into the behavior of the system from the system response. One can look at the modulus of the system function |H(s)| to see whether the input exp(st) is passed (let through) the system or rejected or attenuated by the system (not let through).

[edit] Examples

A simple example of an LTI operator is the derivative:

 \frac{d}{dt} \left( c_1 x_1(t) + c_2 x_2(t) \right) = c_1 x'_1(t) + c_2 x'_2(t),
 \frac{d}{dt} x(t-\tau) = x'(t-\tau).

When the Laplace transform of the derivative is taken, it transforms to a simple multiplication by the Laplace variable s.

 \mathcal{L}\left\{\frac{d}{dt}x(t)\right\} = s X(s)

That the derivative has such a simple Laplace transform partly explains the utility of the transform.

Another simple LTI operator is an averaging operator

 \mathcal{A}\left\{x(t)\right\} = \int_{t-a}^{t+a} x(\lambda) d \lambda .

It is linear because of the linearity of integration

 \mathcal{A}\left\{c_1 x_1(t) + c_2 x_2(t) \right\}
 = \int_{t-a}^{t+a} \left( c_1 x_1(\lambda) + c_2 x_2(\lambda) \right) d \lambda
 = c_1 \int_{t-a}^{t+a} x_1(\lambda) d \lambda + c_2 \int_{t-a}^{t+a} x_2(\lambda) d \lambda
 = c_1 \mathcal{A}\left\{x_1(t) \right\} + c_2 \mathcal{A}\left\{x_2(t) \right\} .

It is time invariant too

 \mathcal{A}\left\{x(t-\tau)\right\}
 = \int_{t-a}^{t+a} x(\lambda-\tau) d \lambda
 = \int_{(t-\tau)-a}^{(t-\tau)+a} x(\xi) d \xi
 =  \mathcal{A}\{x\}(t-\tau) .

Indeed, \mathcal{A} can be written as a convolution with the box function Π(t).

 \mathcal{A}\left\{x(t)\right\} = \int_{-\infty}^\infty \Pi\left(\frac{\lambda-t}{2a}\right) x(\lambda) d \lambda ,

where the box function is

\Pi(t) = \left\{ \begin{matrix} 1 & |t| < 1/2 \\ 0 & |t| > 1/2 \end{matrix} \right. .

[edit] Important system properties

Some of the most important properties of a system are causality and stability. Causality is a necessity if the independent variable is time, but not all systems have time as an independent variable. For example, a system that processes still images does not need to be causal. Non-stable systems can be built and can be useful in many circumstances. Even non-real systems can be built and are very useful in many contexts.

[edit] Causality

Main article: Causal system

A system is causal if the output depends only on present and past inputs. A necessary and sufficient condition for causality is

h(t) = 0 \quad \forall t < 0,

where h(t) is the impulse response. It is not possible in general to determine causality from the Laplace transform, because the inverse transform is not unique. When a region of convergence is specified, then causality can be determined.

[edit] Stability

Main article: BIBO stability

A system is bounded-input, bounded-output stable (BIBO stable) if, for every bounded input, the output is finite. Mathematically, if every input satisfying

\ \|x(t)\|_\infty < \infty

leads to an output satisfying

\ \|y(t)\|_\infty < \infty

(that is, a finite maximum absolute value of x(t) implies a finite maximum absolute value of y(t)), then the system is stable. A necessary and sufficient condition is that h(t), the impulse response, is in L1 (has a finite L1 norm):

\ \|h(t)\|_1 = \int_{-\infty}^\infty |h(t)| dt < \infty.

In the frequency domain, the region of convergence must contain the imaginary axis s = jω.

As an example, the ideal low-pass filter with impulse response equal to a sinc function is not BIBO stable, because the sinc function does not have a finite L1 norm. Thus, for some bounded input, the output of the ideal low-pass filter is unbounded. In particular, if the input is zero for t < 0\, and equal to a sinusoid at the cut-off frequency for t > 0\,, then the output will be unbounded for all times other than the zero crossings.

[edit] Discrete-time systems

Almost everything in continuous-time systems has a counterpart in discrete-time systems.

[edit] Discrete-time systems from continuous-time systems

In many contexts, a discrete time (DT) system is really part of a larger continuous time (CT) system. For example, a digital recording system takes an analog sound, digitizes it, possibly processes the digital signals, and plays back an analog sound for people to listen to.

Formally, the DT signals studied are almost always uniformly sampled versions of CT signals. If x(t) is a CT signal, then an analog to digital converter will transform it to the DT signal x[n], with

x[n] = x(nT),

where T is the sampling period. It is very important to limit the range of frequencies in the input signal for faithful representation in the DT signal, since then the sampling theorem guarantees that no information about the CT signal is lost. A DT signal can only contain a frequency range of 1 / (2T); other frequencies are aliased to the same range.

[edit] Time invariance and linear transformation

Let us start with a time-varying system whose impulse response is a two dimensional function and see how the condition of time-invariance helps us reduce it to one dimension. For example, suppose the input signal is x[n] where its index set is the integers, i.e., n \in \mathbb{Z}. The linear operator \mathcal{H} represents the system operating on the input signal. The appropriate operator for this index set is a two-dimensional function

h[n_1, n_2] \mbox{ where } n_1, n_2 \in \mathbb{Z}.

Since \mathcal{H} is a linear operator, the action of the system on the input signal x[n] is a linear transformation represented by the following superposition sum

y[n_1] = \sum_{n_2=-\infty}^{\infty} h[n_1, n_2] \, x[n_2],

If the linear operator \mathcal{H} is also time-invariant, then

 h[n_1, n_2] = h[n_1 + m, n_2 + m]  \qquad  \forall \, m \in \mathbb{Z}.

If we let

 m = -n_2, \,

then it follows that

h[n_1, n_2] = h[n_1 - n_2, 0]. \,

We usually drop the zero second argument to h[n1,n2] for brevity of notation so that the superposition integral now becomes the familiar convolution sum used in filtering

y[n_1] = \sum_{n_2=-\infty}^{\infty} h[n_1 - n_2] \, x[n_2] = (h * x) [n_1].

Thus, the convolution sum represents the effect of a linear, time-invariant system on any input function. For a finite-dimensional analog, see the article on a circulant matrix.

[edit] Impulse response

If we input a discrete delta function to this system, the result of the LTI transformation is known as the impulse response because the delta function is an ideal impulse. We illustrate this idea as follows:

 (h * \delta) [n] = \sum_{m=-\infty}^{\infty} h[n - m] \, \delta [m] = h[n],

(by the sifting property of the delta function).

Note that

h[n] = h[n_1 - n_2, 0] \,\!\mbox{ where } n = n_1 - n_2,

so that h[n] is the impulse response of the system.

The impulse response can be used to find the response of any input in the following way. Again using the sifting property of the δ[n], we can write any input as a superposition of deltas:

x[n] = \sum_{m=-\infty}^\infty x[m] \delta[n-m].

Applying the system to the input,

\mathcal{H} x[n] = \mathcal{H} \sum_{m=-\infty}^\infty x[m] \delta[n-m]
\quad = \sum_{m=-\infty}^\infty \mathcal{H} x[m] \delta[n-m] (because \mathcal{H} is linear and can pass inside the sum)
\quad = \sum_{m=-\infty}^\infty x[n] \mathcal{H} \delta[n-m] (because x[m] is constant in n and \mathcal{H} is linear)
\quad = \sum_{m=-\infty}^\infty x[m] h[n-m] (by definition of h[n])

The system output is now expressed in terms of the input and the impulse response only, showing that all information about the system is contained in the impulse response h[n].

[edit] Exponentials as eigenfunctions

An eigenfunction is a function for which the output of the operator is the same function, just scaled by some amount. In symbols,

\mathcal{H}f = \lambda f,

where f is the eigenfunction and λ is the eigenvalue, a constant.

The exponential functions zn = esTn, where n \in \mathbb{Z}, are eigenfunctions of a linear, time-invariant operator. T \in \mathbb{R} is the sampling interval, and z = e^{sT}, \ z,s \in \mathbb{C}. A simple proof illustrates this concept.

Suppose the input is x[n] = \,\!z^n. The output of the system with impulse response h[n] is then

\sum_{m=-\infty}^{\infty} h[n-m] \, z^m

which is equivalent to the following by the commutative property of convolution

\sum_{m=-\infty}^{\infty} h[m] \, z^{(n - m)}
 \quad = z^n \sum_{m=-\infty}^{\infty} h[m] \, z^{-m}
 \quad = z^n H(z),

where

H(z) = \sum_{m=-\infty}^\infty h[m] z^{-m}

is dependent only on the parameter z.

So, zn is an eigenfunction of an LTI system because the system response is the same as the input times the constant H(z).

[edit] Z and discrete-time Fourier transforms

The eigenfunction property of exponentials is very useful for both analysis and insight into LTI systems. The Z transform

H(z) = \mathcal{Z}\{h[n]\} = \sum_{n=-\infty}^\infty h[n] z^{-n}

is exactly the way to get the eigenvalues from the impulse response. Of particular interest are pure sinusoids, i.e. exponentials of the form ejωn, where \omega \in \mathbb{R}. These can also be written as zn with z = ejω. These are generally called complex exponentials even though the argument is purely imaginary. The Discrete-time Fourier transform (DTFT) H(e^{j \omega}) = \mathcal{F}\{h[n]\} gives the eigenvalues of pure sinusoids. Both of H(z) and H(ejω) are called the system function, system response, or transfer function'.

The Z transform is usually used in the context of one-sided signals, i.e. signals that are zero for all values of t less than some value. Usually, this "start time" is set to zero, for convenience and without loss of generality. The Fourier transform is used for analyzing signals that are infinite in extent.

Due to the convolution property of both of these transforms, the convolution that gives the output of the system can be transformed to a multiplication in the transform domain.

y[n] = (h*x)[n] = \sum_{m=-\infty}^\infty h[n-m] x[m]
\quad = \mathcal{Z}^{-1}\{H(z)X(z)\}

Just as with the Laplace transform transfer function in continuous-time system analysis, the Z transform makes it easier to analyze systems and gain insight into their behavior. One can look at the modulus of the system function |H(z)| to see whether the input zn is passed (let through) by the system, or rejected or attenuated by the system (not let through).

[edit] Examples

A simple example of an LTI operator is the delay operator D{x}[n]: = x[n − 1].

 D \left( c_1 x_1[n] + c_2 x_2[n] \right) = c_1 x_1[n-1] + c_2 x_2[n-1] = c_1 Dx_1[n] + c_2 Dx_2[n],
 D\{x[n-m]\} = x[n-m-1] = x[(n-1)-m] = D\{x\}[n-m]. \,

When the Z transform of the delay operator is taken, it transforms to a simple multiplication by z-1:

 \mathcal{Z}\left\{Dx[n]\right\} = z^{-1} X(z).

That the delay operator has such a simple Z transform partly explains the utility of the transform.

Another simple LTI operator is an averaging operator

 \mathcal{A}\left\{x[n]\right\} = \sum_{k=n-a}^{n+a} x[k].

It is linear because of the linearity of sums:

 \mathcal{A}\left\{c_1 x_1[n] + c_2 x_2[n] \right\}
 = \sum_{k=n-a}^{n+a} \left( c_1 x_1[k] + c_2 x_2[k] \right)
 = c_1 \sum_{k=n-a}^{n+a} x_1[k] + c_2 \sum_{k=n-a}^{n+a} x_2[k]
 = c_1 \mathcal{A}\left\{x_1[n] \right\} + c_2 \mathcal{A}\left\{x_2[n] \right\} .

It is time invariant too:

 \mathcal{A}\left\{x[n-m]\right\}
 = \sum_{k=n-a}^{n+a} x[k-m]
 = \sum_{k'=(n-m)-a}^{(n-m)+a} x[k']
 =  \mathcal{A}\left\{x\right\}[n-m] .

[edit] Important system properties

Some of the most important properties of a system are causality and stability. Unlike CT systems, non-causal DT systems can be realized. It is trivial to make an acausal FIR system causal by adding delays. It is even possible to make acausal IIR systems (See Vaidyanathan and Chen, 1995). Non-stable systems can be built and can be useful in many circumstances. Even non-real systems can be built and are very useful in many contexts.

[edit] Causality

Main article: Causal system

A system is causal if the output depends only on present and past inputs. A necessary and sufficient condition for causality is

h[n] = 0 \ \forall n < 0,

where h[n] is the impulse response. It is not possible in general to determine causality from the Z transform, because the inverse transform is not unique. When a region of convergence is specified, then causality can be determined.

[edit] Stability

Main article: BIBO stability

A system is bounded input, bounded output stable (BIBO stable) if, for every bounded input, the output is finite. Mathematically, if

\ ||x[n]||_\infty < \infty

implies that

\ ||y[n]||_\infty < \infty

(that is, if bounded input implies bounded output, in the sense that the maximum absolute values of x[n] and y[n] are finite), then the system is stable. A necessary and sufficient condition is that h[n], the impulse response, satisfies

||h[n]||_1 = \sum_{n = -\infty}^\infty |h[n]| < \infty.

In the frequency domain, the region of convergence must contain the unit circle | z | = 1.

[edit] See also

[edit] References

  • Boaz Porat: A Course in Digital Signal Processing, Wiley, ISBN 0471149616
  • P. P. Vaidyanathan and T. Chen (May 1995). "Role of anticausal inverses in multirate filter banks -- Part I: system theoretic fundamentals". IEEE Trans. Signal Proc. 43: 1090. doi:10.1109/78.382395. 
  • P. P. Vaidyanathan and T. Chen (May 1995). "Role of anticausal inverses in multirate filter banks -- Part II: the FIR case, factorizations, and biorthogonal lapped transforms". IEEE Trans. Signal Proc. 43: 1103. doi:10.1109/78.382396.