Filter design
From Wikipedia, the free encyclopedia
Filter design is the process of designing a filter (in the sense in which the term is used in signal processing, statistics, and applied mathematics), often a linear shift-invariant filter, which satisfies a set of requirements, some of which are contradictory. The problem is to find a realization of the filter which meets each of the requirements to a sufficient degree to make it useful.
The filter design process can be described as an optimization problem where each requirement contributes with a term to an error function which should be minimized. Certain parts of the design process can be automated, but normally an experienced electrical engineer is needed to get a good result.
Contents |
[edit] Typical design requirements
Typical requirements which are considered in the design process are
- The filter should have a specific frequency function
- The filter should have a specific impulse response
- The filter should be causal
- The filter should be stable
- The filter should be localized
- The computational complexity of the filter should be low
- The filter should be implemented in a particular hardware or software.
[edit] The frequency function
Typical examples of frequency function are
- A low-pass filter is used to block unwanted high-frequency signals.
- A high-pass filter passes high frequencies fairly well; it is helpful as a filter to block any unwanted low frequency components.
- A band-pass filter passes a limited range of frequencies.
- A band-stop filter passes frequencies above or below a certain range. This is also known as a notch filter.
- A low-shelf filter passes all frequencies, but boosts or cuts frequencies below the cutoff frequency by specified amount.
- A high-shelf filter passes all frequencies, but boosts or cuts frequencies above the cutoff frequency by specified amount.
- A peak EQ filter makes a peak or a dip in the frequency response, commonly used in graphic equalizers.
- An all-pass filter passes through all frequencies unchanged, but changes the phase of the signal. This is a filter commonly used in phaser effects.
An important parameter is the required frequency response. In particular, the steepness and complexity of the response curve is a deciding factor for the filter order and feasibility.
A first order recursive filter will only have a single frequency-dependent component. This means that the slope of the frequency response is limited to 6 dB per octave. For many purposes, this is not sufficient. To achieve steeper slopes, higher order filters are required.
In relation to the desired frequency function, there may also be an accompanying weighting function which describes, for each frequency, how important it is that the resulting frequency function approximates the desired one. The larger weight, the more important is a close approximation.
[edit] The impulse response
There is a direct correspondence between the filter's frequency function and its impulse response, the former is the Fourier transform of the latter. This means that any requirement on the frequency function is a requirement on the impulse response, and vice versa.
However, in certain applications it may be the filter's impulse response which is explicit and the design process then aims at producing as close an approximation as possible to the requested impulse response given all other requirements.
In some cases it may even be relevant to consider a frequency function and impulse response of the filter which are chosen independently from each other. For example, we may both want a specific frequency function of the filter and that the resulting filter have a small effective width in the signal domain as possible. The latter condition can be realized by considering a very narrow function as the wanted impulse response of the filter even though this function has no relation to the desired frequency function. The goal of the design process is then to realize a filter which tries to meet both these contradicting design goals as much as possible.
[edit] Causality
In order to be implementable, any time-dependent filter must be causal: the filter response only depends on the current and past inputs. A standard approach is to leave this requirement until the final step. If the resulting filter is not causal, it can be made causal by introducing an appropriate time-shift (or delay). If the filter is a part of a larger system (which it normally is) these types of delays have to be introduced with care since they affect the operation of the entire system.
[edit] Stability
A stable filter assures that every limited input signal produces a limited filter response. A filter which does not meet this requirement may in some situations prove useless or even harmful. Certain design approaches can guarantee stability, for example by using only feed-forward circuits such as an FIR filter. On the other hand, filter based on feedback circuits have other advantages and may therefore be preferred, even if this class of filters include unstable filters. In this case, the filters must be carefully designed in order to avoid instability.
[edit] Locality
In certain applications we have to deal with signals which contain components which can be described as local phenomena, for example pulses or steps, which have certain time duration. A consequence of applying a filter to a signal is, in intuitive terms, that the duration of the local phenomena is extended by the width of the filter. This implies that it is sometimes important to keep the width of the filter's impulse response function as short as possible.
According to the uncertainty relation of the Fourier transform, the product of the width of the filter's impulse response function and the width of its frequency function must exceed a certain constant. This means that any requirement on the filter's locality also implies a bound on its frequency function's width. Consequently, it may not be possible to simultaneously meet requirements on the locality of the filter's impulse response function as well as on its frequency function. This is a typical example of contradicting requirements.
[edit] Computational complexity
A general desire in any design is that the number of operations (additions and multiplications) needed to compute the filter response is as low as possible. In certain applications, this desire is a strict requirement, for example due to limited computational resources, limited power resources, or limited time. The last limitation is typical in real-time applications.
There are several ways in which a filter can have different computational complexity. For example, the order of a filter is more or less proportional to the number of operations. This means that by choosing a low order filter, the computation time can be reduced.
For discrete filters the computational complexity is more or less proportional to the number of filter coefficients. If the filter has many coefficients, for example in the case of multidimensional signals such as tomography data, it may be relevant to reduce the number of coefficients by removing those which are sufficiently close to zero.
Another issue related to computational complexity is separability, that is, if and how a filter can be written the convolution of two or more simpler filters. In particular, this issue is of importance for multidimensional filters, e.g., 2D filter which are used in image processing. In this case, a significant reduction in computational complexity can be obtained if the filter can be separated as the convolution of one 1D filter in the horizontal direction and one 1D filter in the vertical direction. A result of the filter design process may, e.g., be to approximate some desired filter as a separable filter or as a sum of separable filters.
[edit] Other considerations
It must also be decided how the filter is going to be implemented:
[edit] Analog filters
The design of linear analog filters is for the most part covered in the linear filter section.
[edit] Digital filters
Digital filters are classified into one of two basic forms, according to how they respond to an unit impulse:
- Finite impulse response, or FIR, filters express each output sample as a weighted sum of the last N inputs, where N is the order of the filter. Since they do not use feedback, they are inherently stable. If the coefficients are symmetrical (the usual case), then such a filter is linear phase, so it delays signals of all frequencies equally. This is important in many applications. It is also straightforward to avoid overflow in an FIR filter. The main disadvantage is that they may require significantly more processing and memory resources than cleverly designed IIR variants. FIR filters are generally easier to design than IIR filters - the Remez exchange algorithm is one suitable method for designing quite good filters semi-automatically. (See Methodology.)
- Infinite impulse response, or IIR, filters are the digital counterpart to analog filters. Such a filter contains internal state, and the output and the next internal state are determined by a linear combination of the previous inputs and outputs (in other words, they use feedback, which FIR filters do not). In theory, the impulse response of such a filter never dies out completely, hence the name IIR, though in practice, this is not true given the finite resolution of computer arithmetic. IIR filters normally require less computing resources than an FIR filter of similar performance. However, due to the feedback, high order IIR filters may have problems with instability, arithmetic overflow, and limit cycles, and require careful design to avoid such pitfalls. Additionally, since the phase shift is inherently a non-linear function of frequency, the time delay through such a filter is frequency-dependent, which can be a problem in many situations. 2nd order IIR filters are often called 'biquads' and a common implementation of higher order filters is to cascade biquads. A useful reference for computing biquad coefficients is the RBJ Audio EQ Cookbook.
[edit] Sample rate
Unless the sample rate is fixed by some outside constraint, selecting a suitable sample rate is an important design decision. A high rate will require more in terms of computational resources, but less in terms of anti-aliasing filters. Interference and beating with other signals in the system may also be an issue.
[edit] Anti-aliasing
For any digital filter design, it is crucial to analyze and avoid aliasing effects. Often, this is done by adding analog anti-aliasing filters at the input and output, thus avoiding any frequency component above the Nyquist frequency. The complexity (i.e., steepness) of such filters depends on the required signal to noise ratio and the ratio between the sampling rate and the highest frequency of the signal.
[edit] Theoretical basis
Parts of the design problem relate to the fact that certain requirements are described in the frequency domain while others are expressed in the signal domain and that these may contradict. For example, it is not possible to obtain a filter which has both an arbitrary impulse response and arbitrary frequency function. Other effects which refer to relations between the signal and frequency domain are
- The uncertainty principle between the signal and frequency domains
- The variance extension theorem
- The asymptotic behaviour of one domain versus discontinuities in the other
[edit] The uncertainty principle
As stated in the uncertainty principle, the product of the width of the frequency function and the width of the impulse response cannot be smaller than a specific constant. This implies that if a specific frequency function is requested, corresponding to a specific frequency width, the minimum width of the filter in the signal domain is set. Vice versa, if the maximum width of the response is given, this determines the smallest possible width in the frequency. This is a typical example of contradicting requirements where the filter design process may try to find a useful compromise.
[edit] The variance extension theorem
Let be the variance of the input signal and let be the variance of the filter. The variance of the filter response, , is then given by
- = +
This means that σr > σf and implies that the localization of various features such as pulses or steps in the filter response is limited by the filter width in the signal domain. If a precise localization is requested, we need a filter of small width in the signal domain and, via the uncertainty principle, its width in the frequency domain cannot be arbitrary small.
[edit] Discontinuities versus asymptotic behaviour
Let f(t) be a function and let F(ω) be its Fourier transform. There is a theorem which states that if the first derivative of F which is discontinuous has order , then f has an asymptotic decay like t − n − 1.
A consequence of this theorem is that the frequency function of a filter should be as smooth as possible to allow its impulse response to have a fast decay, and thereby a short width.
[edit] Methodology
One common method for designing FIR filters is the Remez exchange algorithm. Here the user specifies a desired frequency response, a weighting function for errors from this response, and a filter order N. The algorithm then finds the set of N coefficients that minimize the maximum deviation from the ideal. Intuitively, this finds the filter that is as close as you can get to the desired response given that you can use only N coefficients. This method is particularly easy in practice since at least one text[1] includes a program that takes the desired filter and N and returns the optimum coefficients. One possible drawback to filters designed this way is that they contain many small ripples in the passband(s), since such a filter minimizes the peak error.
Another method to finding a discrete FIR filter is filter optimization described in Knutsson et al., which minimizes the integral of the square of the error, instead of its maximum value. In its basic form this approach requires that an ideal frequency function of the filter FI(ω) is specified together with a frequency weighting function W(ω) and set of coordinates xk in the signal domain where the filter coefficients are located.
An error function is defined as
where f(x) is the discrete filter and is the discrete-time Fourier transform defined on the specified set of coordinates. The norm used here is, formally, the usual norm on L2 spaces. This means that measures the deviation between the requested frequency function of the filter, FI, and the actual frequency function of the realized filter, . However, the deviation is also subject to the weighting function W before the error function is computed.
Once the error function is established, the optimal filter is given by the coefficients f(x) which minimize . This can be done by solving the corresponding least squares problem. In practice, the L2 norm has to be approximated by means of a suitable sum over discrete points in the frequency domain. In general, however, these points should be significantly more than the number of coefficients in the signal domain to obtain a useful approximation.
[edit] Simultaneous optimization in both domains
The previous method can be extended to include an additional error term related to a desired filter impulse response in the signal domain, with a corresponding weighting function. The ideal impulse response can be chosen independently of the ideal frequency function and is in practice used to limit the effective width and to remove ringing effects of the resulting filter in the signal domain. This is done by choosing a narrow ideal filter impulse response function, e.g., an impulse, and a weighting function which grows fast with the distance from the origin, e.g., the distance squared. The optimal filter can still be calculated by solving a simple least squares problem and the resulting filter is then a "compromise" which has a total optimal fit to the ideal functions in both domains. An important parameter is the relative strength of the two weighting functions which determines in which domain it is more important to have a good fit relative to the ideal function.
[edit] References
- ^ Rabiner, Lawrence R., and Gold, Bernard, 1975: Theory and Application of Digital Signal Processing (Englewood Cliffs, New Jersey: Prentice-Hall, Inc.) ISBN 0139141014
- A. Antoniou (1993). Digital Filters: Analysis, Design, and Applications. McGraw-Hill, New York, NY. ISBN 0070021171.
- S.W.A. Bergen and A. Antoniou (2005). "Design of Nonrecursive Digital Filters Using the Ultraspherical Window Function". EURASIP Journal on Applied Signal Processing 2005 (12): 1910. doi: .
- A.G. Deczky (October 1972). "Synthesis of Recursive Digital Filters Using the Minimum p-Error Criterion". IEEE Trans. Audio Electroacoust AU-20: pp. 257-263.
- J.K. Kaiser (1974). "Nonrecursive Digital Filter Design Using the Io-sinh Window Function". Proc. 1974 IEEE Int. Symp. Circuit Theory: pp 20-23.
- H. Knutsson, M. Andersson and J. Wiklund (June 1999). "Advanced Filter Design". Proc. Scandinavian Symposium on Image Analysis, Kangerlussuaq, Greenland.
- S.K. Mitra (1998). Digital Signal Processing: A Computer-Based Approach. McGraw-Hill, New York, NY. ISBN 0072865466.
- A.V. Oppenheim and R.W. Schafer and J.R. Buck (1999). Discrete-Time Signal Processing. Prentice-Hall, Upper Saddle River, NJ. ISBN 0137549202.
- T.W. Parks and J.H. McClellan (March 1972). "Chebyshev Approximation for Nonrecursive Digital Filters with Linear Phase". IEEE Trans. Circuit Theory CT-19: pp. 189-194.
- L.R. Rabiner, J.H. McClellan, and T.W. Parks (April 1975). "FIR Digital Filter Design Techniques Using Weighted Chebyshev Approximation". Proc. IEEE 63: pp. 595-610.
[edit] See also
[edit] External links
- A list of digital filter design software at dspGuru
- DISPRO – Free digital filter design software
- Free Analog acitve filter design software – from Texas Instruments
- Analog Filter Design Demystified
- Yehar's digital sound processing tutorial for the braindead! This paper explains simply (between others topics) filters design theory and give some examples]