Open Access

Optimal multiobjective design of digital filters using spiral optimization technique

  • Abderrahmane Ouadi1,
  • Hamid Bentarzi1 and
  • Abdelmadjid Recioui1Email author
SpringerPlus20132:461

DOI: 10.1186/2193-1801-2-461

Received: 16 June 2013

Accepted: 4 September 2013

Published: 13 September 2013

Abstract

The multiobjective design of digital filters using spiral optimization technique is considered in this paper. This new optimization tool is a metaheuristic technique inspired by the dynamics of spirals. It is characterized by its robustness, immunity to local optima trapping, relative fast convergence and ease of implementation. The objectives of filter design include matching some desired frequency response while having minimum linear phase; hence, reducing the time response. The results demonstrate that the proposed problem solving approach blended with the use of the spiral optimization technique produced filters which fulfill the desired characteristics and are of practical use.

Keywords

Multiobjective filter design Spiral optimization technique Magnitude response Minimum linear phase Group delay

Introduction

Digital filters exist in two types: Finite impulse response (FIR) and Infinite impulse response (IIR) or recursive. FIR filters suffer from the problem of high order (hence implementation and performance issues) if strict requirements are imposed at the design stage. Furthermore, IIR filters can have smaller group delay than its equivalent FIR filters (Antoniou 1993 Antoniou 2005). The optimal design of an infinite impulse response (IIR) filter consists in choosing a set of coefficients of the filter to have a frequency response that optimally approximates the desired response (Antoniou 2005; Dumitrescu and Niemisto 2004; Lai 2009; Tseng and Lee 2002 Tseng 2004; Ho et al. 2008; Quelhas and Petraglia 2009; Sanathanan and Koerner 1963).

Different techniques exist for the design of digital filters. Windowing method; in which the ideal impulse response is multiplied by a window function, is the most popular. There are various kinds of window functions (Butterworth, Chebyshev, Kaiser etc.), depending on the requirements on ripples in the passband and stopband, stopband attenuation and the transition width. These various windows limit the infinite length impulse response of ideal filter into a finite window to design an actual response. Furthermore, windowing methods do not allow sufficient control of the frequency response in the various frequency bands and other filter parameters such as transition width. The designer always has to compromise between the design specifications (Antoniou 1993 Antoniou 2005).

Due to the presence of the denominator of the transfer function, the stability condition of the filter should be taken into account in the optimal design (Antoniou 1993 Antoniou 2005; Lu et al. 1998; Saab et al. 1999; Tseng and Lee 2002; Omoifo and Hinamoto 2004; Tseng 2004; Jiang and Kwan 2009 Jiang and Kwan 2010a; Lai and Lin 2010), resulting in a constrained optimization problem. Several sufficient conditions (Lang 2000; Lu 2000; Antoniou 2005; Ho et al. 2008; Pan 2009; Jiang and Kwan 2010b) have been established for the parameterization that represents the filter’s denominator by a single polynomial. The triangle-based stability conditions (Antoniou 1993) are necessary and sufficient and have been incorporated into several design procedures (Lu 1998; Lu and Hinamoto 2003; Lu 2006) that formularize the filter’s denominator by cascaded second-order sections (SOSs). In (Lu 1998), variable transformation is used to convert the finite stability region into the entire coefficient space, such that the original constrained design problem becomes an unconstrained one in the transformed space. However, the transformation increases the nonlinearity of the objective function, which makes it hard to find good (global optimum) solutions in general. In (Lu and Hinamoto 2003), a perturbed stability triangle is proposed to guarantee the SOS to have its zeroes inside a circle of given radius. It is combined with the Gauss–Newton strategy, resulting in an improved design. In (Lang 1999), the conditions presented for the SOS with zeros inside a circle of given radius enclose a triangular stability domain and can be easily incorporated into any constrained optimization formulations based on the SOS parameterization. A method that divides the overall design of an IIR filter into successive designs of its second-order sections is presented in (Saab et al. 1999), where one section is first designed, and then, another section is appended until all sections are designed.

Because of finite word length effects occurring in practical implementations of the designed filters, not only stability of the filter is of great importance but a stability margin is necessary as well. The poles of the transfer function should not lie too close to the unit circle. The sensitivity of pole locations to coefficient quantization increases with decreasing distance from the unit circle. Poles close to the unit circle may considerably enhance quantization noise and increase the maximum amplitude of small scale limit cycle. Consequently, it is desirable to have control over the maximum radius when designing IIR filters.

Linear-phase filters are usually designed as non-recursive (FIR) filters which can have constant group delay over the entire base-band. However, when highly selective filters are required, a very high filter order is needed which makes these filters uneconomical or impractical. To eliminate this problem, attempts have been made to develop methods to design recursive (IIR) filters whose delay characteristics approximate a constant value in the passband. This includes IIR filter design approach that can satisfy both magnitude and phase characteristics simultaneously (Inukai 1980; Cortelazzo and Lightne 1984; Sullivan and Adams 1997; Lang 2000; Lertniphonphun and McClellan 2001). The design of IIR filters with constant group delay in the passband is also carried out by using allpass structures through evaluation of phase response instead of approximating the group delay directly (Jing 1987; Ikehara et al. 1992; Lang and Laakso 1994; Zhang and Iwakura 1999). Some other methods used an indirect approach based on model reduction techniques where a linear-phase FIR filter that meets the required specifications is first designed and then a lower order IIR filter that meets the original amplitude specifications while maintaining a linear-phase response in the passband is obtained (Sreeram and Agathoklis 1992; Peng et al. 1992; Beliczynski et al. 1992). Coretlazzo and Lightner (Coretlazzo and Lightner 1984) have achieved the simultaneous design in both magnitude and group delay of IIR and FIR filters based on multiple criterion optimizations. Lutova (Lutova 1997) has developed a new design method for elliptic IIR filters that provide the implementation of half of the multiplication constants with few shifters and adders. Sullivan et al. (Sullivan James and Adams 1998) have proposed the algorithm based on the peak–constrained least–squares optimality criterion for cascaded IIR filters, which can design a filter that has an equalized group delay without the use of all pass filters, and it can simultaneously meet the frequency response magnitude specifications by using all of the filter coefficients available to optimize the filter. Lang (Lang 2000) has used least square method for designing IIR filter with prescribed magnitude and phase response. This parameterization of the transfer function has been used for designing IIR filters. GordanaJavanovic (Gordana 2006) has proposed a method for the design of IIR notch filters with desired magnitude characteristic, which can be either maximally flat or equi-ripple. Xi Zhang (Zhang 2008) have proposed a novel method for designing maximally flat IIR filters with flat group delay responses in the pass-band.

Under these circumstances, evolutionary and metaheuristic optimization methods find their place. These are referred to as global optimizers while the more familiar, traditional techniques such as conjugate gradient and the quasi-Newtonian methods are classified as local optimizers. The distinction between local and global search of optimization techniques is that the local techniques produce results that are highly dependent on the starting point or initial guess, while the global methods are totally independent of the initial conditions (Recioui 2012). Though they possess the characteristic of being fast in convergence, local techniques, in particular the quasi-Newtonian techniques have direct dependence on the existence of at least the first derivative. In addition, they place constraints on the solution space such as differentiability and continuity; conditions that are hard or even impossible to satisfy in some situations (Recioui 2012).

Previously, global optimization techniques have been implemented in the design of digital filters. One such approach using neural networks has been described in (Wang et al. 2006). Also, use of PSO in the design of frequency sampling finite impulse response (FIR) filter has been described in (Wang et al. 2004; Krusienski and Jenkins 2004; Chen and Chen 2006). Differential evolution has been used in the design of digital filters has been implemented in (Karaboga D 2004; Storn 1996 2005; Karaboga N 2005). (Kit Sang and Kim-Fung 1998; Karaboga N 2004) have used Hierarchical Genetic Algorithms to the design and optimization of IIR filter structures. Use of Particle Swarm Optimization (PSO) and Genetic Algorithms (GA) in the design of digital filters is described in (Ababneh, and Bataineh 2007).

In this work, the application of the novel optimization technique called spiral optimization to the design of digital filters is considered. The purpose is to design a filter that can simultaneously satisfy multiobjective criteria including frequency response and linear phase with the least possible group delay.

Problem formulation

Digital filters find their applications in different areas. One area is power system protection where measurement systems involve faulted signals associated with DC decaying signals, harmonic and sub-harmonic components. To eliminate these unwanted components, a digital filter design based on multi-objective optimization technique to satisfy different specifications such as high speed response for a real-time application and frequency domain requirements.

Digital filtering approach

A digital filter based solution is proposed to remove unwanted disturbances using digital filter design techniques. The filter time response must be included in the requirements. The present filtering application imposes different kind of specifications. On one hand, the time domain requirement where both a high speed and accurate system response are needed. On the other hand, the frequency domain requirements (DC, sub-synchronous and harmonic components elimination) which are the magnitude response within small bandwidth including sharp frequency edges as well as an approximately constant group delay in this band are required too. Usually the best optimum value of all the objective functions of this filter design can be obtained for some values of design variables. A compromise or a trade-off between the objective functions must be made to achieve a satisfactory filter design.

The considered recursive digital filter must satisfy three multi-objective functions. These functions are: 1) meet a specified or a desired magnitude response specification; 2) an approximately constant group delay; and 3) a minimum time response or settling time which involves a minimum phase or a group delay. The optimization approach considers the discrete-time transfer function which is formulated on the basis of some desired amplitude response and a stability margin parameter. A norm of the weighted error function is then minimized with respect to the transfer-function coefficients with a prescribed maximum pole radius referred to as stability margin. The stability margin parameter is varied to optimize the filter coefficients which minimizes mainly the magnitude response, satisfies the best approximately constant group delay and the lowest group delay that leads to minimum settling time or time delay of the system dynamic response.

Filter transfer functions

In the general case an IIR filter can be described by its discrete-time difference equation
y n + i = 1 2 N 1 c i y n 1 = j = 0 2 M d j x n j
(1)
Where x[n] and y[n] are discrete-time input and output signals. Equation (1) can be transformed into the Z-domain and assuming ci and di are real coefficients a second order form transfer function can be obtained, having 2 M conjugate zeros and 2 N conjugate poles; called second order sections (SOS), as:
H z = H 0 j = 1 2 M a 0 j + a 1 j z + z 2 j = 1 2 N b 0 j + b 1 j z + z 2
(2)
Where aii and bjj are real coefficients and H0 is a positive multiplier constant. The polar formulation is also useful and is written as:
H z = H 0 j = 1 2 M z r aj e j θ aj z r aj e j θ aj j = 1 2 N z r bj e j θ bj z r bj e j θ bj
(3)

Where raj, θaj and rbjbj are the radii and angles of the zeros and poles, respectively.

Filter stability margin

As the poles are moved toward the origin (decreasing the pole radius), the system stability margin parameter increases and the system settling time decreases. This action, in fact, brings two required and important properties to the designed system. First, the system time or dynamic response is enhanced as settling time is decreased. Second, the system stability becomes more robust which is a very useful property, particularly in practical implementation. Indeed, the rounding or truncation of the filter coefficients may lead to an unstable implementation if the stability margin is too small. It is therefore desirable to approximate a given response by a transfer function with a prescribed maximum pole radius named stability margin as shown in Figure 1.
Figure 1

Stability region and stability margin parameter δ.

Magnitude response objective function

The amplitude and the phase responses of a recursive filter is given by
M x , ω = H e jωt , φ x , ω = arg H e jωt
(4)
Where ω is the frequency and x is a column vector with 2 M + 2 N + 1 elements and depend on the used formulation, that is a Cartesian form as
x = a 01 a 11 a 21 b 01 b 11 b 21 b 2 M H 0
(5)
Or in polar form as
x = ρ a ρ b H 0
(6)
Where
ρ a = r a 1 , θ a 1 , , r aM , θ aM
(7)
And
ρ b = r b 1 , θ b 1 , , r bN , θ bM
(8)
The superscript T denotes the transpose operation. An approximation error can be formulated as the difference between the actual amplitude response M(x, ω) and the desired amplitude or magnitude response Μ d (ω) as:
e x = M x , ω Μ d ω
(9)
By sampling the error function e(x,ωi), the actual and the desired amplitude responses M(x,ωi) and Md(ω) at frequencies: ω1, ω2, …, ωk, the column error vector is constructed as:
E x = e 1 x , e 2 x , , e k x T
(10)
Where e i (x) = M(x, ω i ) − M d (ω i ) and {ω i  : i = 1, 2, k} is a dense set of frequencies which are distributed over in the pass-band and stop-band of the filter. A weighting or penalty error is included to control portions of the actual filter response curve that are most important to the filter response. This involves modifying the error to the form:
e ¯ i x = W ω M x , ω i M d ω i
(11)

Where W(ω) is a weighting piece wise constant function over all frequency space, which is assigned a positive value greater than one or less than one to increase or decrease the magnitude approximation in a given band.

A recursive filter can be designed by finding a point x = x ^ in (11) such that
e ¯ i x 0 for i = 1 , 2 , k
(12)
Such a point can be obtained when solving the optimization problem by minimizing the error function e ¯ i x . The design of a recursive filter that approaches a specified response M d (ω i ), can be performed by minimizing the error objective function in terms of Lp norm error which is formulated as:
min x μ x
(13)

Subject to r aj  ≤  1 − δ

Where
μ x = e max x i = 1 l W i ω e i x e max x p 1 p
(14)
And
e max x = max 1 i l e i x

The p is a positive integer. The spiral technique is used to minimize μ(x) for increasing values of p.

Group-delay objective function

The group delay is derived from the phase relation, as given in equation (4), and is defined as:
τ x , ω = x , ω
(15)
Where φ(x, ω) is the phase response of the filter, and stated as
φ x , ω = arg H e jωT
(16)
The group delay for a synthesized recursive filter is desired to be unchanged in the considered region on one hand. On the other hand, a minimum group delay is another required property needed in certain applications. In the present design, a constant group delay property is considered which is defined as:
τ x , ω i = τ c for ω i ω p
(17)
Where τ c is constant, and ω p is the passband regions of the filter. In the present application τ c is an unknown but can be considered as the mean value over the passband region, which can be determined as:
τ c , m = 1 n k i = 1 n k τ x ¯ m , ω i for ω i ω p
(18)

Where x ¯ m is the optimal filter coefficient determined by minimizing the magnitude objective function for an m-th stability margin parameter.

The stability margin parameter is varied for discrete values, from which an optimal constant group delay is determined by minimizing the following objective error function:
min m E τ m = 1 τ c , m i = 1 n k τ x ¯ m , ω i τ c , m for ω i ω p
(19)

The multi-objective optimization problem is solved by discretizing the stability margin parameter, the magnitude optimization algorithm is used to generate the corresponding filter coefficients’, in which basis the group delay is synthesized where a feasible and optimal solution can be obtained by minimizing the objective function (19).

The final multiobjective design is obtained by minimizing the sum combination of the errors in equations (14) and (19).

The spiral inspired optimization method

Compared with traditional optimization techniques and other global optimizers, the spiral optimization method is easy to implement and very efficient in reaching optimum solutions. Spiral optimization method has been recently developed based on the analogy to spiral phenomena (Tamura and Yasuda 2011a; Tamura and Yasuda 2011b).

Patterns in nature are visible regularities of form found in the natural world. These patterns recur in different contexts and can sometimes be modelled mathematically. Natural patterns include symmetries, trees, spirals, meanders, waves, foams, arrays, cracks and stripes. Mathematics, physics and chemistry can explain patterns in nature at different levels. Patterns in living things are explained by the biological processes of natural selection and sexual selection. Studies of pattern formation make use of computer models to simulate a wide range of patterns.

Among the natural patterns, spirals are common in plants and in some animals. For example, in the nautilus (Figure 2a), each chamber of its shell is an approximate copy of the next one, scaled by a constant factor and arranged in a logarithmic spiral.
Figure 2

Natural spiral patterns and their mathematical model. (a) Cutaway of a nautilus shell. (b) Logarithmic spiral.

The spiral phenomena occurring in nature (like the one in Figure 2a) are approximated to logarithmic spirals as in Figure 2b. Examples of natural spiral dynamics include whirling currents, low pressure fonts, nautilus shells and arms of spiral galaxies. Logarithmic spirals discrete processes to generate spirals that can form an effective behaviour in metaheuristics. A two-dimensional algorithm has been first proposed (Tamura and Yasuda 2011a), and then, a more generalized n-dimensional version has been recently suggested (Tamura and Yasuda 2011b).

In the present work, the use of the spiral optimization technique is presented and used to solve the multiobjective IIR filter design. First, based on practical requirements involved in power system application, a detailed mathematical IIR filter design formulation is presented.

Next, the spiral optimization algorithm detailed and implemented to solve the optimization task. A first optimization example aims at matching a desired magnitude response only. Later, the design is improved by considering the other requirements including minimum and constant group delay.

Before presenting the n-dimensional spiral optimization algorithm, it is worth understanding the two dimensional optimization model as some results are just extended over.

Two-dimensional spiral optimization

Rotating a point in a 2-dimensional orthogonal coordinate system (as shown in Figure 3) to the left around the origin by θ can be expressed as:
x ' = R θ x
(20)
Figure 3

Rotation in × 1 2 plane.

Where
R 2 θ = cos θ sin θ sin θ cos θ
(21)
Hence, the two dimensional algorithm moves from one point to another as:
x 1 k + 1 x 2 k + 1 = r R 2 θ x 1 k x 2 k
(22)

Where θ is the rotation angle around the origin (0 ≤ θ ≤ 2π) and r is the convergence rate of distance between a point and the origin at each iteration k (0 < r < 1). The parameter r can be regarded as the “scaling” or radius of the logarithmic spiral curvature.

The spiral model presented earlier has a center only at the origin. Hence, it should be extended to have center at an arbitrary point x* as:
x k + 1 = r R 2 θ x k r R 2 θ I 2 x *
(23)
Based on the previous formulation, the following optimization algorithm may be developed:
  • Preparation: select the number of search points m > 2, the parameters θ and r and the maximum number of iterations k max .

  • Initialization: initialize randomly the points; x i (0) i = 1..m; in the feasible region and the center x* as the point with the least fitness value.

  • Updating x i :
    x i k + 1 = r R 2 θ x i k r R 2 θ I 2 x * for i = 1 m .
    (24)
  • Updating x*: Select x* as the point with the least fitness function in the updated set of points.

  • Check for termination criterion: If k = k max then stop. Otherwise, start a new iteration.

n-dimensional spiral optimization

The extension of the two-dimensional optimization algorithm presented earlier is easy to do as one must understand how rotation in an n-dimensional space is done. Rotation in n-dimension is performed in the same way as the two-dimensional rotation taking two dimensions at a time. This is defined for dimensions i, j as:
R i , j = 1 i j 1 i cos θ sin θ 1 1 j sin θ cos θ 1 1
(25)

Where the blank elements are zeros.

Hence, there are n n 1 2 rotation matrices. The resulting rotation matrix is then (Tamura and Yasuda 2011b):
R n θ = i = 1 n 1 ( j = 1 i R n i , n + 1 j θ )
(26)
Hence the n-dimensional algorithm may be formulated similar to the two-dimensional algorithm as:
  • Preparation: select the number of search points m > 2, the parameters θ and r and the maximum number of iterations k max .

  • Initialization: initialize randomly the points; x i (0) i = 1..m; in the feasible region and the center x* as the point with the least fitness value.

  • Updating x i :
    x i k + 1 = r R n θ x i k r R n θ I n x * for i = 1 m
    (27)
  • Updating: x*: Select x* as the point with the least fitness function in the updated set of points.

  • Check for termination criterion: If k = k max then stop. Otherwise, start a new iteration.

The flowchart in Figure 4 summarizes the spiral optimization procedure.
Figure 4

The spiral optimization algorithm flowchart.

Results and discussions

The digital filter to be optimized is to be used to eliminate harmonics and sub-harmonics in a power network with a fundamental frequency of 50 Hz. The filter is a bandpass type and is desired to satisfy the magnitude response to ideally pass only frequencies confined in the interval [45 Hz, 55 Hz] while rejecting all other frequency content. The sampling frequency is taken to be 1800 Hz. The filter is of order 10 and is hence composed of 5 cascaded SOSs. At start up, the filter is optimized to match the magnitude response specifications only. Next, more constraints are added to the optimization process including minimum and linear phase and constant group delay to enhance the designed filter performance.

Single objective design

The purpose of this part is to design a filter which only satisfies the magnitude response described earlier without considering any other performance criteria. The filter is designated single objective optimized filter (SOOF) hereafter. Figure 5 shows both the desired and the optimized magnitude response of the digital filter. It can be noted that the filter fulfils well the requirements of magnitude response. Indeed, the filter response falls exactly within the desired frequency range [45 Hz, 55 Hz] and it attenuates all other frequencies as the overall Sidelobe level is lower than −16 dB.
Figure 5

Magnitude filter response (SOOF).

However, the optimized filter is not of practical use as it suffers from drawbacks in the dynamic properties. First, as shown in Figure 6, the time delay is in the order of 100 ms which is not suitable in the present applications as the requirements specify that the time delay should not exceed one cycle (20 ms). In addition, the group delay is not constant as shown in Figure 7. Furthermore, the phase response is nonlinear as shown in Figure 8 which explains the non constant nature of the group delay. As a result, it is necessary to include all the preceding performance criteria into the design process and hence, the problem becomes a multiobjective optimization task. The results are summarized in Table 1 where are presented the filter SOS coefficients and SOS gains.
Figure 6

Step response of the filter and time delay (about 100 ms) (SOOF).

Figure 7

Goup delay response of single objective optimized filter (SOOF).

Figure 8

Phase response of filter with non linear phase or non constant group delay (SOOF).

Table 1

Single-objective optimized filter (SOOF) SOS coefficients and gains

Section

Numerator

Denominator

Gain

1

1

−1,9520335260

0,9999998000

1

−1,8473812327

0,8836

0,00089772300915

2

1

−1,9999996000

0,99999980000

1

−1,9999996000

0,999999800000010

1

3

1

−1,9999996000

0,99999980000

1

−1,8473812327

0,88360000000

1

4

1

−1,8851094580

0,99999980000

1

−1,8473812327

0,883600000000

1

5

1

1,99999980000

0,99999980000

1

−1,8473812327

0,88360000000

1

Multiobjective filter design

The inclusion of the constant and minimum group delay in the optimization task besides magnitude response criterion produced a filter which satisfies almost all requirements. The filter is thereafter called multiobjective optimized filter (MOOF). The magnitude response of the MOOF is shown in Figure 9. The filter magnitude response is not as good as the magnitude response characteristic of the SOOF. Indeed, the SOOF bandwidth is narrower than the MOOF which makes it having better selectivity.
Figure 9

Magnitude filter response (MOOF).

However, the MOOF performs well in terms of the other criteria. In terms of phase response and group delay, Figures 10 and 11 illustrate the phase and group delay responses of the MOOF. The filter is characterized by a linear phase with an almost constant group delay in the pass-band contrary to the SOOF phase response and group delay illustrated in Figures 7 and 8. Furthermore, the group delay is low yielding a small time response as shown in Figure 12. Indeed, the time delay of the MOOF is about 14.9 ms which conforms to the desired specifications. This is due to the fact that the stability margin or equivalently the pole radii have been taken into account in the optimization process and these latter have been lowered to a value of 0.595 instead of being closer to unity. Table 2 summarizes the SOS coefficients and gains of the MOOF.
Figure 10

Optimized filter phase response MOOF.

Figure 11

Group delay response of optimized filter MOOF.

Figure 12

Optimized filter step response and time delay (14.9 ms) MOOF.

Table 2

Multi-objective optimized filter (MOOF) SOS coefficients and gains

Section

Numerator

Denominator

Gain

1

1

−1,82933833319

0,999999800000

1

−1,0914383222

0,35402500000

0,012499922808

2

1

−1,4247565639

0,99999980000

1

−1,0914383222

0,35402500000

1

3

1

−1,1469437066

0,99999980000

1

−1,1899988100

0,35402500000

1

4

1

0,227981442276607

0,999999800000010

1

−1,09143832226114

0,354025000000000

1

5

1

−0,467539619942456

−0,532460226811516

1

−1,11373644574629

0,354025000000000

1

To validate the obtained results, the MOOF optimized filter is compared with the classical filter designs widely used in literature. The filters have the same order and should meet the same requirements described earlier. The filters considered are the least pth optimized filter, the elliptic and Chebyshev type II filters.

The classical filters have magnitude responses that meet well the requirements as compared to the MOOF. This can be explained by the fact that these filters are mainly designed to match a given magnitude response which can be considered to be similar to the SOOF optimized filter. However, according to Figures 11 and 13 the MOOF has lower and constant group delay over the pass-band region compared to the classical filters. In fact, the delays of the classical filters are in orders of 200 and 300 sample delay (the MOOF delay is at 14 samples) which is not acceptable for practical requirements. Also, the classical design approach relies on matching a required magnitude response followed by an all-pass filter equalizer. Hence, the classical design adds on the system order and the computational complexity. The latter approach is not suitable for real time applications.
Figure 13

Group delay comparison between the MOOF and the classical filter designs.

To better assess the performance of the filter (MOOF), two tests have been performed. In the first, a step sinusoidal signal of 50 Hz is input to the filter and Figure 14 shows both filter input and output waveforms. It is seen that the filter output matches exactly the input except for a phase shift and a time delay of less than one cycle. Hence, this filter proves to be practical for high speed measurement systems where the system accuracy is of great importance. In the second test, the previous step sinusoidal signal is corrupted with a DC offset, harmonic and subharmonic components. The subharmonic component is set to 25 Hz and the harmonics to 100, 150 and 200 Hz. Both input and output signals are shown in Figure 15. As it is clearly seen, the filter succeeded in eliminating the DC and harmonic components and mitigating the subharmonic component.
Figure 14

Test signal: step of sinusoidal signal input and filter output response.

Figure 15

Second test signal by considering a step sinusoidal signal mixed with harmonic and subharmonic components.

Conclusion

The application of the spiral optimization method to design a multiobjective digital filter has been considered in this paper. The objectives of the filter design were to match a desired magnitude response while having a minimum and linear phase. At start up, only magnitude response has been considered in the optimization task. The resulting filter was good in terms of this characteristic while it showed awful dynamic and phase performance. Next, the dynamic properties were included in the optimization algorithm to solve a multiobjective task. The spiral optimization method has succeeded in attaining the optimal design in terms of the previous requirements by achieving a compromise between them. The optimized filter has been tested and it showed good performance with required practical characteristics.

Declarations

Authors’ Affiliations

(1)
Laboratory signals and systems, Institute of Electrical and Electronic Engineering, University M’hamed Bougara Boumerdes

References

  1. Ababneh JI, Bataineh MH: Linear phase FIR filter design using particle swarm optimization and genetic algorithms. Digit signal process 2007. doi:10.1016/j.dsp.2007.05.011Google Scholar
  2. Antoniou A: Digital filters: analysis and design. New York: McGraw Hill; 1993.Google Scholar
  3. Antoniou A: Digital signal processing: signals, systems and filters. New York: McGraw Hill-Professional; 2005.Google Scholar
  4. Beliczynski B, Kale I, Cain GD: Approximation of FIR and IIR digital filters: an algorithm based on balanced model reduction. IEEE Trans on Signal Process 1992, 40(3):532-542. 10.1109/78.120796View ArticleGoogle Scholar
  5. Chen HC, Chen OT: Particle swarm optimization incorporating a preferential velocity-updating mechanism and its application in IIR filter design. IEEE Trans Syst Man Cybern 2006, 1190-1195. SMC’06 October 08-11, 2006, Taipei, TaiwanGoogle Scholar
  6. Cortelazzo G, Lightner MR: Simultaneous design in both magnitude and group-delay of IIR and FIR filters based on multiple criterion optimization. IEEE Trans Circuits Syst 1984, 32(5):949-967.Google Scholar
  7. Dumitrescu B, Niemisto R: Multistage IIR filter design using convex stability domains defined by positive realness. IEEE Trans Signal Process 2004, 52(4):962-974. 10.1109/TSP.2004.823497View ArticleGoogle Scholar
  8. Gordana J-D: Design of IIR notch filters with maximally flat or equiripple magnitude characteristics. 14th edition. Florence, Italy: Europeon Signal Processing Conference (EUSIPCO 2006); 2006.Google Scholar
  9. Ho CYF, Ling BWK, Chi ZW, Shikh-Bahaei M, Liu YQ, Teo KL: Design of near-allpass strictly stable minimal-phase real valued rational IIR filters. IEEE Trans Circuits Syst II Exp Briefs 2008, 55(8):781-785.View ArticleGoogle Scholar
  10. Ikehara M, Funaishi M, Kuroda H: Design of complex all-pass networks using Remez algorithm. IEEE Trans Circuits Syst II 1992, 39(8):549-556. 10.1109/82.168947View ArticleGoogle Scholar
  11. Inukai T: A unified approach to optimal recursive digital filter design. IEEE Trans Circuits Syst 1980, CAS-27(7):646-649.View ArticleGoogle Scholar
  12. Jiang A, Kwan HK: IIR digital filter design with new stability constraint based on argument principle. IEEE Trans Circuits Syst I Reg Papers 2009, 56(3):583-593.View ArticleGoogle Scholar
  13. Jiang A, Kwan HK: Minimax design of IIR digital filters using iterative SOCP. IEEE Trans Circuits Syst I Reg Papers 2010, 57(6):1326-1337.View ArticleGoogle Scholar
  14. Jiang A, Kwan HK: Minimax design of IIR digital filters using SDP relaxation techniques. IEEE Trans Circuits Syst I Reg Papers 2010, 57(2):378-390.View ArticleGoogle Scholar
  15. Jing Z: A New method for digital allpass filter design. IEEE Trans ASSP 1987, 35(11):1557-1564. 10.1109/TASSP.1987.1165067View ArticleGoogle Scholar
  16. Karaboga D: Simple and global algorithm for engineering problems, DE. Turk J Elect Engin 2004, 12(1):53-60.Google Scholar
  17. Karaboga N Proceedings of the 6th Nordic signal processing symposium-NORSIG2004, June 09-11, 2004. In Design of minimum phase digital IIR filters by using genetic algorithm. Finland: Espo; 2004.Google Scholar
  18. Karaboga N: Digital filter design using differential evolution algorithm. EURASIP J Appl Signal Process 2005, 8: 1269-1276.View ArticleGoogle Scholar
  19. Kit Sang T, Kim-Fung M: Design and optimization of IIR filter structure using hierarchical genetic algorithms. IEEE Trans Ind Electron 1998, 45(3):481-487. 10.1109/41.679006View ArticleGoogle Scholar
  20. Krusienski DJ, Jenkins WK: Particle swarm optimization for adaptive IIR filter structures. Congr Evol Comput 2004, 1: 965-970.Google Scholar
  21. Lai XP: Optimal design of nonlinear-phase FIR filters with prescribed phase error. IEEE Trans Signal Process 2009, 57(9):3399-3410.View ArticleGoogle Scholar
  22. Lai XP, Lin ZP: Minimax design of IIR digital filters using a sequential constrained least-squares method. IEEE Trans Signal Process 2010, 58(7):3901-3906.View ArticleGoogle Scholar
  23. Lang M, Laakso TI: Simple and robust method for the design of allpass filters using least-squares phase error criterion. IEEE Tarns Circuits Syst II 1994, 41(1):40-48.View ArticleGoogle Scholar
  24. Lang MC Ph.D. Dissertation. In Constrained design of digital filters with arbitrary magnitude and phase responses. Vienna, Austria: Vienna Univ. Technol; 1999.Google Scholar
  25. Lang MC: Least-squares design IIR filters with prescribed magnitude and phase response and a pole radius constraint. IEEE Trans Signal Process 2000, 48(11):3109-3121. 10.1109/78.875468View ArticleGoogle Scholar
  26. Lertniphonphun W, McClellan JH Proceedings of the IEEE international conference on accoustics, speech and signal processing 6. In Unified design algorithm for complex FIR and IIR filters. Salt Lake City, Utah; 2001:3801-3804. May 2001Google Scholar
  27. Lu WS: Design of recursive digital filters with prescribed stability margin: a parameterization approach. IEEE Trans Circuits Syst II Analog Digit Signal Process 1998, 45(9):1289-1298. 10.1109/82.718597View ArticleGoogle Scholar
  28. Lu WS: Design of stable minimax IIR digital filters using semidefinite programming. Proc IEEE Int Symp Circuits Syst 2000, 1: 355-358. Geneva, SwitzerlandGoogle Scholar
  29. Lu WS: An argument-principle based stability criterion and application to the design of IIR digital filters. Proc IEEE Int Symp Circuits Syst 2006, 4431-4434. ISCAS2006, Island of Kos: Greece, May 2006Google Scholar
  30. Lu WS, Hinamoto T: Optimal design of IIR digital filters with robust stability using conic quadratic-programming updates. IEEE Trans Signal Process 2003, 51(6):1581-1592. 10.1109/TSP.2003.811229View ArticleGoogle Scholar
  31. Lu WS, Pei SC, Tseng CC: A weighted least-squares method for the design of stable 1-D and 2-D IIR digital filters. IEEE Trans Signal Process 1998, 46(1):1-10. 10.1109/78.651159View ArticleGoogle Scholar
  32. Lutova A: Design of computationally efficient elliptic IIR filters with a reduced number of shift-and-add operations in multipliers. IEEE Trans Signal Process 1997., 45(10):Google Scholar
  33. Omoifo OI, Hinamoto T: Optimal design of stable recursive digital filters using unconstrained optimization methods. Proc Int Midwest Symp Circuits Syst 2004, 2: II-49-II-52.Google Scholar
  34. Pan ST: Design of robust D-stable IIR filters using genetic algorithms with embedded stability criterion. IEEE Trans Signal Process 2009, 57(8):3008-3016.View ArticleGoogle Scholar
  35. Peng SC, Chen BS, Chiou BW: Simultaneous design in both magnitude and group-delay of IIR and FIR filters based on multiple criterion optimization. IEE Proc G 1992, 139(5):586-590.Google Scholar
  36. Quelhas MF, Petraglia A: Digital filter design optimization using partial cost functions. Proc IEEE Int Symp Circuits Syst 2009, 285-288. Taipei, TaiwanGoogle Scholar
  37. Recioui A: Sidelobe level reduction in linear array pattern synthesis using particle swarm optimization. J Optimiz Theory Appl 2012, 153(2):497-512. doi:10.1007/s10957-011-9953-9 10.1007/s10957-011-9953-9View ArticleGoogle Scholar
  38. Saab S, Lu WS, Antoniou A: Design and implementation of lowpower IIR digital filter systems. Proc IEEE Int Symp Circuits Syst 1999, 391-394. Orlando, FL3Google Scholar
  39. Sanathanan CK, Koerner J: Transfer function synthesis as a ratio of two complex polynomials. IEEE Trans Autom Control 1963, AC-8(1):56-58.View ArticleGoogle Scholar
  40. Sreeram V, Agathoklis P: Design of linear-phase IIR filters via impulse- response gramians. IEEE Trans. Signal Process 1992, 40(2):389-394. 10.1109/78.124948View ArticleGoogle Scholar
  41. Storn R: Differential evolution design of an IIR filter. IEEE Int Conf Evol Comput 1996, 268-273. 20-22 May, 1996, Nagoya, JapanGoogle Scholar
  42. Storn R: Designing nonstandard filters with differential evolution. Signal processing magazine. IEEE 2005, 22: 103-106.Google Scholar
  43. Sullivan James L, Adams JW: PCLS IIR digital filters with simultaneous frequency response magnitude and group delay specifications. IEEE Trans Signal Process 1998, 46(11):306-321.Google Scholar
  44. Sullivan JL, Adams JW Conference record of the thirty-first asilomar; conference on signals, systems and Computers1. In PCLS IIR filters with simultaneous frequency response magnitude and phase related specifications. California: Pacific Grove; 1997:705-709.Google Scholar
  45. Tamura K, Yasuda K: Primary study of spiral dynamics inspired optimization. IEEJ Trans electrical Electron Eng 2011, 6(S1):S98-S100. 10.1002/tee.20628View ArticleGoogle Scholar
  46. Tamura K, Yasuda K: Spiral dynamics inspired optimization. J Adv Comput Intell Intell Inform 2011, 15(8):1116-1122.Google Scholar
  47. Tseng CC: Design of stable IIR digital filter based on least P-power error criterion. IEEE Trans Circuits Syst I Reg Papers 2004, 51(9):1879-1888. 10.1109/TCSI.2004.834526View ArticleGoogle Scholar
  48. Tseng CC, Lee SL: Minimax design of stable IIR digital filter with prescribed magnitude and phase responses. IEEE Trans Circuits Syst I Fundam Theory Appl 2002, 49(4):547-551. 10.1109/81.995676View ArticleGoogle Scholar
  49. Wang WP, Zhou LF, Qian JX: Fir filter design: frequency sampling filters by particle swarm optimization algorithm. Proc Int Conf Mach Learn Cybern 2004, 4: 2332-2327.View ArticleGoogle Scholar
  50. Wang X, Meng X, He Y: A novel neural networks-based approach for designing FIR filters, The sixth world congress on intelligent control and automation. 2006, 1: 4029-4032.Google Scholar
  51. Zhang X: Design of maximally flat IIR filters with flat group delay responses. Signal Process 2008, 88: 1792-1800. 10.1016/j.sigpro.2008.01.016View ArticleGoogle Scholar
  52. Zhang X, Iwakura H: Design of IIR digital allpass filters based on eigenvalue problem. IEEE Trans Acoustics Speech Signal Process 1999, 47(2):554-559. 10.1109/78.740143View ArticleGoogle Scholar

Copyright

© Ouadi et al.; licensee Springer. 2013

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.