Identification for Wiener Systems with RTF Subsystems*

Identification for Wiener Systems with RTF Subsystems*

European Journal of Control (2006)6:581–594 # 2006 EUCA Identification for Wiener Systems with RTF Subsystems Xiao-Li Hu and Han-Fu Chen** Institute...

242KB Sizes 0 Downloads 68 Views

European Journal of Control (2006)6:581–594 # 2006 EUCA

Identification for Wiener Systems with RTF Subsystems Xiao-Li Hu and Han-Fu Chen** Institute of Systems Science, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing 100080, P. R. China

Recursive algorithms are proposed for identification of the Wiener system composed of a subsystem with rational transfer function (RTF) cascaded with a static nonlinearity. The nonlinear block is not necessary to be invertible. The strong consistency is proved for estimates for the coefficients of RTF and for the value of the nonlinear function at any point. Some numerical examples are given showing that the simulation results are consistent with theoretical analysis. Keywords: Recursive identification; Wiener systems; rational transfer function; strong consistency; stochastic approximation

1.

Introduction

The Wiener system consisting of a linear subsystem followed by a static nonlinearity is an important model in practice. For recent years, its identification issue has been attracting a great attention from researchers, e.g., [2,4,7,8,10–20] among others. The SISO Wiener system considered in the paper is presented by Fig. 1. To present the nonlinearity there are the parametric [1,5,9,12,14–20] and the nonparametric approaches [2,7,8,10,11,13]. The present paper belongs to the latter and is the continuation of [10,11]. If vk were observed, then identification of the linear subsystem would be an easy task. Therefore, it is important to extract information concerning vk from observations. In the case where fðÞ is invertible, vk can Supported by the NSFC under Grants No. G0221301, 60334040, 60474004. Correspondence to: H.-F. Chen, e-mail: [email protected]

be expressed by system output: vk ¼ f1 ðyk Þ, and such an expression may simplify analysis in many cases, in particular, for the case where the linear subsystem is also nonparametric [7,8,13]. However, the invertibility condition on fðÞ excludes many important nonlinearities including saturations from consideration. This motivated us to consider identification of Wiener systems without requiring invertibility of fðÞ in [10,11], where the recursive identification algorithms and their strong consistency were established for Wiener systems with general nonlinearities. However, the linear subsystem considered there is limited to an finite impulse response (FIR) system. In this paper the FIR system is extended to the RTF system, which clearly is with infinite impulse response. Similar to [10,11], the parameters of the linear part are first estimated, then the intermediate signals vk are approximated by using the estimated linear subsystem. Finally, the nonlinear block is identified on the basis of the approximated intermediate signals and the observations fzk g. Such an extension is technically nontrivial by the following observation. In the convergence analysis itP is of importance to have convergence of series like 1 k¼1 ak vk where fak g are the stepsizes used in the algorithms to be defined later on. P For an FIR linear subsystem the series 1 k¼1 ak vk can be split into a finite number of convergent subseries (due to independence), while for an RTF subsystem such a direct separation is impossible and the treatment is more delicate. The estimates for coefficients of the linear subsystem and for the value fðvÞ of the static nonlinearity Received 10 July 2006; Accepted 15 January 2007 Recommended by A. Karimi and L. Ljung

582

X.-L. Hu and H.-F. Chen

Fig. 1. Wiener system.

at any v are given by the stochastic approximation (SA) algorithms with expanding truncations [3]. These algorithms together with conditions for their convergence are presented in Section 2. The limiting behaviors of some correlation functions and series are demonstrated in Section 3. In order to deal with the difficulty arising from the infinite impulse response of the linear subsystem mentioned above, the totality of integers is separated into a countable union of disjoint subsets, and convergence of several series is established with the help of such a separation. This is done in Section 4. The strong consistency of the estimates given in Section 2 is proved in Section 5, and a few numerical examples are given in Section 6. In the last section some concluding remarks are given.

2. 2.1.

Problem Setting and Estimation Algorithms Problem Setting

Let the linear subsystem be with RTF CðzÞvkþ1 ¼ DðzÞuk ,

ð1Þ

where CðzÞ ¼ 1 þ c1 z þ . . . þ cp zp ,

ð2Þ

DðzÞ ¼ 1 þ d1 z þ . . . þ dq zq

ð3Þ

are polynomials in backward-shift operator zuk ¼ uk1 . It is noticed that in the special case CðzÞ ¼ 1, the linear subsystem is with FIR and is considered in [10,11]. The purpose of this paper is to estimate the coefficients ci , i ¼ 1, . . . , p and dj , j ¼ 1, . . . , q and the value fðvÞ of the nonlinearity at any v on the basis of the designed inputs fuk g and observations fzk g: zk ¼ yk þ " k ,

yk ¼ fðvk Þ,

where f"k g is the observation noise. We now list conditions to be used in the sequel.

ð4Þ

H0 CðzÞ and DðzÞ are coprime and CðzÞ is stable. H1 The function fðÞ satisfies the following condition jfðxÞ  fðyÞj < Lð1 þ jxj þ jyj Þjx  yj 8x,y 2 R; ð5Þ where L > 0 and   0 are some constants. H2 fk g is a sequence of mutually independent random variables with Ek ¼ 0 and supk E2k < 1, and is independent of fuk g. Remark 1. Condition H1 proposed here is probably used for the first time. H1 is more general than the global Lipschitz condition, which requires  ¼ 0 in (5), but it is stronger than the local Lipschitz condition, by which in lieu of Lð1 þ jxj þ jyj Þ in (5) there may be any continuous function of x, y. Under H1 the growth rate of fðzÞ is not faster than a polynomial as z tends to infinity. This property is used in (40) to guarantee an acceptable approximation to yk whenever vk is well estimated, which is needed for deriving convergence of some stochastic series in Lemma 8. Remark 2. We do not claim any necessity of H1. As shown later (Section 6, Example 3), even if H1 does not hold, the identification algorithm given in the paper still provides satisfactory estimates for both linear and nonlinear subsystems. However, it is not clear if it works in a feedback case. Since fuk g is at designer’s disposal, let uk ¼ 0, 8k  0. Then under Condition H0, (1) can be expressed as k1 X hj ukj , h0 ¼ 1, ð6Þ vkþ1 ¼ C1 ðzÞDðzÞuk ¼ j¼0

where the coefficients hj exponentially decay:   jhj j ¼ O ej ,  > 0: 2.2.

ð7Þ

Stochastic Approximation Method

Since our identification algorithms are based on an improved version of classical stochastic approximation (SA) algorithm, we first briefly introduce SA and its improved version.

583

Identification for Wiener Systems

For an unknown function gðÞ with single root denoted by x0 , the classical Robbins-Monro (RM) algorithm, being the first version of SA, suggests to seek the root x0 by the simple recursion: xkþ1 ¼ xk þ ak gkþ1 ,

k ¼ 0, 1, . . . ,

ð8Þ

where 

gkþ1 ¼ gðxk Þ þ kþ1

ð9Þ

is the noisy observation of gðxk Þ, xk is the k-th estimate for x0 , kþ1 is the observation noise, and fak g is a sequence of positive real numbers tending to zero as k ! 1. For convergence of xk to x0 , a certain growth rate restriction on gðÞ or an a priori boundedness assumption on fxk g are required in the classical theory in addition to conditions on the noise fk g. These requirements quite restrict the application of SA. Moreover, the noise is usually required to analyze for the whole sequence of estimates fxk g. In order to extend the range of applicability of SA method, the RM algorithm (8) is modified to be truncated at expanding bounds in [3]. By this the increasing rate of fxk g is controlled. Let fMk g be an increasing sequence of positive real numbers with Mk ! 1, and let xk be generated by k!1

the following algorithm:  xk þ ak gkþ1 , if jjxk þ ak gkþ1 jj  Mk , xkþ1 ¼ otherwise, x ; ð10Þ k ¼ k1 þ I½jxk1 þak1 gk j>Mk1  ,

0 ¼ 0,

ð11Þ

where function IA denotes the indicator function of a set A, and k is the number of truncations occurred until time k  1. From (11) it is seen that k  k  1, and from (10) jjxkþ1 jj  jjx jj ^ Mk1 . This means that the growth of jjxkþ1 jj should not be faster than that of Mk . At any time if jjxk þ ak gkþ1 jj exceeds the truncation bound Mk , then we pull xkþ1 back to the fixed point x and simultaneously extend the truncation bound from Mk to Mk þ1 . Otherwise, it develops as (8). Since the region where x0 is located is unknown, it is important to allow xk to grow up in order to have possibility to reach x0 . For convergence of fxk g we must have ak gðxk Þ ! 0. Thus, it is reasonable to require k!1

ak ! 0. However, if ak tends to zero too fast so that Pk!1 1 may be unable to k¼1 ak < 1, then the estimate xkP reach x0 . Thus, we have to require 1 k¼1 ak ¼ 1 (see 1 [3]). A common choice is ak ¼ k. The selection of

fMk g is defined by that how fast we allow fxk g to grow up. The following fact to be used in what follows is a special case of the general convergence theorem (GCT) Theorem 2.2.1 of [3]. Let gðxÞ ¼ ðx  x0 Þ and ak ¼ 1=k. Then xk defined by (10)(11) converges to x0 for any sample paths satisfying   mðnX  k , Tk Þ   ai iþ1  ¼ 0 8Tk 2 ½0, T lim lim sup T1   i¼n  T!0 k!1 k

ð12Þ for any nk such that fxnk g converges, where   m X aj  T : mðk, TÞ ¼ max m : j¼k

Further, if fk g in (9) for the sample path ! under consideration can be decomposed into two parts k ¼ k0 þ k00 such that 1 X

0 a1 k k < 1,

k00 ¼ Oðak Þ

k¼1

for some  2 ð0, 1, then by the convergence rate theorem (CRT) Theorem 3.1.1 of [3] xk given by (9)(10)(11) converges to x0 with the following rate: k xk  x0 k ¼ oðak Þ: It is worth noting that Condition (12) is required to verify only along convergent subsequences fxnk g rather than along the whole sequence fxn g. As pointed in [3], in many cases (12) cannot be directly verified along the whole sequence fxn g, but it can be done for convergent subsequences. This is the case in Theorem 2 of Section 5. The method verifying Condition (12) along convergent subsequences has shown a great advantage over verification along the whole sequence, and it is called as trajectory-subsequence (TS) method in [3]. 2.3.

Identification Algorithms

As stated in introduction, we first estimate the parameters ci , i ¼ 1, . . . , p, dj , j ¼ 1, . . . , q of the linear part, then construct v^k to approximate the output vk of the linear subsystem; finally, identify the nonlinear block by using the data f^ vk , zk g. We take the input sequence fuk g1 1 to be iid and uk 2 N ð0, 1Þ. Under condition H1, let us set Z 1 X t2 1  2  2 hi , and  ¼ pffiffiffiffiffiffi tfðtÞeð22 Þ dt:  ¼ 2 3 R i¼0 ð13Þ

584

X.-L. Hu and H.-F. Chen

P 2 Clearly, vk obeying N ð0, k2 i¼0 hi Þ is asymptoti2 cally distributed as N ð0,  Þ, and  is the asymptotic correlation coefficient of the input and output of the nonlinearity. Define the real sequences to be used in algorithms: ak ¼ k1,

bk ¼ k1 ,

Mk ¼ M0 þ k

and

ð14Þ

with > 0, > 0, M0 > 0, and 3 þ < 12. Similar to [4,10,11], hi , i ¼ 0, 1, . . . , p þ q are estimated by the SA algorithms with expanding truncations: kþ1 ðiÞ ¼ ½ k ðiÞ  ak ð k ðiÞ  uk zkþiþ1 Þ  I½j k ðiÞak ð k ðiÞuk zkþiþ1 ÞjM k ðiÞ  ,

ð15Þ

Then (19) can be written in the vector form ½hqþ1 hqþ2 . . . hqþp T ¼ ½c1 c2 . . . cp T ,

which is called the Yule–Walker equation. Replacing hi , i ¼ q, . . . , p þ q, in (21) with their estimates  ðiÞ k  if k ð0Þ 6¼ 0 hik ¼ k ð0Þ 0 otherwise and denoting by k the resulting matrix corresponding to , we derive a linear equation with respect to ½c1 . . . cp T at time k. The solution ½c1k . . . cpk T whenever exists to the obtained linear equation serves as an estimate of ½c1 . . . cp T , i.e., T ½c1k . . . cpk T ¼ 1 k ½hqþ1, k . . . hpþq, k  :

k ðiÞ ¼ k1 ðiÞ þ I½j k1 ðiÞak1 ð k1 ðiÞuk1 zkþi Þj>M k1 ðiÞ  , ð16Þ

0 ðiÞ ¼ 0



E½vkþlþ1 uk  ¼ hl ,

l ¼ 0, 1, . . . :

ð17Þ

Multiplying both sides of (1) by uki and taking the mathematical expectation lead to hi ¼ E½vkþ1 uki  ¼ 

i^p X

cl hil þ di ,

i ¼ 1, . . . , q,

l¼1

hj ¼ E½vkþ1 ukj  ¼ 

i^p X

clk hil, k ,

i ¼ 1, . . . , q:

Thus, we have estimated all coefficients of the linear subsystem. It remains to estimate fðvÞ at any fixed v. Let us use these estimates to approximate the intermediate signal vk first. Define 0 1 1 0  0 c1 B c 0 1  0 C B C 2 B C B    C C ¼ B  C , B C 0  1 A @ cp1 0 cp 0 0    0 pp 0 1 1 d 1    dq B 0 0  0 C B C D¼B , C @   A 0

0

Uk ¼ ½uk . . . ukq T ,



0

pðqþ1Þ

H ¼ ½1 0 . . . 01p

xkþ1 ¼ Cxk þ DUkþ1 :

cl hjl ,

ð23Þ

l¼1

ð18Þ j^p X

ð22Þ

Motivated by (18), the estimates for di are given by dik ¼ hik þ

with arbitrary initial values 0 ðiÞ, i ¼ 0, 1, . . . , p þ q. We note that k ð0Þ serves as the estimate for , and ð k ðiÞ= k ð0ÞÞ if exists serves as the estimate for hi , i ¼ 1, . . . , p þ q. In order to estimate the coefficients of the linear subsystem let us first derive the corresponding Yule– Walker equation. Since fuk g is iid with Euk ¼ 0, from (6) it follows that

ð21Þ

ð24Þ

l¼1

j ¼ q þ 1, . . . , p þ q,

ð19Þ

where and hereafter a ^ b denotes minða, bÞ. Define 0

hq h B qþ1 ¼B @  hpþq1 

hq1 hq  hpþq2

1    hqpþ1    hqpþ2 C C,   A  hq

where hi ¼ 0 for i > 0 by definition.

It is clear that with x0 appropriately chosen the output vk of the linear subsystem coincides with Hxk : vk ¼ Hxk :

ð20Þ

ð25Þ

Replacing the elements of matrices C and D with their estimates derived above gives Ck and Dk which naturally serve as estimates for C and D, respectively. Define x^kþ1 ¼ Ckþ1 x^k þDkþ1 Ukþ1 ,

v^k ¼ H x^k

ð26Þ

with an arbitrary x^0 . We use v^k as an estimate for vk .

585

Identification for Wiener Systems

For estimating fðvÞ we need the kernel function 

wk ¼

pffiffiffiffi 2 bk exp

  2  vkbv þ k

v2k 220

 ð27Þ

,

where 0 > 0 is a fixed constant. The kernel function wk can be estimated by 

^k ¼ w

pffiffiffiffi 2 bk exp

(

v^k v 2  þ bk

) v^2k 220

:

ð28Þ

Finally, fðvÞ is estimated by k ðvÞ given by the following recursive algorithm:

Consequently, we have lim Ewk yk

k!1

Z pffiffiffiffi





x2 22 0

x2

2 ðxv bk Þ þ

 1 22 k fðxÞpffiffiffiffiffiffi e dx 2 k (

) Z ðvþbk tÞ2 1 1 t2 1 dt  ¼ lim e fðvþbk tÞk exp k!1 R 2 20 2k pffiffiffi ¼ fðvÞðvÞ: &

¼ lim b2 e k!1 k R

Lemma 2. Assume H1 holds. Then lim E½ukj ykþ1  ¼ hj ,

j ¼ 0, 1, . . . , p þ q,

k!1

^k ðk ðvÞ  zk Þ kþ1 ðvÞ ¼ ½k ðvÞ  ak w  I½jk ðvÞak w^k ðk ðvÞzk ÞjMk ðvÞ  ,

ð29Þ

k ðvÞ ¼ k1 ðvÞ þ I½jk1 ðvÞak1 w^k1  ðk1 ðvÞ  yk1 Þj > Mk1 ðvÞ ,

0 ðvÞ ¼ 0 ð30Þ

with any given initial value 0 ðvÞ.

3.

Limiting Behaviors of Some Random Variables

We now establish the limiting behaviors of some random variables connected with the kernel and the a.s. convergence of some series. Lemma 1. Assume H1 holds. For wk defined by (27) the following limits take place: pffiffiffi pffiffiffi ðvÞ, lim E½wk yk  ¼ fðvÞðvÞ, k!1 k!1 pffiffiffi ð31Þ lim E½wk jyk j ¼ jfðvÞjðvÞ, lim Ewk ¼

k!1

pffiffiffiffiffi pffiffi lim E½ bk wk 2 ¼ 2ðvÞ, k!1 pffiffiffiffiffi pffiffi lim E½ bk wk yk 2 ¼ 2fðvÞðvÞ,

where  and  are defined by (13). Proof. Noticing that ukj and vkþ1 are jointly Gaussian and by (6) h i h E ukj  2j vkþ1 vkþ1 ¼ 0, k

h

we find that ukj  2j vkþ1 is independent of vkþ1 . k Taking the conditional expectation with respect to vkþ1 leads to

hj E ukj jvkþ1 ¼ 2 vkþ1 : k By this we have



 E ukj fðvkþ1 Þ ¼ E E ukj fðvkþ1 Þjvkþ1 

 ¼ E fðvkþ1 ÞE ukj jvkþ1 ¼ hj

Lemma 3. Assume H2 holds. Then the following series converge a.s.:



n 2 v 2

1 0

 12

ak w k  k ,

k¼1 1 X

k!1

where ðvÞ ¼ 1 exp

E½vkþ1 fðvkþ1 Þ , 2k

which by (13) yields the desired result by letting k ! 1 because 2k ! 2 and vk is asymptotically & distributed as N ð0, 2 Þ.

1 X

ð32Þ

ð33Þ

1 X k¼1

ak wk ðjk j  Ejk jÞ, ð34Þ

ak uk kþj :

k¼1

o .

Proof. It suffices to show limk!1 Ewk yk , and the rest can be provedP in a similar manner. k2 hj ukj1 , and hence vk 2 N ð0, 2k Þ By (6) vk ¼ j¼0 P  k2 2 with 2k ¼ j¼0 hj .

Proof. The proof is essentially based on the following fact (see Chapter 11.1, Theorem 3 of [6]): Let fSn , F n , n  1g be a martingale with supn EjSn j < 1, and let fYn , F n1 , n  1g be with supn jYn j < 1 a.s. 1 P Xn Yn < 1 a.s. where Xn ¼ Sn  Sn1 . Then n¼1

586

X.-L. Hu and H.-F. Chen

For example, ak wk k can be written as ak wk k ¼ Xk Yk    2 . Since and Yk ¼ exp  vkbv k

pffiffiffiffi 2 with Xk ¼ k1

k

0 < < 1/6, applying the above cited result leads to 1 P Xn Yn < 1 a.s. n¼1

The proof for convergence of the second series is similar. The convergence of the last series follows from the fact that fuk kþj g is a martingale difference sequence. In what follows we need the divergence rates of uk and yk as k ! 1. Lemma 4. For any  > 0 the following rate takes place uk a:s: ! 0: k k ! 1

ð35Þ

If C(z) is stable and H1 holds, then k a:s: ! 0, k k ! 1

yk a:s: ! 0: k k ! 1

vk a:s ! 0, k k!1

  1 X juk j P  > " < 1: k k¼1

4.

&

Convergence Analysis via Separation of Integer Set

Since the linear subsystem is with RTF, its output is no longer a combination of inputs of the system. This makes convergence analysis for some series difficult. To overcome the difficulty we truncate the expression (6) for vkþ1 as follows vkþ1 ¼

k X i¼0

hi uki ,

 k ¼ ½k , 0 <  < 1,

ð39Þ By (5) of Condition H1 and Lemma 4, we further have vkþ1 Þj jykþ1  ykþ1 j ¼ jfðvkþ1 Þ  fð  < Lð1 þ jvkþ1 j þ j vkþ1 j Þjvkþ1  vkþ1 j    ¼ Lð1 þ jvkþ1 j þ j vkþ1 j ÞO e2k    a:s:, ð40Þ ¼ O e 4 k and similarly,

which incorporating with H1 implies

yk a:s ! 0. k k!1

Let us explain the idea of truncations applied in (37). Assume  ¼ 1=2. Then vk2 þ1 is a linear combination of fuk2 k , uk2 kþ1 , . . . , uk2 g for any positive integer k. By the independence of the sequence fuk g and the fact k2  k > ðk  1Þ2 , it is clear that the subsequence f vk2 þ1 , k ¼ 0, 1, . . .g now turns to be a sequence of mutually independent random variables. On the other hand, when the linear subsystem is stable, the slowly truncated vkþ1 approximates vkþ1 as k ! 1 based on the following analysis. By (7) and (35) it is clear that   X  k1      hi uki  ¼ O e2k jvkþ1  vkþ1 j ¼  a:s:   

ð36Þ

Then by the Borel–Cantelli lemma (35) follows. By stability of C(z) from (24)(25) and (35) we have

ð38Þ

i¼kþ1

Proof. For any  > 0 by the Tchebychev inequality " #   juk j juk j2= 1 P  >" ¼P > "2= < 2= 2 Ejuk j2= , k2 k " k which implies

where and hereafter ½x denotes the integer part of a real number x. Correspondingly, we set (

) pffiffiffiffiffiffi vk v 2 v2k 2 k ¼ yk ¼ fð vk Þ, w exp  þ 2 : bk bk 20

   kþ1 j ¼ O e4k jwkþ1  w

a:s:,

ð41Þ

   2kþ1 j ¼ O e4k jw2kþ1  w

a:s:

ð42Þ

PTo analyze convergence of the series like ak ðuk ykþ1  Euk ykþ1 Þ, we now separate the totality of integers into a countable union of disjoint sets. Let f k g be a sequence of strictly increasing integers such that 0 ¼ 0, 1 ¼ 1, f kþ1  k g also nondecreasingly diverges to infinity, and 0 < lim inf k!1

k ka

 lim sup k!1

k ka

<1

for some a > 1: ð43Þ

Define ð37Þ



A0 ð0Þ ¼ f k : k ¼ 1, 2, . . .g;

ð44Þ

587

Identification for Wiener Systems 

Aj ðiÞ ¼ f k þ i : k ¼ j, j þ 1, . . .g, 

Aj ¼

i[ 2 ðjÞ

Aj ðiÞ,

ð45Þ



j ¼ 1, 2, . . . ,

l ðj, iÞ ¼ 

i¼i1 ðjÞ 



where i runs from i1 ðjÞ ¼ j  j1 to i2 ðjÞ ¼

jþ1  j  1. If i2 ðjÞ < i1 ðjÞ, then set Aj ¼ . We denote by J the totality of integers f1, 2, . . .g. Lemma 5. The integer sets Aj ðiÞ, j ¼ 0, 1, . . . , compose a disjoint separation of J, i.e., Aj ðiÞ, i ¼ i1 ðjÞ, . . . , i2 ðjÞ, j ¼ 0, 1, . . ., are disjoint and A0 ð0Þ [

1 i[ 2 ðjÞ [

Aj ðiÞ ¼ J:

Set

ð46Þ

jas1 1=2 s  j1þl þi :

j1þl þ i

By (43) we have ! jas1 1=2 jas1 1=2 1  s < ¼O

sj1þl

j1þl þ i ðj  1 þ lÞ1 þ1=2

1 ð50Þ ¼ O  þ1=2 , l1 which combining with (49) yields El ðj, iÞ ¼ 0 and

j¼1 i¼i1 ðjÞ

El2 ðj, iÞ < Proof. The assertions of the lemma follow from the following set operations by noticing that f k þ 1, . . . , kþ1  1g, k ¼ 1, 2, . . ., are disjoint: 1 i[ 2 ðjÞ [

Aj ðiÞ ¼

 j 1 [ 1 jþ1[ 1 [

f k þig

j¼1 i¼ j  j1 k¼j

j¼1 i¼i1 ðjÞ

¼

 j 1 1 [ 1 jþ1[ [

f k þig

¼

C l1þ21

ð51Þ

for some constant C > 0. By the condition of the lemma for fixed j and ifl ðj, iÞg is a sequence of mutually independent random variables with possible exception of a finite number of terms. Then, by the convergence theorem for martingale difference sequences (mds)

j¼1 k¼j i¼ j  j1 1 [ k [

Sji ðmÞ ¼

f k þ j  j1 ,

k þ j  j1 þ1, ..., k þ jþ1  j 1g 1 [ & ¼ f k þ1, k þ2, ..., kþ1 1g: Lemma 6. Let fk g be a sequence of random variables with zero mean and supk Ek2 < 1. If for any fixed i and j, i ¼ i1 ðjÞ, . . . , i2 ðjÞ whenever i2 ðjÞ  i1 ðjÞ, j ¼ 1, 2, . . . , the subsequence fk : k 2 Aj ðiÞg is composed of mutually independent random variables with possible exception of a finite number of k , then

for s >  3 2

1 a,

l ðj, iÞ

a.s. converges to some random variable Sji as m ! 1. Given  > 0 by (50)(51) we have

P jSji ðmÞj > j 

k¼1

a:s:

m X l¼1

k¼1 j¼1

1 X 1  <1 s k k k¼1

ð49Þ

ð47Þ

where a is given in (43).

3 1 1 a1 Proof. From a1 s > 2  a 1it  is clear that as  a þ 2 > 2 . Let  2 2 , as  a þ 2 and let 1 > 0 such that 1 ð48Þ as  a þ    1 > 0: 2 Since s > 12, under the conditions of the lemma it is P 1 clear that ks k < 1 a.s. Consequently, the k2A0 ð0Þ

index set A0 ð0Þ can be excluded from consideration.



1 2 j2

EjSji ðmÞj2

1 C X 1 C0 ¼ 2 2 , 2 2 1þ2 1  j  j l¼1 l

P 1 where C0 ¼ C 1 l¼1 l1þ21 . Letting m ! 1 we have

P jSji j > j  C2 j02 :

ð52Þ

Define   j ðÞ¼ ! : jSji j  j , i ¼ i1 ðjÞ, . ..,i2 ðjÞ , ð53Þ 1 \  j ðÞ: ðÞ¼ j¼1

In the case a  2 it is clear that i2 ðjÞ  i1 ðjÞ  1 ¼ jþ1  2 j þ j1 ¼ Oððj þ 1Þa  2ja þ ðj  1Þa Þ ¼ Oðja2 Þ:

588

X.-L. Hu and H.-F. Chen

Thus, similar to (55) for ! 2 j ðÞ by Lemma 5 it follows that

By (52)(53) it follows that !

1 [

PððÞÞ ¼ 1  P

cj ðÞ

>1

j¼1

>1

1 X j¼1

1  X P cj ðÞ j¼1

  i2 ðjÞ X 1 1 X X 1  jSji j      k s as  k¼1 k  j¼1, A 6¼ j 1 1=2 i¼i ðjÞ j

  C0 O ja2 2 2  j

¼

! 1 1 X 1 >1O 2 :  j¼1 j2þ2a

j¼1, Aj 6¼

ð54Þ

By the definition of , we have that 2 þ 2  a > 1 and, hence, PððÞÞ ! 1. !1

For ! 2 ðÞ by Lemma 5 it follows that      i2 ðjÞ X X  1 1 X 1 X   1   1   ¼    k s jþl1 þi   s  k¼1 k   j¼1 ð jþl1 þ iÞ  i ðjÞ l¼1

¼ O ð 1Þ



j¼1 i¼i1 ðjÞ

¼O 

!

jas1 ða2Þ1=2

i

< 1:

< 1:

!1

Therefore,

0

>1

jas1 1=2

j¼1

i¼1

Similar to (56), for j ðÞ defined by (53) we have   ! : jSji j  kj , , if i2 ðjÞ ¼ i1 ðjÞ ¼ i, j ðkÞ ¼  otherwise:

1 [

1 X

j¼1, Aj 6¼ 1 X

1 cj ðÞA

 P cj ðÞ

C0 2 j2  j¼1, Aj 6¼ 1 X 1 ¼ 1  O ð 1Þ 2 ! 1: 2 ia1 !1  i¼1

>1

1

1 1 a1 ðas1 1=2Þ

j¼1, Aj 6¼

jSji j

1 X

1 X

PððÞÞ ¼ 1  P@

   X i2 ðjÞ X 1 1 X l ðj, iÞ  ¼  as 1=2    j¼1 i1 ðjÞ l¼1 j 1

Oð1Þ jas1 1=2

To complete the proof it remains to show PððÞÞ ! 1.

1

i2 ðjÞ 1 X X

1

1 X

&

ð55Þ

Lemma 7. Assume C(z) is stable. Let s 2 ð12, 32Þ, P2 ð0, s  12Þ, and vk be defined by (37). Then 1  ks vk < 1 a.s.

In the case 1 < a < 2 we note that for sufficiently large j, ð jþ1  j Þ  ð j  j1 Þ equals either 0 or 1. In other words, i2 ðjÞ  i1 ðjÞ equals either  1 or 0, and correspondingly,

Proof. Since fuk g is iid and Euk ¼ 0, we have E vk ¼ 0 1 2 , 32s Þ. It is clear that and supk v2k < 1. Take a 2 ð1 3 1 s > 2  a, which is required in Lemma 6. It suffices to show that f vk : k 2 Aj ðiÞg is mutually independent. If this is done, then the assertion of the lemma follows from Lemma 6. By (37) v½ðkþ1Þa þi is a linear combination of u½ðkþ1Þa þi , . . . , u½ðkþ1Þa þi½ðkþ1Þa þi , where by (31)

Aj ¼

i[ 2 ðjÞ

 Aj ðiÞ ¼

i¼i1 ðjÞ

Aj ðiÞ; if i2 ðjÞ ¼ i1 ðjÞ ¼ i, , otherwise: ð56Þ 1

By noticing that i1 ðjÞ ¼ i implies j ¼ Oðia1 Þ, for ! 2 j ðÞ we have jSji j as j 1 1=2





! 0:

jas1 1=2 j!1

½ðk þ 1Þa  þ i ¼ ½ð½ðk þ 1Þa  þ iÞ . The total number of these terms is ½ðk þ 1Þa  þ i þ 1 ¼ Oðka Þ. Let us denote by Uk the minimal subset of fuj , j ¼ 1, 2, . . .g such that v k þi is the linear combination of elements from Uk. On the other hand, the neighboring terms in f vk , k 2 Aj ðiÞg are with subscripts kþ1 þ i and k þ i, and ð kþ1 þ iÞ  ð k þ iÞ ¼ Oðka1 Þ. This means that

589

Identification for Wiener Systems

starting from some large enough k0 fUk , k ¼ k0 , k0 þ 1, . . .g are disjoint subsets of fuj , j ¼ 1, 2, . . .g whenever a < a  1. Therefore, f vk : k 2 Aj ðiÞg is a sequence of mutually independent random variables with possible exception of finite number of terms. Thus, the conditions of Lemma 6 are satisfied. Lemma 8. Assume C(z) is stable and H1 holds. Then the following series converge a.s. 1 X ak ðukj ykþ1  Eukj ykþ1 Þ 8j  0, k¼1 1 X

ð57Þ ak ðwk yk  Ewk yk Þ,

k¼1 1 X

k ð0Þ !  k!1

a:s:,

k ðiÞ ! hi k!1

ð61Þ

a:s:,

i ¼ 1, . . . , p þ q with convergence rates j k ð0Þ  j ¼ oðk Þ a:s:,

j k ðiÞ  hi j

ð62Þ

a:s

ð58Þ ak ðjwk  Ewk j  Ejwk  Ewk jÞ,

in addition,  6¼ 0 and det  6¼ 0, then cik ! ci , k!1

a:s

i ¼ 1, . . . , p, dik ! dj , j ¼ 1, . . . , q with the same k!1

rate as that given in (62), where cik and djk are given by (22)(23), respectively. ak ðwk jyk j  E½wk jyk jÞ:

ð59Þ

k¼1

Proof. Let us prove convergence of the first series in (57). The rest can be done by a similar treatment. By (35)(40) and H1 it is clear that 1 X k¼1 1 X

ak ukj ðykþ1  ykþ1 Þ < 1 a:s:,

and

ak ½Eukj ðykþ1  ykþ1 Þ < 1 a:s:

k¼1

Thus, for the first series in (57) it suffices to show 1 X

Theorem 1. Assume H0-H2 hold. Then

where 0 <  < 1=2 and  is given in (13). Further, if,

ak ðwk  Ewk Þ,

k¼1 1 X

We now in a position to establish the strong consistency of estimation algorithms given in Section 2.

¼ oðk Þ a:s:, i ¼ 1, . . . , p þ q,

k¼1 1 X

5. Convergence of Identification Algorithms

ak ðukj ykþ1 Eukj ykþ1 Þ < 1 a:s:

ð60Þ

k¼1

Notice that Eðukj ykþ1 Eukj ykþ1 Þ ¼ 0 and sup Ejukj ykþ1 Eukj ykþ1 j2 < 1: k

Similar to Lemma 7, separating fuj , j ¼ 1, 2, . . .g to the disjoint subsets Uk with s ¼ 1 and  2 ð0, 1=2Þ, we find that fukj ykþ1 Eukj ykþ1 , k 2 Aj ðiÞg is a sequence of mutually independent random variables with possible exception of a finite number of terms. Then (54) follows from Lemma 6. & Remark 3. As shown in Lemma 6, the stepsize ak in (57)-(59) may be replaced by k1s with s > 12.

Proof. Rewrite (15) as 8 k ðiÞak ð k ðiÞhi Þak kþ1 ðiÞ, > > > < if j ðiÞa ð ðiÞh Þa  ðiÞj k k k i k kþ1 kþ1 ðiÞ¼ > M ðiÞ,

k > > : 0, otherwise, where kþ1 ðiÞ ¼ uk zkþiþ1 þ hi ,

ð63Þ

i ¼ 0, 1, . . . , p þ q:

ð64Þ In the sequel we will use the general convergence theorem (GCT) and the convergence rate theorem (CRT) for the SA algorithm with expanding truncations mentioned in Section 2.2 and cited in [11] with reference to [3] for detailed proof. From (63) it is seen that by GCT with regression  functions gðiÞ ðxÞ ¼ ðx  hi Þ for (61) it suffices to show   mðn, tÞ  1  X  lim lim sup  aj jþ1 ðiÞ ¼ 0  T!0 n!1 T j¼n

ð65Þ

8t 2 ½0, T,

i ¼ 0, 1, . . . , r, ( ) m P where mðn, tÞ ¼ max m : aj  t . j¼n Noticing kþ1 ðiÞ ¼ ðuk ykþiþ1  E½uk ykþiþ1 Þ  ðE½uk ykþiþ1   di Þ  uk kþiþ1 , we conclude (65) by Lemma 2 and the convergence of the first series in (57).

590

X.-L. Hu and H.-F. Chen

For (62) it suffices to set  ¼ 1, F ¼ 1, and 00 k ¼ 0 in CRT stated in [11]. The assertions concerning ci and dj follow from (22), (23) and the assumptions  6¼ 0 and det  6¼ 0. &

By stability of C and by the convergence Ck ! C, there exist N1 > 0 and 2 ð0, 1Þ such that

Theorem 2. Assume that H0-H2 hold,  6¼ 0 and det() 6¼ 0 where  and  are given in (13) and (20), respectively. Then for k(v) given by (29)(30)

where

k ðvÞ ! fðvÞ

Sð , kÞ ¼ N2

ð66Þ

a:s:

k!1

x0  x0 jj þ Sð , kÞ, jj x^kþ1 xkþ1 jj  N1 kþ1 j j^

kþ1 X

ð69Þ

kjþ1

j¼1

 ðjjCj  Cjj  jjxj1 jj þ jjDj  Djj  jjUj jjÞ

Proof. We rewrite (29) as pffiffiffi 8 k ðvÞ  ak ðvÞðk ðvÞ  fðvÞÞ þ ak ek ðvÞ, > > pffiffiffi > < if jk ðvÞ  ak ðvÞðk ðvÞ  fðvÞÞ kþ1 ðvÞ ¼ > þak ek ðvÞj  Mvk ðvÞ, ð67Þ > > : 0, otherwise, where ek ðvÞ ¼

pffiffiffi ^k ðk ðvÞ  zk Þ: ðvÞðk ðvÞ  fðvÞÞ  w

Byffiffiffi noticing that f(v) is the root of the linear function p ðx  fðvÞÞ, by GCT for the theorem it suffices to show   mðki , Ti Þ  1  X  ðsÞ aj ej ðvÞ ¼ 0, ð68Þ lim lim sup   T!0 i!1 T j¼k i

s ¼ 1, 2, 3, 4

8Ti 2 ½0, T

for any convergent subsequence ki ðvÞ ! ðvÞ, where ð1Þ

ð2Þ

ð3Þ

ð4Þ

ek ðvÞ þ ek ðvÞ þ ek ðvÞ þ ek ðvÞ ¼ ek ðvÞ, ð1Þ

^k Þðk ðvÞ  zk Þ, ek ðvÞ ¼ ðwk  w pffiffiffi ð2Þ ek ðvÞ ¼ ð ðvÞ  wk Þk ðuÞ, pffiffiffi ð3Þ ek ðvÞ ¼ wk yk  ðvÞfðvÞ, ð4Þ

ð4Þ

ek ðvÞ ¼ ek ¼ wk k : By the second limit in (31) and the convergence of the second series in (57) it follows that (68) holds for s ¼ 3. Further, by Lemma 3 (the first series in (34)) (68) holds for s ¼ 4. We now show (68) for s ¼ 1. For this we estimate ^k wk j. Since 12  3  > 0, there is a $ > 0 jw such that 12  3   3$ > 0. By (24)(26) it follows that x^kþ1 xkþ1 ¼ Ckþ1 ð^ xk xk Þ þ ðCkþ1  CÞxk þ ðDkþ1  DÞUkþ1 :

and

N2 > 0:

By Theorem 1 jjCk  Cjj and jjDk  Djj are of order oðk1=2$ Þ, and by Lemma 4, a:s:

xk ! 0. k$ k!1

Uk a:s: ! k$ k!1

0,

Then we have

Sð , kÞ ¼

kþ1 X



kjþ1 o

j¼1



1



j1=22$

X kþ1

1=22$ k 1=22$ j k j¼1

X

1=22$ k 1 k j ¼ o 1=22$

kjþ1 k j¼0 1 0 k

X k 2 X C 1 B ¼ o 1=22$ @ þ A j

k j¼0 k

¼o

1



kjþ1

1=22$

j¼ 2 þ1

k  kjþ1



k 1 1 2 þ1 þ 2 ¼ o 1=22$  k2  k 1 k

1 ¼ o 1=22$ : k Noticing that the first term on the right-hand side of (69) exponentially decays, we find that

jj x^k xk jj ¼ o

1

k1=22$

a:s:

Since j v^k vk j and jj x^k xk jj are of the same order, we further have



1 1 ^k wk j ¼ o 3 j v^k vk j ¼ o 3 jj x^k xk jj jw b bk

k 1 ð70Þ ¼ o 1=23 2$ : k

591

Identification for Wiener Systems

By the selection of Mk defined in (14) from (29) it follows that jk ðvÞj  k , which together with (70) implies ! 1 1 X X 1 ^k wk Þk ðvÞj ¼ o ak jðw : k3=23  2$ k¼1 k¼1 ð71Þ By Lemma 4 and (70) we have 1 1 X X ^k wk Þyk j ¼ ak jðw

1  yk  ^ ð w w Þ   k k k$ k1$ k¼1 ! 1 X 1 ¼o < 1: k3=23 3$ k¼1

k¼1

ð72Þ Again by (70) it is seen that 1 X

and s, ki ¼ 1 þ OðTÞ 8s 2 ½ki , . . . , mðki , TÞ

ð77Þ

as i ! 1 and T ! 0. By (76) and the convergence of the second series in (34) it follows that   X  s   s, jþ1 aj wj j  ¼ OðTÞ 8s 2 ½ki , . . . , mðki , TÞ   j¼k  i

as i ! 1 and T ! 0. By the last limit in (31) and (59)(77) it is seen that   X  s   s, jþ1 aj wj yj     j¼ki



s X

s, jþ1 aj ðwj jyj j  E½wj jyj j þ E½wj jyj jÞ

j¼ki

¼ OðTÞ

^k  wk Þk j ak jðw

8s 2 ½ki , . . . , mðki , TÞ. From the recursion (74) it follows that

k¼1

 ¼

1 X k¼1 1 X k¼1

^k  wk Þj  jk j ak jðw

o

ð78Þ

ðjk j  Ejk jÞ þ Ejk j k3=23 2$

< 1:

ð73Þ

Combining (71)–(73) leads to (68) for s ¼ 1. Finally, we prove (68) for s ¼ 2. For this we first show that there exists an 0 with P0 ¼ 1 such that for any ! 2 0 if ki ðvÞ is a convergent subsequence, then ^s ðs ðvÞ  zs Þ, sþ1 ðvÞ ¼ s ðvÞ  as w

ð74Þ

sþ1 ðvÞ ¼ s ðvÞ  as ws ðs ðvÞ  zs Þ ^s Þðs ðvÞ  zs Þ þ as ðws  w s X ¼ s, ki ki ðvÞ þ s, jþ1 aj wj zj j¼ki s X

þ

^j Þðj ðvÞ  zj Þ, s, jþ1 aj ðwj  w

j¼ki

which incorporating with (71)–(73) yields sþ1 ðvÞ ¼ ki ðvÞ þ OðTÞ 8s 2 ½ki , . . . , mðki , TÞ: ð79Þ

and jjsþ1 ðvÞ  ki ðvÞjj  cT, s ¼ ki ,ki þ 1, ...,mðki ,TÞ ð75Þ provided i is large enough and T > 0 is sufficiently small, where c > 0 does not depend on ki . Consider the recursion (74) starting from ki ðvÞ. Let 

i, j ¼ ð1  ai wi Þ . . . ð1  aj wj Þ,

i  j,

j, jþ1 ¼ 1:

By the first limit in (31) and the convergence of the first series in (58) it is clear that s X

aj wj ¼ OðTÞ 8s 2 ½ki , . . . , mðki , TÞ,

j¼ki

which implies log s, ki ¼ O

s X j¼ki

! aj wj

ð76Þ

This means that the algorithm (29)(30) has no truncation for s 2 ½ki , . . . , mðki , TÞ, when i is large enough and T > 0 is sufficiently small. Since Lemmas 1, 3, 4 and 8 are all proved a.s., the possibly exceptional set is with probability zero. This means that 0 may be taken with P0 ¼ 1 such that for all ! 2 0 (74)(75) are verified for large i and small T > 0. Let ki ðvÞ ! ðvÞ. Then we have   mðki , Ti Þ  1  X  ð2Þ aj ej ðvÞ   T j¼k i   mðki , Ti Þ  1  X  aj ðj ðvÞ  ðvÞÞðwj  Ewj Þ    T  j¼k i   mðk  i , Ti Þ  X  1 aj ðwj  Ewj Þ þ TðvÞ  j¼k  i   mðk i , Ti Þ X pffiffiffi  1 aj j ðvÞðEwj  Þ: ð80Þ þ T  j¼k  i

592

X.-L. Hu and H.-F. Chen

By the convergencepofffiffiffi the second series in (58), (75), and limk!1 Ewk ¼ ðvÞ (Lemma 1) we see that the first term on the right-hand side of (80) tends to zero as i ! 1 and T ! 0, while its second term tends to zero as i ! 1 by the convergence of the first pffiffiffi series in (58). Finally, by (75) and limk!1 Ewk ¼ ðvÞ the last term in (80) also tends to zero as i ! 1 and T ! 0. Thus, (68) for s ¼ 2 is proved and meanwhile the proof of Theorem 2 is completed. & Remark 4. Although the estimation algorithms consisting of (15),(16),(26)–(30) do not apply to the feedback control system, they are not complicated and suitable for implementation. Fig. 2.

6.

Numerical Experiments

In this section we give some examples with and without complete satisfaction of conditions required in theorems. For a better comparison in all simulations the linear subsystem is taken to be the same: vkþ1  0:3vk þ 0:6vk1 ¼ uk þ 0:27uk1  0:5uk2 þ 0:79uk3 with noise k 2 N ð0, 0:22 Þ and with iid input fuk g, uk 2 N ð0, 1Þ. Take Mk ¼ 0:1 þ k0:1 , 0 ¼ 2, and ak ¼ 1=k. The inputs for all simulations are taken from the same totality consisting of 6000 data. The difference between examples to be given consists in the nonlinear function fðÞ and in the selection of in bk ¼ 1=k . In all figures shown below the solid lines denote the estimates while the dotted lines the true values. Example 1. Let the nonlinear function be 8 < v  13 v3 , jvj  1, fðvÞ ¼ 2=3, v > 1, : 2=3, v < 1: Take bk ¼ k0:16 , which means ¼ 0.16. Obviously, all conditions required in Theorems 1 and 2 are satisfied. Figures 2 and 3 demonstrate how the estimates given in Section 2 approach to the true values. Figure 2 is for the coefficients of the linear subsystem, and Fig. 3 is for the nonlinear function. The estimate for fðÞ is obtained by separating the interval ½-1:5,1:5 into 100 subintervals with equal length and the values of fðÞ at endpoints of the subintervals are estimated. As shown in the figures, the simulation is consistent with the theoretical analysis.

Fig. 3.

An interesting question is that what will happen if some conditions required in theorems are not met. We show this by examples. Example 2. In (14) it is required that 3 þ < 1=2, which is violated if ¼ 0:6 or ¼ 0:9. We take the same system as that discussed in Example 1 but with different . It is clear the change of does not effect the estimates for coefficients of the linear subsystem. Figures 4 and 5 demonstrate the estimates for the nonlinear function for ¼ 0:6 and ¼ 0:9, respectively. From the figures it is seen that the larger the worse the estimates. As a matter of fact, from (27) we observe that bk ¼ 1=k characterizes the concentrating rate of vk on v. Further, from (28) we see that in order v^k be also concentrated on v, bk ¼ 1=k should not decrease faster than vk  v^k , which by Theorem 1 tends to zero not faster than 1=k0:5 . This explains why

should be small.

593

Identification for Wiener Systems

Fig. 4.

Fig. 6.

Fig. 5.

Fig. 7.

Example 3. We now consider the situation where fðÞ is discontinuous, i.e., H1 fails to hold. Let  1, v  0, fðvÞ ¼ 1, v < 0, and let the linear subsystem be the same as that in Example 1. Figure 6 corresponds to the case where the parameters of algorithms completely coincide with those in Example 1, while in Fig. 7 the only change is that ¼ 0:6 rather than 0.16. From these figures we find that the algorithms proposed in the paper work well even H1 does not hold, if is appropriately small.

7.

Concluding Remarks

The paper gives the recursive and strongly consistent estimates for the Wiener systems, when its linear

subsystem is an RTF. From the paper we see that in comparison with the FIR type linear subsystem the infinite impulse response makes a great complication in analysis even under a local Lipschitz-type condition H1 on the nonlinearity. Since Gaussian distribution is even,  defined by (13) is zero for any even functions. Thus, Theorem 1 does not apply to identifying Wiener systems with even nonlinearity. Therefore, for further research it is important to solve the even function problem. It is also of interest to weaken the conditions on fðÞ and to seek the conditions guaranteeing nondegeneracy of . To accelerate the algorithms used in the paper is of interest too. It is also worth considering the possible extension to the MIMO case.

References 1. Bendat JS. Nonlinear system analysis and identification from random data. New York Wiley 1999

594 2. Billings SA, Fakhouri FY. Identification of nonlinear systems using the Wiener model. Electron Lett 1978; 13: 502–504 3. Chen HF. Stochastic approximation and its applications. Kluwer, Dordrecht 2002 4. Chen HF. Pathwise convergence of recursive identification algorithms for Hammerstein systems. IEEE Trans Autom Control 2004; 49(10): 1641–1649 5. Chen HF. Recursive identification for Wiener model with nonlinearity being discontinuous piece-wise linear function. IEEE Trans Autom Control 2006; 51(3): 390–400 6. Chow YS, Teicher H. Probability theory. Springer, New York 1978 7. Greblicki W. Nonparametric approach to Wiener system identification. IEEE Trans Circuits Syst-I: Fundam Theory Appl 1997; 44(6): 538–545 8. Greblicki W. Recursive identification of Wiener system. Int J Appl Math Comp Sci 2001; 11(4): 977–991 9. Hasiewicz Z. Identification of a linear system observed though zero-memory nonlinearity. Int J Syst Sci 1987; 18: 1595–1607 10. Hu XL, Chen HF. Strong consistency of recursive identification for Wiener systems. Automatica 2005; 41(11): 1905–1916 11. Hu XL, Chen HF. Recursive identification for Wiener systems using Gaussian signals. to appear in Asian J Control 2007.

X.-L. Hu and H.-F. Chen

12. Hunter IW, Korenberg MJ. The identification of nonlinear biological systems: Wiener and Hammerstein cascade models. Biol Cybern 1986; 55: 135–144 13. Krzyzak A, Sasiadek JZ, Kegl B. Non-parametric identification of dynamic non-linear systems by a Hermite Series Approach. Int J Syst Sci 2001; 32(10): 1261–1285 14. Pajunen GA. Adaptive control of Wiener type nonlinear systems. Automatica 1992; 28: 781–785 15. Nordsjo¨ AE, Zetterberg LH. Identification of certain time-varying nonlinear Wiener and Hammerstein systems. IEEE Trans Signal Proc. 2001; 49: 577–592 16. Verhaegen M, Westwick D. Identifying MIMO Wiener systems in the context of subspace model identification methods. Int J Control 1996; 63(2): 331–349 17. Vo¨ro¨s J. Parameter identification of Wiener systems with discontinuous nonlinearities. Syst Control Lett 2001; 44: 363–372 18. Westwick DT, Kearney RE. A new algorithm for the identification of multiple input Wiener systems. Biol Cybern 1992; 68: 75–85 19. Wigren T. Recursive prediction error identification using the nonlinear Wiener mode. Automatica 1993; 29: 1011–1025 20. Wigren T. Convergence analysis of recursive algorithms based on the nonlinear Wiener model. IEEE Trans Automat Control 1994; 39: 2191–2206