• No results found

Robust High-Gain DNN Observer for Nonlinear Stochastic Continuous Time Systems

N/A
N/A
Protected

Academic year: 2021

Share "Robust High-Gain DNN Observer for Nonlinear Stochastic Continuous Time Systems"

Copied!
9
0
0

Loading.... (view fulltext now)

Full text

(1)

Technical report from Automatic Control at Linköpings universitet

Robust high-gain DNN observer for

non-linear stochastic continuous time systems

D.A. Murano, A.S. Poznyak,

Lennart Ljung

Division of Automatic Control

E-mail: ,

apoznyak@ctrl.cinvestav.mx

,

ljung@isy.liu.se

25th June 2007

Report no.:

LiTH-ISY-R-2803

Accepted for publication in IEEE Conference on Decison and Control,

2001

Address:

Department of Electrical Engineering Linköpings universitet

SE-581 83 Linköping, Sweden

WWW: http://www.control.isy.liu.se

AUTOMATIC CONTROL REGLERTEKNIK LINKÖPINGS UNIVERSITET

Technical reports from the Automatic Control group in Linköping are available from

(2)

Abstract

A class of nonlinear stochastic processes satysfying a Lipschitz-type strip condition and supplied by a linear output equation, is considered. Robust asymptotic (high-gain) state estimation for nonlinear stochastic processes via dierential neural networks is discussed. A new type learning law for the weight dynamics is suggested. By a stochastic Lyapunov-like analysis (with Ito formula implementation), the stability conditions for the state estima-tion error as well as for the neural nertwork weights are established. The upper bound for this error is derived. The numerical example, dealing with module-type nonlinearities, illustrates the eectiveness of the suggested approach.

(3)

Robust High-Gain DNN Observer for Nonlinear Stochastic

Continuous Time Systems

Daishi A. Murano

, Alex S. Poznyak

and Lennart Ljung

CINVESTAV-IPN, Dept. of Automatic Control, AP 14 740, CP 07000,

Mexico D.F., Mexico, e-mail: apoznyak@ctrl.cinvestav.mx ‡Linköping University, Dept. of Electr. Eng.S-581 83, Linköping, Sweden

e-mail: ljung@isy.liu.se

Abstract

A class of nonlinear stochastic processes, satisfying a ”Lipschitz-type strip condition” and supplied by a lin-ear output equation, is considered. Robust asymptotic (high-gain) state estimation for nonlinear stochastic processes via differential neural networks is discussed. A new type learning law for the weight dynamics is sug-gested. By a stochastic Lyapunov-like analysis (with Itô formula implementation), the stability conditions for the state estimation error as well as for the neural network weights are established. The upper bound for this error is derived. The numerical example, dealing with ”module”-type nonlinearities, illustrates the effec-tiveness of the suggested approach.

Keywords: Dynamic Neural Networks, Stochastic Pro-cesses, Nonlinear Observers.

1 Introduction

Due to many successful applications, the Neural Net-work technique seems to be a very effective for identiÞ-cation and controlling of a wide class of complex non-linear systems especially in the absence of a complete model information (”grey-box” models) or, even, con-sidering a controlled plant as ”a black box” (Hunt et al., 1992 [6]). In general, the NN’s can be qualiÞed as static (feedforward) and as dynamic (recurrent or differential) nets. The most of recent publications (see, for exam-ple, Haykin, 1994 [7]; Agarwal, 1997 [1]; Parisini and Zoppoli, 1994. [15]) deals with Static NN, providing the appropriate approximation of a nonlinear operator functions in the right-hand side of dynamic model equa-tions. In spite of successful implementations, they have several important disadvantages such as a slow learning rate (the weights updates do not utilize the information on the local NN structure) and a high sensitivity of the function approximation to the training data (they do not have memory, so their outputs are uniquely deter-mined by the current inputs and weights). Dynamic Neural Networks can successfully overhead these

draw-backs and demonstrate a workable behavior in the pres-ence of essential unmodelled dynamics because their structure incorporate feedback. They have been intro-duced by HopÞeld, 1984 [5] and then studied in (Sand-berg, 1991 [23](the approximation aspects), Rivithakis and Christodoulou, 1994 [21] (the direct adaptive reg-ulation via DNN), Poznyak et al., 1999 [19](the adap-tive tracking using DNN)). Several advance results con-cerning DNN have been recently obtained in (Narendra and Li, 1998 [14](the identiÞcation and control), Lewis and Parisini, 1998 [11] (NN feedback control: the sta-bility analysis) and in (Rivithakis and Christodoulou, 2000 [22]). All of these publications concern NN ap-plications considered speciÞcally only for the external environment of the deterministic nature. There are a few publications dealing with the identiÞcation or con-trol of stochastic processes with unknown dynamics via NN (see Kosmatopulos and Christodoulou, 1994 [9]and Elanayar and Shin, 1994 ). In the recent paper of the authors (Poznyak and Ljung, 2001 [20]) the identiÞca-tion of continuous time stochastic system via DNN has been studied but under the assumption that all state measurements are available and are not disturbed by a noise component.

In this paper we spread the ideas of DNN approach to the class of continuous time stochastic processes (with incomplete dynamic model description) based only on the output measurement subjected to stochastic noise corruptions. Dynamic Neural Networks given by dif-ferential equations (Differential Neural Networks) are applied to obtain the on-line state estimates. In fact, a robust high-gain DNN observer is suggested. The main goal of this paperis to show that practically the same robust DNN observers, designed for the deterministic systems with bounded perturbations, is robust with re-spect to perturbations of a stochastic nature which are unbounded (with probability one). The stochastic cal-culus (such as Itô formula and the martingale technique together with the strong large number law implementa-tion) are shown to be the effective instrument for the solution of this problem.

(4)

2 Uncertain Stochastic System Let³Ω, F, {Ft}t≥0, P

´

be a given Þltered probability space ( (Ω,F, P) is complete, the F0contains all the

P-null sets inF, the Þltration {Ft}t≥0is right continuous:

Ft+ := ∩

s>tFs= Ft). On this probability space deÞne

an m-dimensional standard Brownian motion, that is, ¡¯

Wt, t ≥ 0

¢

(with ¯W0 = 0) is an{Ft}t≥0-adapted Rm

-valued process such that

E©W¯t− ¯Ws| Fsª= 0 P− a.s.,

ω ∈ Ω : ¯W0= 0ª= 1

E©£W¯t− ¯Ws¤ £W¯t− ¯Ws¤| | Fsª= (t − s) I P − a.s.

Consider the stochastic nonlinear controlled continuous-time system with the dynamics xt:

xt= x0+ t R s=0 b (s, xs, us) dt + t R s=0 σx(s, x s, us) d ¯Ws yt= Cxt + t R s=0 σy(s, x s, us)d ¯Ws (1) or, in the abstract (symbolic) form,

   dxt= b (t, xt, ut) dt + σx(t, xt, ut) d ¯Wt x0= xgiven, t ∈ [0, ∞) dyt= Cdxt + σy(t, xt, ut)d ¯Wt (2)

The Þrst integral in (1) is an stochastic ordinary inte-gral and the second one is an Itô inteinte-gral. In the above xt∈ Rnis system state at timet, ut∈ U ⊆ Rp is a

con-trol, yt∈ Rk is the output, b : [0,∞) × Rn× U → Rn,

σx: [0, ∞) × Rn× U → Rn×m, σy: [0, ∞) × Rn× U →

Rk×m and C ∈ Rk×nis a constant matrix

characteriz-ing the state-output mappcharacteriz-ing. Below we will use the following notation: W¯x t := t R s=0 σx(s, x s, us) d ¯Ws, ¯Wty := t R s=0 σy(s, x s, us) d ¯Ws. It is assumed that

A1: {Ft}t≥0 is the natural Þltration generated by

¡¯ Wt, t ≥ 0

¢

and augmented by the P-null sets fromF. A2: (U, d) is a separable metric space with a metric d.

DeÞnition 1 The function f : [0, ∞) × Rn × U →

Rn×m is said to be an LAφ¡C2¢-mapping (or, ”strip-cone” type) if it is Borel measurable and is C2in x (al-most everywhere) for any t ∈ [0, ∞) and any u ∈ U and also there exist a matrix A ∈ Rn×n, a constant L and a

modulus of continuity φ : [0, ∞) → [0, ∞) such that for any t ∈ [0, ∞) and for any x, u, ˆx, ˆu ∈ Rn× U × Rn× U

kf (t, x, u) − f (t, ˆx, ˆu) − A (x − ˆx)k ≤ L kx − ˆxk + φ (d (u, ˆu))

kfx(t, x, u) − fx(t, ˆx, ˆu) − Ak ≤ L kx − ˆxk + φ (d (u, ˆu))

kfxx(t, x, u) − fxx(t, ˆx, ˆu)k ≤ φ (kx − ˆxk + d (u, ˆu))

(here fx(·, x, ·) and fxx(·, x, ·) are the partial

deriva-tives of the Þrst and the second order).

A3: b (t, x, u) is LAφ¡C2¢-mappings, σx(t, xt, ut) and

σy(t, x

t, ut) are L0φ

¡

C2¢-mappings (or, for simplicity,

¡ C2¢).

DeÞnition 2 A stochastic control utis called a

feasi-ble in the stochastic sense (or, s-feasifeasi-ble) for the sys-tem (2) if ut ∈ U [0, ∞) := {u : [0, ∞) × Ω → U | ut

is {Bt}t≥0-adapted}, {Bt}t≥0 is the Þltration generated

by ((yτ, uτ : τ ∈ [0, t]) , t ≥ 0) and augmented by the P

- null sets from F, and xt is the unique solution of

(2) in the sense that for any xtand ˆxt, satisfying (2),

P{ω ∈ Ω : xt= ˆxt} = 1.

The set of all s-feasible controls is denoted by Us

f eas[0, ∞). The triple (xt; yt; ut) , where (xt, yt)

is the solution of (2), corresponding to this ut,

is called an s-feasible triple. The assumptions A1-A3 guarantee that any ut from U [0, ∞) is

s-feasible. It is assumed also that the past information about (yτ, uτ : τ ∈ [0, t]) is available for the controller:

ut := u¡t, yτ∈[0,t], uτ∈[0,t)

¢

implying that ut becomes

{Ft}t≥0-adapted too.

The only sources of uncertainty in this system de-scription are the system random noise ¯Wt, the

pri-ori unknown LA φ ¡ C2¢-function b (t, x, u) and L φ ¡ C2¢

-function σx,y(t, x, u). The class Σ

un of uncertain

stochastic systems (2) satisfying A1-A3 is considered below as the main object of the investigation.

3 Differential Neural Networks Observer Consider the following DNN observer (Poznyak and Yu, 2000 [16]) with a single hidden layer1:

½

dˆxt/dt = Aˆxt+ W1,tσ (V1,txˆt) + W2,tϕ (V2,txˆt) γ (ut)

+ K(yt− C ˆxt)

(3) where ˆxt∈ Rnis its state vector at time t, σ : Rr→ Rk

is the given Lφ¡C2¢-mapping, ϕ : Rs → Rl is the

given Lφ

¡

C2¢-mapping, V

1,t ∈ Rr×n, V2,t ∈ Rs×n

are a{Bt}t≥0-adapted adjustable internal hidden layer

weight matrices, W1,t ∈ Rn×k, W2,t ∈ Rn×l are a

{Bt}t≥0-adapted adjustable the output layer weight

matrices, γ : Rp→ Rqthe given Lφ

¡

C2¢-mapping sat-isfyingkγ (u)k ≤ γ01kuk for any u ∈ U, A ∈ Rn×nis

a constant matrix characterized by LA φ

¡

C2¢-mapping.

The initial state vector ˆx0 is supposed to be

quadrati-cally integrable³Enx0k2

o

< ∞´.

1A multi layer DNN can be represented in a single layer form

(5)

Remark 1 The mappings σ and ϕ are usually as-sumed to be of the main diagonal structure with the bounded sigmoidal elements σii and ϕii, that is, σii:=

ai 1 + e−b|iˆx − ci (i = 1, min {r; k}) , ¯ σ := sup x∈Rn kσ (x)k , ¯ ϕ := sup

x∈Rnkϕ (x)k which obviously belong to the class of

Lφ¡C2¢-functions.

4 The Learning Law for IdentiÞcation Process The matrices collection (A, V1,t, V2,t, W1,t, W2,t) should

be selected to provide ”a good enough” state esti-mation process for any possible uncertain stochastic system from Σun.

Accept also some technical assumption which will be use in the main result formulation.

A4: the applied s-feasible control ut ∈ U is

quadrati-cally integrable (uniformly on t) as well the correspond-ing stochastic system dynamics xt, that is, lim sup t→∞

[Enkutk2

o

+Enkxtk2

o

] < ∞. In fact, this assump-tion restricts the consideraassump-tion by the class of BIBO-stable systems (with bounded second moment of state vector).

A5: The pair (A, C) is detectable.

A6: The gain matrix K is selected such a way that the joint matrix A := A− KC is stable and there ex-ist positive deÞnite matrices Λσ, Λ3, Λ1, Λ2, Λf, Λy, Q0

providing the existence of a positive solution to the fol-lowing algebraic matrix Riccati equation

G := P A + A|P + P RP + Q = 0 R := 2 ¯W1+ 2 ¯W2+ ΛΛ−1y Λ|+ Λ−1f

Q := Λσ+ δΛ3+ δΛy+ 2δΛ1+ Q0

(4)

has a positive solution P = P| > 0. and ¯W1 :=

W1,0Λ−11 W1,0| , ¯W2:= W2,0Λ−12 W2,0| .

A7. The large number law (Poznyak, 2000 [19]) is valid for the considered processes

       T−1 RT t=0 ³ kxtk2− E kxtk2 ´ dta.s.→ 0 T−1 RT t=0 ³ kutk2− E kutk2 ´ dta.s.→ 0 (5)

The last assumptions is not so restrictive since the only decreasing of the autocorrelation functions of the con-sidered processes is required to fulÞll (5) that automat-ically holds for the accepted constructions (Poznyak et al, 2000 [18]).

Introduce the following notations: ˆ σt:= σ (V1,txˆt) − σ (V1,0xˆt) ÿ σt:= σ (V1,0xˆt) − σ (V1,0xt) ˆ ϕt:= ϕ (V2,txˆt) − ϕ (V2,0xˆt) ÿ ϕt:= ϕ (V2,0xˆt) − ϕ (V2,0xt) σt:= σ(V1,txˆt), ϕt:= ϕ(V2,txˆt) σt(0) = 0, ϕt(0) = 0

In view of the differentiability and the properties A3 it follows ˆ σt= Dσ(V1,txˆt) ˜V1,txˆt+ ησ,t ˆ ϕtγ (ut) = q P i=1 γi(ut) Diϕ(V2,txˆt) ˜V2,txˆt+ ηϕ,tγ (ut) Dσ:= Dσ(V1,txˆt) := ∂ ∂zσ (z) |z=V1,txˆt Diϕ:= Diϕ(V2,txˆt) := ∂ ∂zϕi(z) |z=V2,txˆt ÿ σ|tΛ1ÿσt≤ ∆|tΛσ∆t (ÿϕtγ (ut))|Λ2(ÿϕtγ (ut)) ≤ λmax(Λ2)ÿϕ2tkγ (ut)k2 (6) ° °ησ,t ° °2 ≤ l1 ° ° ° ˜V1,txˆt ° ° °2, °°ηϕ,t°°2≤ l2 ° ° ° ˜V2,txˆt ° ° °2 The weights in (3) are suggested to be adjusted by the following Differential Learning Law :

                             LW1,t:= 2dtdW˜1|K1−1+ (2σte|t(Cδ+)|P + σtσ|tW˜1,t| P H3P ) = 0 LW 2,t:= 2dtdW˜2|K2−1+ 2ϕtγ (ut) e|t(Cδ+)|P + ϕtγ (ut) γ (ut)|ϕ|tW˜2,t| P H3P = 0 LV 1,t:= 2dtdV˜ | 1K3−1+ 2ˆxte|t(Cδ+)|P W1,0Dσ+ ˆ xtxˆTtV˜1,t|[D|σW1,0| P H1P W1,0Dσ+ l1Λ1] = 0 LV 2,t:= 2dtdV˜ | 2K4−1+ 2ˆxte|t(Cδ+)|P W2,0(Dϕut )| + ˆxtxˆ|tV˜2,t| £ Dϕut W2,0| P H1P W2,0(Dϕut )| + l2γ (ut) γ (ut)|Λ2] = 0 (7) et:= C ˆxt− yt ˜

Wi,t := Wi,t− Wi,0, ˜Vi,t:= Vi,t− Vi,0(i = 1, 2)

Cδ+:= (C|C + δI)−1C|, Nδ:= (C|C + δI)−1, δ > 0 H1:= (δNδΛ−11 Nδ|+ C + δΛ−1y (Cδ+)|) H3:= ¡ δNδΛ−13 Nδ|+ C + δΛ−1y (Cδ+)| ¢ Dϕut := q P i=1 γi(ut) Diϕ(V2,txˆt)

with any nonzero initial weights V1,0 and V2,0. DeÞne

the state estimation error by ∆t:= ˆxt− x and

ρ+:= lim sup t→∞ E {ρt} ρt:= 4( ¯Wty)|ΛyW¯ty+ tr {σxσx|P } + 2λmax(Λ2)˜ϕ2 ³ γ2 0+ γ21kutk2 ´ + kΛfk ³ f0+ f1kxtk2+ f2kutk2 ´ (8)

Theorem 1 (Averaged Upper Bound) If, under the assumptions A1-A7, the learning law (7) is applied,

(6)

then the following propertie of the corresponding state estimation process is guaranteed: the averaged quadratic state estimation error turns out to be bounded with the probability one: lim sup t→∞ t−1 t Z τ =0 ∆|τQ0∆τdτ a.s. ≤ ρ+ (9)

Theorem 2 (On Expected Quadratic Variation) The assumptions A1-A7 imply

h 1 − µ/pE {∆|tP ∆t} i +t→∞→ 0 µ :=qρ+ min¡P−1/2Q0P−1/2¢

where the operator [·]+ is deÞned as [z]+ :=

½

z if z > 0 0 if z ≤ 0 .

5 Simulation

5.1 Example 1 (benchmark nonlinear system) The nonlinear stochastic process generated by

       dx1,t= [−5x1,t+ 3 |x2,t| + u1,t] dt + 0.088d ¯W1,t dx2,t= [−10x2,t+ 2 |x1,t| + u2,t] dt + 0.088d ¯W2,t dyt= Cdxt+ 0.0088d ¯Wt x1,0= 0.1, x2,0= −0.2 (10) is considered. The programmed controls are the sine-wave and the saw-tooth periodic functions, that is, u1.t = u1,0sin (ωt) , u1,0 = ω = 1 and u2,t = u2,0·

(t − kτ) , t ∈ [kτ , (k + 1) τ) , u2,0 = τ = 0.25, k =

0, 1, 2, ... The DNN applied for the identiÞcations is as follows dˆxt= [Aˆxt+ W1,tσ (V1,txˆt) + W2,tϕ (V2,txˆt) ut + K(yt− C ˆxt)] dt ˆ x1,0= 0.2, ˆx2,0= 0.1 (11) σi(z) = 2/ (1 + exp (−rszi)) − 1, rs= 0.8, rf = 0.5 ϕ1,2= ϕ2,1= 0 ϕi,i(z) = 0.5/ (1 + exp (−rfzi)) − 0.05 (12) and the weights W1,t ∈ R2×4, V1,t ∈ R4×2, W2,t, V2,t ∈

R2×2 are adjusted by the learning law (7) with q =

2, γi(u) = ui(i = 1, 2). The selected parameter values

are A = µ −3.9 0 0 −2 ¶ , K = µ 5 5 ¶ , l1= l2= 0.3 Λ1= Λ2= Λ3= Λy= 0.3I, k1,v1= k1,v2 = 20000 k1,w1 = 1500, k1,w2 = 800 R = 0.0122 µ 1 0 0 1 ¶ , Q = µ 0.30643 0.12043 0.12043 4.80043 ¶ P = µ 0.05528 0.00362 0.00362 0.05530 ¶

and the initial weight values equal to

W1,0= µ 0.5 0.1 0 2 0.2 0.5 0.6 0 ¶ , V1,0=     1 0.5 0.3 1.3 0.5 1 0.2 0.6     W2,0= µ 0.5 0.1 0 0.2 ¶ , V2,0= µ 1 0.1 0.1 2 ¶

The correspondent simulation results, realized only with the use of 6 neurons, are depicted at the Þgures Fig.1-Fig.4. The value of the performance index It,

ob-tained after t = 20 sec ., is equal to

It=20= t−1 t

Z

τ =0

k∆τk2dτ = 0.00137

Figure 1: X1 and X1e.

Figure 2: X2 and X2e.

(7)

Figure 4:W1 Weights.

5.2 Example 2 (a double tank liquid system) The model, considered below, describes a double tank liquid system and corresponds to a standard laboratory process including two water tanks in series (see Fig.5)

Figure 5: The double tank processes.

The input signal ut is the voltage to the pump and the

output signal ytis the level of the lower tank. The

con-trol goal is to concon-trol the level x2of the lower tank and

indirectly the level x1of the upper tank. In the general

case, these two tanks levels are both measurable. But here we will consider the situation when the only level x2in the lower tank can be estimated directly. Our aim

is to estimate the level x1of the upper tank based only

on the output measurements corrupted by ” a white noise”. The following state space description holds:

   dx1/dt = −α1√x1+ βut dx2/dt = α1√x1− α2√x2 y = x2+ 0.01Wy

where utis the control variable (water ßow) selected as

” a single step-function”, x = (x1, x2)| are the states

of this process, Wy is the white noise component, y is

the measurable output. The parameters of the process are as follows: α1= 1.5, α2= 1.5, β = 0.3. And the

parameters of the DNN observer are as follows: σi(z) = 2/ (1 + exp (−rszi)) − 1, rs= 8, rf = 8

ϕ1,2= ϕ2,1= 0

ϕi,i(z) = 0.5/ (1 + exp (−rfzi)) − 0.05

and the weights W1,t ∈ R2×4, V1,t ∈ R4×2, W2,t, V2,t ∈

R2×2are adjusted by the learning law (7) with q = 2, γi(u) = ui (i = 1, 2). The selected parameter values

are A = µ −5.89 0.4 0 −2 ¶ , K = 5 µ 1 1 ¶ Λ1= Λ2= Λ3= Λy = 0.3I, l1= l2= 0.3 k1,w1 = 15000, k1,w2= 8000 k1,v1= 20000, k1,v2 = 200000 R = 0.0122 µ 1 0 0 1 ¶ , Q = µ 0.30643 0.12043 0.12043 4.80043 ¶ P = µ 0.00256 0.00026 0.00026 0.01108 ¶

We obtained the simulation results depicted at Figures 6-7. The value of the performance index It, obtained

after t = 20 sec .,is equal to It=20= t−1

t

Z

τ =0

k∆τk2dτ = 0.00007717

Figure 6: X1 and X1e

Figure 7: X2 and X2e

6 Conclusion

In this paper a new asymptotic differential neuro ob-server is proposed to estimate the states of continuous time stochastic process without a priory knowledge of the plant-structure. The corresponding weight matrices of this DNN - observer are adjusted by a special differ-ential learning law containing two terms: the Þrst one

(8)

is standard resemble to the ”back-propagation” scheme and the second one is proportional to the output er-ror. Even the sigmoidal nonlinearities are bounded and, in the Þrst glance, the hidden layer weights variation may not effect the identiÞcation quality, the presented theoretical study (as well as the corresponding simula-tions) show that the hidden layer weights updating sig-niÞcantly improves the observation process. Stochastic Lyapunov analysis (with the use of the Itô’s formula) is applied to prove the existence of the upper bound to the averaged quadratic estimation error. The bound-ness ”in average” is stated also for the weight matrices which are time varying during all learning time in the opposite to the static DNN which weights obligatory converge. The simulation results also demonstrate the high effectiveness of the suggested DNN-observer appli-cation even its structure is simple enough( containing the only 4 neurons). The next natural study step is to extend this approach to the case of the Adaptive Track-ing based on such DNN-observers, that is, to consider the workability of this DNN in a stochastic close-loop system.

References

[1] Agarwal M., 1997, A Systematic ClassiÞcation of Neural Network Based Control, IEEE Control Systems Magazine, 17, 75-93.

[2] Gard T.C., Introduction to Stochastic Differential equations, 1988, Marcel Dekker Inc., NY and Basel. [3] Duncan T.E., Lei Guo and B. Passik-Duncan, 1999, Adaptive Continuous-time Linear Quadratic Gaussian Control, IEEE Transaction on AC, 44 (9), 1653-1662.

[4] Elenayar and Shin , 1994, Approximation and Estimation of Nonlinear Stochastic Dynamic Systems Using Radial Basis Function Neural Network, IEEE Trans. on Neural Networks, 5, 594-603.

[5] HopÞeld J.J., 1984, Neurons with Grade Re-sponce Have Collective Computational Properties Like Those of a Two-state Neurons, Proc. Nat. Academy Sci., USA, 8, 3088-3092.

[6] Hunt K.J., Sharbaro D., Zbikovski R. and P.J.Gawthrop, 1992, Neural Networks for Control Sys-tems - A Survey”, Automatica, 28, 1083-1112.

[7] Haykin S., 1994, Neural Networks: A Compre-hensive Foundation, IEEE Press, NY.

[8] Isidori A., Nonlinear Control Systems, 3.rd Ed., 1995, New York, Springer- Verlag.

[9] Kosmatopulos E.B. and M.Christodolou, 1994, The Boltzmann g-RHONN: A Learning Machine for estimating Unknown Probability Distributions, Neural Networks, 7 (2), 271-278.

[10] Krsti´c M., Kanellakopulos I. and P.Kokotovich, Nonlinear and Adaptive Control Design, 1995, John Walley & Sons, NY.

[11] Lewis F.L. and T.Parisini, 1998, Neural Network Feedback Control with Guaranteed Stability, Int. J. Control, 70 , 337-340.

[12] Ljung L. System for IdentiÞcation. Theory for the Users (Second Edition). 1999. Prentice Hall PTR, Up-per Saddle River, NJ 07458.

[13] Ljung L. and S.Gunnarson, Adaptation and Tracking in System IdentiÞcation - A Survey, 1990, Au-tomatica, 26 , 7-21.

[14] Narendra K.S. and S.M.Li, 1998, Control of Non-linear Time-Varying Systems Using NN, Proc. of 10-th Yale Workshop on Adaptive and learning Systems, 2-18.

[15] Parisini T. and R.Zoppoli, 1994, Neural Networks for Feedback Feedforward Nonlinear Control Systems, IEEE on NN, 5, 436-449.

[16] Poznyak A.S. and Wen Yu, 2000, ”Robust Asymptotic Neuro-Observer with Time Delay Term”, Int.Journal of Robust and Nonlinear Control, 10, 535— 559.

[17] Poznyak A.S, E.N. Sanchez, Wen Yu and J.P. Perez, 1999, Nonlinear Adaptive Trajectory Tracking Using Dynamic Neural Networks, IEEE on NN, 10 (6), 1402-1411.

[18] Poznyak A.S., Martinez-Guerra R. and A.Osorio-Cordero, 2000, Robust High-gain Observer for Non-linear Closed-Loop Stochastic Systems, Mathematical Methods in Engineering Practice, 6, pp.31-60.

[19] Poznyak A.S., 2000, A New Version of the Strong Law of Large Number for Dependent Vector Processes with Decreasing Correlation, Proceedings of CDC-2000, Sydney, 2881-2882.

[20] Poznyak A. and L.Ljung, 2001, On-line IdentiÞ-cation and Adaptive Trajectory Tracking for Nonlinear Stochastic Continuous Time Systems Using Differen-tial Neural Networks, Automatica IFAC, especial issue (submitted).

[21] Rovithakis G.A. and M.A.Cristodoulou, 1994, Direct Adaptive regulation of Unknown Plants using DNN, IEEE on SMC, 24 (3), 400-412.

[22] Rovithakis G.A. and M.A.Cristodoulou, 2000, Adaptive control with Recurrent High-Order Neural Networks: Theory and Applications, Springer.

[23] Sandberg I.W., 1991, Approximation Theorems for Discrete Time Systems, IEEE on AC, 38, 564-566.

(9)

Avdelning, Institution Division, Department

Division of Automatic Control Department of Electrical Engineering

Datum Date 2007-06-25 Språk Language  Svenska/Swedish  Engelska/English   Rapporttyp Report category  Licentiatavhandling  Examensarbete  C-uppsats  D-uppsats  Övrig rapport  

URL för elektronisk version

http://www.control.isy.liu.se

ISBN  ISRN



Serietitel och serienummer

Title of series, numbering ISSN1400-3902

LiTH-ISY-R-2803

Titel

Title Robust high-gain DNN observer for nonlinear stochastic continuous time systems

Författare

Author D.A. Murano, A.S. Poznyak, Lennart Ljung Sammanfattning

Abstract

A class of nonlinear stochastic processes satysfying a Lipschitz-type strip condition and supplied by a linear output equation, is considered. Robust asymptotic (high-gain) state estimation for nonlinear stochastic processes via dierential neural networks is discussed. A new type learning law for the weight dynamics is suggested. By a stochastic Lyapunov-like analysis (with Ito formula implementation), the stability conditions for the state estimation error as well as for the neural nertwork weights are established. The upper bound for this error is derived. The numerical example, dealing with module-type nonlinearities, illustrates the eectiveness of the suggested approach.

Nyckelord

References

Related documents

Host cell membranes are equipped with a number of inhibitors to protect them against attack by complement, including membrane cofactor protein (MCP; CD46), complement receptor 1

As a preschool teacher in the Swedish preschool context, I have personally experienced challenges regarding the participation and to what extent are children a part in the

at lunch, we would have to also model as undesirable those states where it is lunch and the pills are not taken, although this is not as bad as not having taken the pills at all for

 Ett  par  av  informanterna  anser  dock  att   det  kan  vara  enklare  för  patienten  att  uttrycka  sig  genom  digitala  vårdbesök  och   menar  på  att

This study found that Rhythmia was feasible for mapping procedures during radiofrequency ablation of different complex arrhythmias with 100% technically success.. Lackermair

När det gäller regionala frågor har den turkiska regeringen vid flera tillfällen förklarat att den förblir fast besluten att stödja en övergripande lösning på Cypernfrågan

utan att informanterna avsäger sig det andra medborgarskapet och att nationalstaten definieras genom dess medborgare, istället för att medborgarna definieras genom nationalstaten

3 presents the fees and resulting improvement in social surplus when applying mean distance or time MSC 5. prices and optimal fees respectively, for both network wide and