• No results found

Discussion on ''On the Use of Minimal Parametrizations in Multivariable ARMAX Identification'' by R.P. Guidorzi

N/A
N/A
Protected

Academic year: 2021

Share "Discussion on ''On the Use of Minimal Parametrizations in Multivariable ARMAX Identification'' by R.P. Guidorzi"

Copied!
7
0
0

Loading.... (view fulltext now)

Full text

(1)

Discussion on “On the Use of Minimal

Parametrizations in Multivariable ARMAX

Identification” by R. P. Guidorzi

Tomas McKelvey

Department of Electrical Engineering

Link¨

oping University, S-581 83 Link¨

oping, Sweden

WWW:

http://www.control.isy.liu.se

Email:

tomas@isy.liu.se

March 1999

REGLERTEKNIK

AUTOMATIC CONTROL

LINKÖPING

Report no.: LiTH-ISY-R-2132

Published in European Journal of Control, vol 4, 1998, pp 93-98.

Technical reports from the Automatic Control group in Link¨oping are available by anonymous ftp at the address ftp.control.isy.liu.se. This report is contained in the pdf-file 2132.pdf.

(2)

Discussion on “On the Use of Minimal

Parametrizations in Multivariable ARMAX

Identification” by R. P. Guidorzi

Tomas McKelvey

Department of Electrical Engineering

Link¨

oping University

SE-581 83 Link¨

oping, Sweden

email: tomas@isy.liu.se

May 1998

This discussion section will focus on three issues:

1. Minimal parametrizations Is a minimal parametrization a necessity for a successful identification result?

2. The structural issue Does the choice of structural indices influence the identification result?

3. Gauss-Newton optimization How should the GN-search direction be calculated in a numerically stable way?

Minimal Parametrizations

System identification fundamentally deals with selecting a linear system based on some metric which depends on the data. In the paper the prediction error method is employed which corresponds to minimizing the sum of the square of the prediction errors. In order to execute this selection a model structure and parametrization needs to be defined which forms a mappingM from the space of real parameters Rl to the space of prediction models Σ. If this mapping M is injective the parametrization is known as minimal, i.e. a prediction model corresponds to at most one point in the parameter space. A fundamental result states that for multi output ARMAX systems there exists no injective mapping M which covers all linear models of a given McMillan degree. Hence a set of parametrizations is needed, Mi, i = 1, . . . . In the paper one such set is presented which is based on polynomial input output models. In its original form it is not suitable for predictions but a clever reformulation, see section 2.1, enables a minimal ARMAX parametrization of a predictor.

Is it necessary to use a minimal parametrization? Consider the state-space representations, see equation (1), and let all matrix elements inA, B, C, K rep-resent the parametrization of the linear model [4]. Trivially this parametrization is a surjective mapM between the parameter space Rl and the space of linear

(3)

models Σ, i.e. every linear system of McMillan degreen has infinitely many rep-resentations in the parameter space. We call this an overparametrized model structure. It is important to notice that the overparametrized model structure and the set of minimal ARMAX parametrizations describes the same set of lin-ear systems (1). Hence the selection of models from this common set by some metric does not depend on the parametrization used, minimal or not. The iden-tification result will, if we disregard numerical issues, be identical. Also notice that from a stochastic point of view the statistical properties of the identified model (such as the variance of the transfer function) is also invariant with re-spect to the parametrization used [7]. The advantage of an overparametrization over a minimal one is that it is sufficient to consider one parametrization as well as the possibility to use a state-space basis which is well conditioned, e.g. balanced forms. An apparent negative property of the overparametrized mod-els is the increased parameter dimension. However this can be reduced by a tridiagonal parametrization [5] or dynamically, during the optimization, form a minimal parametrization [6].

The Structural Issue

Each minimal ARMAX parametrization Mi corresponding to a certain struc-tural indexν is generic, i.e. the set of systems which cannot be described by the parametrization is of measure zero. Hence, given an arbitrary system, any of the ARMAX parametrizationsMiwill, w.p. 1, exactly be able to model it, [3]. This fact can lead you to the conclusion, which the author does, that the structural issue is only a theoretical one without practical identification implications. To show the converse first the results of a multivariable identification problem is presented followed by an analysis of the underlying reason which leads us to the third and last issue.

Identification of a pendulum system The problem is to identify an output-error model of an educational pendulum system from measured data using PEM. The system consists of a pendulum attached to a moving cart driven by a DC-motor. The input is the Voltage to the DC-motor and the 4 measurements are, cart position and velocity, pendulum angle and velocity. In figure 1 the criterion function V (θ) is plotted as a function of the number of Gauss-Newton itera-tions for two different parametrizaitera-tions, the observable state-space structure [3, p. 119] which is a minimal parametrization and a full state-space parametriza-tion [6]. The same Gauss-Newton optimizaparametriza-tion algorithm has been used for both parametrizations. Clearly there is a fundamental difference in the speed of convergence which is due to the type of parametrization used.

Gauss-Newton optimization

The quadratic criterion function to be minimized, see equations (26-31) of the paper, can be written as

V (θ) = e◦(θ)Te(θ)

where e◦(θ) is a Nm−dimensional vector of all prediction errors. Denote by −Hψ(θ) the Jacobian matrix of dimension Nm × l associated with e◦(θ). The

(4)

0 5 10 15 20 25 30 101 102 103 104 105 Criterion function Iteration # V( θ ) Full param. Minimal param.

Figure 1: Identification results during Gauss-Newton search. The graph shows de evolution of the criterion functionV (θ) during the minimization.

Gauss-Newton search direction γ(θ) can be formulated as the solution to the following Least-Squares (LS) problem

γ(θ) = arg min x ke

(θ) − H ψ(θ)xk2 and the parameter update is

θk+1=θk+δkγ(θk)

where δk is adjusted to ensure convergence (damped Gauss-Newton). The Gauss-Newton direction kan be determined by straight forward solution of the quadratic problem via the normal equations, c.f. equation (33) in the paper

γ(θ) =(θ)T(θ)−1HT

ψ(θ)e(θ)

Since the Gauss-Newton step is the solution to a Least-Squares problem its relative accuracy is proportional to the square of the condition number of(θ) [2]. In figure 2 the development of the condition number is depicted for the pendulum identification example. Clearly the minimal parametrization has a higher condition number. The relative sensitivity of the GN-direction will be higher and the search direction becomes less accurate. This explains the slow convergence.

Slow convergence is thus correlated with an ill-conditioned(θ) matrix. One possibility to improve conditioning is to use a pseudo-inverse [2] of when solving for the GN-direction or to add a scaled identity matrix to Hψ (Levenberg-Marquardt) [1]. When the pseudo-inverse is employed the singular

(5)

0 5 10 15 20 25 30

104

105

106

107

Condition number of Psi

Iteration #

Condition number

Full param. Minimal param.

Figure 2: Identification results during Gauss-Newton search. The graph illus-trates the development of the condition number of Hψ, the Jacobian matrix during the optimization.

values of matrix is determined an all less thanζ¯σ are set to zero, where ¯σ is the largest singular value of and ζ is a tunable constant. When pseudo-inverse technique is employed with the minimal parametrization and a choice of ζ = 10−4, we get the results shown in figures 3 and 4. An improved convergence to a point near the model obtained with the full parametrization can be noticed. The condition number of (the unmodified)Hψ has also greatly improved which indicates that the parameter path towards convergence has been shifted towards a path whereHψ is better conditioned.

Summary

When performing practical identification the structural issue often cannot be neglected since the conditioning of the parameterization influence the conver-gence of the optimization. Solutions are to carfully monitor the conditioning of the associated LS problem or to use alternative parametrizations (minimal or non-minimal) which give better conditioning.

References

[1] J. E. Dennis and R. B. Schnabel. Numerical Methods for Unconstrained Optimization and Nonlinear Equations. Prentice-Hall, Englewood Cliffs, New Jersey, 1983.

(6)

0 5 10 15 20 25 30 101 102 103 104 105 Criterion function Iteration # V( θ )

Minimal param. with mod GN dir. Minimal param.

Figure 3: Identification results during Gauss-Newton search using modified and non-modified GN-direction. The graph shows de evolution of the criterion func-tionV (θ) during the minimization.

0 5 10 15 20 25 30

104

105

106

107

Condition number of Psi

Iteration #

Condition number

Minimal param. with mod GN dir. Minimal param.

Figure 4: Identification results during Gauss-Newton search using modified and non-modified GN-direction. The graph illustrates the development of the con-dition number of, the Jacobian matrix during the optimization.

(7)

[2] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press, Baltimore, Maryland, second edition, 1989.

[3] L. Ljung. System Identification: Theory for the User. Prentice-Hall, Engle-wood Cliffs, New Jersey, 1987.

[4] T. McKelvey. Fully parametrized state-space models in system identification. In Proc. of the 10th IFAC Symposium on System Identification, volume 2, pages 373–378, Copenhagen, Denmark, July 1994.

[5] T. McKelvey and A. Helmersson. State-space parametrizations of multi-variable linear systems using tridiagonal matrix forms. In Proc. 35th IEEE Conference on Decision and Control, pages 3654–3659, Kobe, Japan, De-cember 1996.

[6] T. McKelvey and A. Helmersson. System identification using an over-parametrized model class - improving the optimization algorithm. In Proc. 36th IEEE Conference on Decision and Control, pages 2984–2989, San Diego, California, USA, December 1997.

[7] R. Pintelon, J. Schoukens, T. McKelvey, and Y. Rolain. Minimum variance bounds for overparameterized models. IEEE Trans. on Automatic Control, 41(5):719–720, May 1996.

References

Related documents

Flertalet av de inhämtade artiklarna till denna studie hävdar att enkelrum visade sig vara en stor fördel för patienternas välbefinnande under deras sjukhusvistelse men också att de

Han menar också att arbete med mönster i matematik är viktigt för att elever ska kunna lösa problem samt utveckla förståelse för begrepp och relationer i

Prospektiv studie 399 patienter med 1263 implantat 6 månader Från de behandlade patienterna 27 implantat misslyckades av de 1263 implantat Nej så länge diabetes är

Utifrån resultat i studien är slutsatsen att barnen och förskollärarna använder sig till stor del av naturmaterial för att arbeta med olika matematiska begrepp i utemiljön..

Genom att studera det lokala skolutvecklingsprojektet, Den Tionde Dagen, är avsikten att försöka bidra med kunskap inom värdediskursen vars kontext kommer att bestå av lokal

Eftersom vi använder samma system för teckentolkningen som våra lärare har i studion blir det ”strömlinjeformat” då vi bara behöver hantera ett system och får den

Vid analys av knäckning användes Ansys toolbox ”Linear Buckling” och beräkningar genomfördes endast för den slutgiltiga modellen.. Svaret som fås från Ansys är

Träets egenskaper kan förbättras genom impregnering...