• No results found

S-58183 Linkoping, Sweden svante@isy.liu.se, mino@isy.liu.se

N/A
N/A
Protected

Academic year: 2021

Share "S-58183 Linkoping, Sweden svante@isy.liu.se, mino@isy.liu.se"

Copied!
16
0
0

Loading.... (view fulltext now)

Full text

(1)

Svante Gunnarsson Mikael Norrlof

Department of Electrical Engineering, Linkoping University,

S-58183 Linkoping, Sweden svante@isy.liu.se, mino@isy.liu.se

January 1997

Abstract

An introduction to iterative learning control (ILC) is given. The basic prin- ciple behind ILC in both open loop and closed loop problems is explained. A general class of algorithms for updating of the ILC input signal is presented and the choice of lters in the update algorithm is discussed with respect to convergence, robustness and disturbance sensitivity.

1

(2)

1 Introduction

Iterative learning control (ILC) has been an active research area for more than a decade, and the paper of Arimoto and co-authors 1] is often referred to as the main source of inspiration for research in this area. Several papers have been written during this period, and only a minor part of these will be referred to in this report. Further references can be found in e.g. 2], 3] and 4].

The main idea in iterative learning control is to utilize the situation that the system to be controlled will carry out the same operation several times. It will then be possible to gradually improve the performance of the control system by using the results from one operation when choosing the input signal for the next operation.

The main area of application for ILC is control of industrial robots, and di erent types of real or simulated robots are used as test examples in a large number of publications.

A general discussion of the use of ILC in robotics is given in 5].

A major issue when applying ILC is convergence, i.e. that the iterative update of the input signal converges to a signal giving good performance. The convergence aspects were discussed already in 1] where some convergence criteria were derived. These criteria were however restrictive and a great e ort has been spent on nding more realistic conditions. A number of these results are found in 6], 7], 8], 9] and 10].

The aim of this report is to give a brief introduction to the area of iterative learning control. Even though industrial robots, which are the most important applications of ILC, are highly nonlinear we shall in this report restrict ourselves to linear systems.

The report is organized as follows. In Section 2 we start by giving a short introduction to the control problem in general. In Sections 3 and 4 we then present the iterative learning control concept in open loop and closed loop control systems. These sections describe the ILC idea in general and present some of the basic convergence criteria.

Section 5 then contains a more general discussion of di erent approaches to the up- dating of the learning control signal and how these approaches a ect the convergence conditions. In Section 6 we then investigate the e ects of load and measurement distur- bances, while we in Section 7 present some model and optimization based approaches to learning control. Finally in Section 8 we briey discuss the e ects of friction and mechanical exibilities, which are two areas of big practical importance in robotics.

Some conclusions are given in Section 9.

2

(3)

2 Control

To support the discussion of iterative learning control and introduce some useful equa- tions we shall here give a brief introduction to the control problem in general.

Let us therefore consider a scalar linear system with input

U

(

s

) and output

Y

(

s

) described by

Y

(

s

) =

G

(

s

)

U

(

s

) +

V

(

s

) (1) where

G

(

s

) denotes the transfer function of the system and

V

(

s

) is a load disturbance.

We shall mainly consider continuous-time systems, but since in reality the control is carried out using sampled data a discrete-time treatment is more suitable. The main part of the equations are however applicable also for discrete-time problems.

The aim is to choose the input signal

U

(

s

) such that the output signal follows a desired output signal

YD

(

s

) as closely as possible. One approach to this problem is to apply feed-forward (open loop) control and generate the input signal as

U

(

s

) =

Ff

(

s

)

YD

(

s

) (2) where

Ff

(

s

) is a transfer function. Introducing the error signal

E

(

s

) =

YD

(

s

)

;Y

(

s

) (3) equation (2) implies that the error is given by

E

(

s

) = (1

;G

(

s

)

Ff

(

s

))

YD

(

s

)

;V

(

s

) (4) In equation (4) we see that we ideally would like to choose

Ff

(

s

) as the inverse of

G

(

s

), since the rst term on the right hand side then becomes zero. In order to obtain this however, we need an exact description of the system, and in reality

Ff

(

s

) has to be an approximate model of the system. Furthermore, since typically

G

(

s

) is a strictly proper transfer function, the input signal will contain derivatives of the reference signal.

Exact di erentiation is not possible to obtain in practice, but in, for example, some robot applications both the reference position and the reference velocity are specied.

In such a case it is possible to obtain a feed-forward input signal using the derivative of the reference signal. A third limitation of the feed-forward approach is that

Ff

(

s

) has to be stable, which means that we cannot apply feed-forward to a system having zeros in the right half plane. Finally it should be noticed that the second term on the right hand side of equation (4), i.e. the load disturbance, is not a ected by feed-forward control.

A natural extension of the open loop (feed-forward) control above is to combine it with feed-back according to

U

(

s

) =

F

(

s

)(

YD

(

s

)

;Y

(

s

)) +

Ff

(

s

)

YD

(

s

) (5)

3

(4)

where

F

(

s

) is the transfer function of the feed-back regulator. Equation (5) implies that the error signal dened in (3) is given by

E

(

s

) =

S

(

s

)(1

;G

(

s

)

Ff

(

s

))

YD

(

s

)

;S

(

s

)

V

(

s

) (6) where

S

(

s

) = 1

1 +

F

(

s

)

G

(

s

) (7)

is the sensitivity function of the closed loop system. Comparing equations (4) and (6) we see that the error is reduced by the feedback in the frequency range where

jS

(

i!

)

j<

1 (8)

Typically

j S

(

i!

)

j

is less than one in the low frequency range, while it tends to one for high frequencies. The bandwidth over which the sensitivity function can be made small is limited by factors like available input signal energy, measurement noise and model uncertainties.

In a robot application it is sometimes of interest to use the control signal from the feed-back regulator as error signal. We thus consider

E

(

s

) =

F

(

s

)(

YD

(

s

)

;Y

(

s

)) (9) Since

F

(

s

) in a robot application typically is a PD-regulator the error signal is a combination of the position and velocity error, appropriately scaled into a torque signal.

This denition gives

E

(

s

) =

GC

(

s

)((

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;F

(

s

)

S

(

s

)

V

(

s

) (10) where

G

C

(

s

) =

F

(

s

)

G

(

s

)

1 +

F

(

s

)

G

(

s

) (11)

is the transfer function of the closed loop system.

3 Open Loop Iterative Learning Control

The main idea in ILC is that the system carries out some given operation several times.

It is furthermore assumed that each operation is carried out over a nite time interval.

All signals are hence dened over a nite time interval, and in the discrete time case

e

k

(

t

) will denote the error signal at time instant

t

and iteration

k

. The variable

ek

will then be a vector of values representing the error at the sampling points.

4

(5)

To illustrate the basic idea we shall start by considering ILC in an open loop context, and for simplicity we consider a servo problem and neglect the load disturbance.

At iteration

k

, the output

Yk

(

s

) of the system is generated as

Y

k

(

s

) =

G

(

s

)

Uk

(

s

) (12) which gives the error signal

E

k

(

s

) =

YD

(

s

)

;Yk

(

s

) (13) The idea is then to use the error signal

Ek

(

s

) to improve the system performance by computing a new input signal

Uk+1

(

s

) using

Ek

(

s

). Several methods for carrying out the updating of the input signal have been discussed in the literature. We shall here for simplicity consider the update equation

U

k+1

(

s

) =

Uk

(

s

) +

H

(

s

)

Ek

(

s

) (14) where

H

(

s

) is a lter. Di erent choices of the lter

H

(

s

) have been discussed, and the simplest form is of course to use a constant, i.e.

H

(

s

) =



(15)

In 1] the derivative of the error signal is used, which means

H

(

s

) =

s

(16)

A combination of these two alternatives gives an ILC algorithm of, so called, PD-type, which corresponds to

H

(

s

) =

1

+

2s

(17)

It is also important to note how the error signal is dened. Dening

ek

(

t

) as the velocity error in a position control problem and using

H

(

s

) =



is equivalent to dening

ek

(

t

) as the position error and using

H

(

s

) =

s

. The rst alternative is used in e.g. 11]. A more general discussion of the choice of updating algorithms will be given in Section 5 below.

It is now of interest to investigate what happens with the error signal as the iterations continue. We therefore consider

E

k+1

(

s

) =

YD

(

s

)

;Yk+1

(

s

) =

YD

(

s

)

;G

(

s

)

Uk+1

(

s

) (18) which using equation (14) gives

E

k+1

(

s

) =

YD

(

s

)

;G

(

s

)

Uk

(

s

)

;G

(

s

)

H

(

s

)

Ek

(

s

) (19)

=

Ek

(

s

)

;G

(

s

)

H

(

s

)

Ek

(

s

) = (1

;G

(

s

)

H

(

s

))

Ek

(

s

)

5

(6)

In the continuous-time open loop case we see that provided

j

1

;H

(

i!

)

G

(

i!

)

j<

1

8 !

(20) the error will tend to zero, and hence to output signal will follow the desired one exactly. The condition in equation (20) has the interpretation that the Nyquist curve

H

(

i!

)

G

(

i!

) has to be inside a circle of radius one with the center in one. This circle is in the literature denoted learning circle. When

H

(

s

) is just a constant this condition becomes very restrictive since it requires that the argument of

G

(

i!

) never goes below

;

90



. For real systems the condition is typically violated in the high frequency range which implies that high frequency components of the error signal increases as the iterations proceed.

4 Closed Loop Iterative Learning Control

We shall in this report mainly consider ILC in combination with conventional feed-back and feed-forward control, as discussed in 5]. The structure of the problem is described by Figure 1. The basic idea is also here that the system carries out the same movement repeatedly, and a correction signal 

uk

is updated after each iteration.

+ +

+ -

F G

F

f

y

D

y

k u

k



uk

e

k

Figure 1: A feed-forward and feed-back control system with iterative learning control Di erent types of feed-back and feed-forward can be covered by this structure, and the most common case is that the feed-back consists of a PD-regulator. Alternative control strategies like, for example, state space methods, have also been used, and one example is given in 12].

According to the block diagram the input signal is now given by

U

k

(

s

) =

Ff

(

s

)

YD

(

s

) +

F

(

s

)(

YD

(

s

)

;Yk

(

s

)) + 

Uk

(

s

) (21)

6

(7)

which implies that the output of the closed loop system is given by

Y

k

(

s

) = 1

1 +

F

(

s

)

G

(

s

)(

F

(

s

)

G

(

s

)

YD

(

s

) +

Ff

(

s

)

G

(

s

)

YD

(

s

) +

G

(

s

)

Uk

(

s

)) (22) Using the output of the feed-back regulator as error signal, i.e. equation (9), the error signal becomes

E

k

(

s

) =

GC

(

s

)((

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;



Uk

(

s

)) (23) Initially we shall consider the same updating algorithm as in the open loop case, i.e.

U

k+1

(

s

) =

Uk

(

s

) +

H

(

s

)

Ek

(

s

) (24) Using equation (23) we get

E

k+1

(

s

) =

GC

(

s

)(

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;GC

(

s

)

Uk+1

(

s

) (25) which, inserting (24), gives

E

k+1

(

s

) =

GC

(

s

)(

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;GC

(

s

)

Uk

(

s

) (26)

; G

C

(

s

)

H

(

s

)

Ek

(

s

) =

Ek

(

s

)

;GC

(

s

)

H

(

s

)

Ek

(

s

) i.e.

E

k+1

(

s

) = (1

;H

(

s

)

GC

(

s

))

Ek

(

s

) (27) In analogy to what was nd in the open loop case we see that provided that

j

1

;H

(

i!

)

GC

(

s

)(

i!

)

j<

1

8!

(28) the error will tend to zero. The condition is the same as for the open loop problem with the di erence that the open loop transfer function

G

(

s

) has been replaced by the closed loop transfer function.

5 Update Equations

It is clear that the properties of the ILC algorithm will depend on how the update of the control signal is carried out, and in this section we shall discuss some possible approaches.

Considering only linear operations a general formulation of the updating of the correc- tion signal can be expressed in the frequency domain as



Uk+1

(

s

) =

Xk

j=0



H

j

(

s

)

Ej

(

s

) (29)

7

(8)

where 

Hj

(

s

)

 j

= 0

:::k

are linear lters.

Let us initially for simplicity assume that the lters 

Hj

(

s

) are just constants, i.e.



H

j

(

s

) = 

hj 8 j

(30)

This gives, in the time domain,



uk+1

(

t

) =

Xk

j=0



h

j e

j

(

t

) (31)

which means that the input signal at time

t

in iteration

k

+ 1 will be a weighted sum of the errors at time

t

in the previous iterations. Keeping

t

xed and considering

k

as the time index the computation of the new correction signal 

uk+1

is a ltering of the error signal

ek

using a lter with impulse response coecients 

hj

. Provided that the coecients 

hj

decay exponentially equation (31) can be rewritten in a recursive formulation as a conventional di erence equation having

ek

as input and 

uk

as output.

Applying this way of thinking to the update equation



uk+1

= 

uk

+

ek

(32)

we nd that it can be seen as a ltering of

ek

through a discrete time lter

L

(

z

) =



z;

1 (33)

i.e. a pure integrator. By including also

ek;1

in the update equation we obtain a, so called, two-step algorithm



uk+1

= 

uk

+

1ek

+

2ek;1

(34) which corresponds to the lter

L

(

z

) =

1

+

2z;1

z;

1 (35)

which is of PI-type. This way of describing the update equation is thoroughly discussed in 8], where two dimensional transforms are used to describe the involved signals both in time and iteration number.

For convenience we shall here however consider recursive update equations on the form



Uk+1

(

s

) =

H1

(

s

)

Uk

(

s

) +

H2

(

s

)

Ek

(

s

) (36) where

H1

(

s

) and

H2

(

s

) are linear lters. Since the ltering in equation (36) is carried out o -line we can allow

H1

and

H2

to be non-causal.

8

(9)

We shall now investigate how the error signal behaves when the update equation (36) is applied. Let us recall equation (23)

E

k

(

s

) =

GC

(

s

)((

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;



Uk

(

s

)) (37) and introduce the signal

E0

(

s

) dened by

E

0

(

s

) =

GC

(

s

)((

G;1

(

s

)

;Ff

(

s

))

YD

(

s

) (38) which is the error signal obtained in the rst iteration when no correction signal is added, i.e. 

U0

(

s

)



0. Using equation (38) we get

E

k+1

(

s

) =

E0

(

s

)

;GC

(

s

)

Uk+1

(

s

) (39) and inserting equation (36) we obtain

E

k+1

(

s

) =

E0

(

s

)

;GC

(

s

)

H1

(

s

)

Uk;GC

(

s

)

H2

(

s

)

Ek

(

s

) (40) By adding and subtracting

H1

(

s

)

E0

(

s

) on the right hand side of equation (40) we get

E

k+1

(

s

) = (1

;H1

(

s

))

E0

(

s

) + (

H1

(

s

)

;H2

(

s

)

GC

(

s

))

Ek

(

s

) (41) This result can be compared with the error equation in, for example, 13] where the analogous equation for the open loop control case are derived. In our case, where the closed loop case is considered, the driving signal in the update equation is

E0

(

s

), i.e.

the error obtained without the correction signal 

U

(

s

).

The convergence condition now becomes

jH

1

(

i!

)

;H2

(

i!

)

GC

(

i!

)

j<

1

8 !

(42) A di erent representation, inspired by 14], of the lters in the update equation is given

by 

Uk+1

(

s

) = 1

1 +

W

(

s

)(

Uk

(

s

) +

HEk

(

s

)) (43) which means

H

1

(

s

) = 1

1 +

W

(

s

)

H2

(

s

) = 1 +

HW

(

s

) (44) The convergence criterion then becomes

j

1

;H

(

i!

)

GC

(

i!

)

j<j

1 +

W

(

i!

)

j 8!

(45) As shown in 14], where a constant

H

(

s

) is used, the lter

W

(

s

) can be used to extend the region where the Nyquist curve

H

(

i!

)

GC

(

i!

) has to be located in order to get convergence.

9

(10)

Equation (36) covers the majority of the algorithms that have been considered in the literature, and the lters

H1

(

s

) and

H2

(

s

) are used in di erent ways by di erent authors. The algorithm considered in the original paper 1] corresponds to

H1

(

s

)



1 and

H2

(

s

) =

 s

, where



is a scalar. One of the rst references where

H1

(

s

)

6

= 1 is used appears to be 14], where it is shown how

H1

(

s

) can be used to obtain less restrictive convergence conditions. In 15] the case

H1

(

s

) =



, where

<

1 is a scalar, is studied. In 13] iterative learning control of a exible robot is studied, and there the authors also use two lters in the update equation.

More or less systematic methods for design of the lters

H1

(

s

) and

H2

(

s

) have also been presented. A tempting alternative is to choose

H

2

(

s

) =

GC

(

s

)

;1

(46) which would yield convergence in one step. This however requires that a perfect model of the closed loop system is available and that this model has a stable inverse. Fur- thermore this choice would result in a lter with very high gain for high frequencies.

A more realistic alternative is to let

H2

(

s

) be equal to the inverse of a model of the closed loop system only in the low frequence range. Such an approach is discussed in, for example, 13]. An interesting method for choosing appropriate lters in the update equation is presented in 16] where methods from design of robust controllers are ap- plied. The lters are designed to give a convergent ILC algorithm despite uncertainties in the process model.

Provided that the updating of the correction signal converges it is of interest to study the asymptotic error signal. By simply replacing the error signal in equation (41) by



E

(

s

) we obtain



E

(

s

) = 1

;H1

(

s

)

1

;H1

(

s

) +

GC

(

s

)

H2

(

s

)

E0

(

s

) (47) A Bode plot of this transfer function shows the benets of applying ILC. We see that by using

H1

(

s

)

6

= 1 we are not able to eliminate the error completely, and this is the price that has to be paid for the improved convergence properties.

6 Disturbances

While we so far mainly have considered servo problems and neglected both load and measurement disturbances, we shall now investigate how these e ects inuence the properties of the ILC algorithm and the performance of the control system.

Let us, at iteration

k

, consider the system

Y

k

(

s

) =

G

(

s

)(

Uk

(

s

) +

Vk

(

s

)) (48)

10

(11)

where

Vk

(

s

) denotes a load disturbance that now acts on the input side of the system.

The system is controlled using the input signal

U

k

(

s

) =

Ff

(

s

)

YD

(

s

) +

F

(

s

)(

YD

(

s

)

;

(

Yk

(

s

) +

Nk

(

s

))) + 

Uk

(

s

) (49) where

Nk

(

s

) is a measurement disturbance. Inserting (49) into (48) gives the closed loop system

Y

k

(

s

) = 1

1 +

F

(

s

)

G

(

s

)(

G

(

s

)(

F

(

s

) +

Ff

(

s

))

YD

(

s

) (50) +

G

(

s

)

Uk

(

s

) +

G

(

s

)

Vk

(

s

)

;F

(

s

)

G

(

s

)

Nk

(

s

))

Considering as before the error signal

E

k

(

s

) =

F

(

s

)(

YD

(

s

)

;Yk

(

s

)) (51) we obtain

E

k

(

s

) =

GC

(

s

)((

G;1

(

s

)

;Ff

(

s

))

YD

(

s

)

;



Uk

(

s

)

;Vk

(

s

) +

F

(

s

)

Nk

(

s

)) (52) Let us recall equation (38)

E

0

(

s

) =

GC

(

s

)(

G;1

(

s

)

;Ff

(

s

))

YD

(

s

) (53) and equation (36)



Uk+1

(

s

) =

H1

(

s

)

Uk

(

s

) +

H2

(

s

)

Ek

(

s

) (54) This gives

E

k+1

(

s

) =

E0

(

s

)

;GC

(

s

)

H1

(

s

)

Uk

(

s

)

;GC

(

s

)

H2

(

s

)

Ek

(

s

) (55)

; G

C

(

s

)

Vk+1

(

s

) +

F

(

s

)

GC

(

s

)

Nk+1

(

s

)

and by adding and subtracting relevant terms on the right hand side we get

E

k+1

(

s

) = (1

;H1

(

s

))

E0

(

s

) +

H1

(

s

)(

E0

(

s

)

;GC

(

s

)

Uk

(

s

)

;GC

(

s

)

Vk

(

s

) (56) +

F

(

s

)

GC

(

s

)

Nk

(

s

))

;GC

(

s

)

H2

(

s

)

Ek

(

s

) +

H1

(

s

)

GC

(

s

)

Vk

(

s

)

; G

C

(

s

)

Vk+1

(

s

)

;H1

(

s

)

F

(

s

)

GC

(

s

)

Nk

(

s

) +

F

(

s

)

GC

(

s

)

Nk+1

(

s

) which implies the following error update equation

E

k+1

(

s

) = (1

;H1

(

s

))

E0

(

s

) + (

H1

(

s

)

;GC

(

s

)

H2

(

s

))

Ek

(

s

) +

GC

(

s

)(

H1

(

s

)

Vk

(

s

)

; V

k+1

(

s

)) +

F

(

s

)

GC

(

s

)(

Nk+1

(

s

)

;H1

(

s

)

Nk

(

s

)) (57) A similar equation is presented in 13] for the open loop case and for load disturbances only. A number of observations can be made using equation (57). Let us rst consider the case

H1

(

s

) = 1, which implies the update equation

E

k+1

(

s

) = (1

;GC

(

s

)

H2

(

s

))

Ek

(

s

) +

GC

(

s

)(

Vk

(

s

)

;Vk+1

(

s

)) (58) +

F

(

s

)

GC

(

s

)(

Nk+1

(

s

)

;Nk

(

s

))

11

(12)

The disturbances contribute to the error equation by their di erences between the iterations. If a disturbance is of repetitive nature in the sense that the disturbance signals

dk

(

t

) =

dk+1

(

t

) and

nk

(

t

) =

nk+1

(

t

) for all

k

the contribution to the error di erence equation is zero. This assumption is more likely for the load disturbance where for example load disturbances due to gravitational forces can be expected to be rather similar during di erent iterations. Measurement disturbances, on the other hand, are more likely to be of random character which means that

nk+1

(

t

)

6

=

nk

(

t

) in general, and there will hence always be a driving term on the right hand side of equation (59) that prevents

Ek

(

s

) from tending to zero.

Let us then return to the situation with

H1

(

s

)

6

= 1, neglect measurement disturbances and assume that

v

k

(

t

) =

v

(

t

)

8 k

(59)

This corresponds to the error di erence equation

E

k+1

(

s

) = (1

;H1

(

s

))

E0

(

s

)+(

H1

(

s

)

;GC

(

s

)

H2

(

s

))

Ek

(

s

)

;GC

(

s

)

V

(

s

)(1

;H1

(

s

)) (60) The load disturbance will then cause a non-zero driving term on the right hand side similar to the term caused by the initial error

E0

(

s

). Both terms will then contribute to the asymptotic error resulting when

k

tends to innity, which is given by



E

(

s

) = 1

;H1

(

s

)

1

;H1

(

s

) +

GC

(

s

)

H2

(

s

)

E0

(

s

)

; GC

(

s

)(1

;H1

(

s

))

1

;H1

(

s

) +

GC

(

s

)

H2

(

s

)

V

(

s

) (61)

7 Model and Optimization Based Methods

In early studies of the topic the hope was that the ILC method should o er a completely model free control method. By just carrying out repeated movements of, for example, a robot it should be possible to determine a suitable input that minimizes the desired performance measure. The convergence results that have been derived have however shown that the property of the system (open or closed loop) itself plays an important role for the behavior of the ILC algorithm. In order to design such an algorithm properly it is hence necessary to have some a priori model of the system that is going to be controlled. This is not a particularly restrictive assumption since fairly accurate models often are available.

In 17] the updating of the learning control signal is carried out in the frequency domain using the DFT of the signals. A (local) model is identied in each iteration by forming the ETFE, i.e. the ratio between the DFT:s of the output and input signals. The inverse of the ETFE is then used in the update of the learning control signal. System identication is also used in 16] where an initial identication experiment is carried

12

(13)

out in order to obtain a model of the system to be controlled. In addition a bound on the modeling error is computed for later use in the design of lters in the update equation. In 18] and 19] another way of obtaining a model of the system i presented.

The model is then used in the updating of the input signal.

The choice of the lters

H1

(

s

) and

H2

(

s

) in the formula for updating of the control signal can also be seen as a step size selection in an iterative minimization procedure.

Let us therefore consider a discrete time problem where, during each iteration, all signals are dened in

N

sampling points. We therefore introduce the vectors

Y

k

= (

yk

(1)

:::yk

(

N

))

T

(62)

U

k

= (

uk

(1)

:::uk

(

N

))

T

(63)

E

k

= (

ek

(1)

:::ek

(

N

))

T

(64) and

Y

D

= (

yD

(1)

:::yD

(

N

))

T

(65) containing the output, input, error and reference signals at iteration

k

and time instants

t

= 1

:::N

. The relationship between input and output is now given by

Y

k

=

GUk

(66)

where

G

=

0

B

B

B

B

B

B

B

@ g

0

0 0

:::

0

g

1 g

0

0

:::

0 ... ... ... ...

g

N g

N;1

::: g

1 g

0 1

C

C

C

C

C

C

C

A

(67)

is a matrix dened by the impulse response coecients of the system

G

(

z

) =

X1

k=0 g

k z

;k

(68)

Consider now the criterion

J

=

EkTEk

+

UkTUk

(69) where

E

k

=

YD ;Yk

(70)

and



is a positive scalar.

We would now like to minimize

J

with respect to the input signal values in the vector

U

, and this is done by di erentiating

J

with respect to

U

and putting the gradient equal to zero. This yields

U

opt

= (

I

+

GTG

)

;1GTYD

(71)

13

(14)

By further imposing a condition on the size of the update step an iterative procedure for computing the input is obtained, and following 19] we get

U

k+1

=

Uk;

(

I

+

rI

+

GTG

)

;1

(

Uk

+

GTEk

) (72) where

r

is a parameter determined by the bound on the update step.

Another type of modeling problem, discussed in 20] and 21], is how to model the ILC input signal. In the standard formulation the signal 

Uk

is a table of numbers representing the input value at each sampling point. For di erent reasons it can be useful to try to represent the signal by fewer parameters. This can be done by using, for example, spline functions, as discussed in 21]. Another benet of this approach is that it, by a proper choice of functions, also can be possible to restrict the frequency content of the signals.

8 Friction and Flexibilities

Friction is a very important problem in robotics and other types of mechanical devices, and a lot of e ort has been spent on developing methods for reducing the negative e ects of friction. A majority of these methods are model based in the sense that they use some kind of mathematical model of the friction. By applying parameter estimation a friction compensation signal can be computed and added to the input signal generated by the control system. Iterative learning control however o ers non- parametric alternative where no explicit model of the friction is used. Such an approach is discussed in, for example, 22] and 23].

Another topic of big practical importance is mechanical exibilities. In the design of robots there is always a trade-o between sti ness and weight. A sti robot, which is easier to control with high precision, is typically heavier and more expensive. A challenge is therefor to design a control system such that also a exible robot can be used in tasks requiring high precision. Iterative learning control approaches to this problem have been discussed in, for example, 24], 13] and 21].

9 Conclusions

An introduction to the area of iterative learning control has been given. The basic principles behind the use of ILC in both open loop and closed loop control has been discussed. The choice of lters in the updating formula and the consequences for convergence and robustness have been treated.

14

(15)

References

1] S. Arimoto, S. Kawamura, and F. Miyazaki. \Bettering Operation of Robots by Learning". Journal of Robotic Systems, pages 123{140, 1984.

2] L. Hideg. Stability of Learning Control Systems. PhD thesis, Oakland University, Rochester, Michigan, 1992.

3] K.L. Moore, M. Daleh, and S.P. Battacharrya. \Iterative Learning Control: A Survey and New Results". Journal of Robotic Systems, 9:563{594, 1992.

4] R. Horowitz. \Learning Control of Robot Manipulators". ASME Journal of Dy- namic Systems, Measurement, and Control, 115:403{411, 1993.

5] J. Craig. Adaptive Control of Mechanical Manipulators. Addison-Wesley Publish- ing Company, 1988.

6] F. Padieu and R. Su. \An

H1

approach to Learning Control Systems". Interna- tional Journal of Adaptive Control and Signal Processing, 4:465{474, 1990.

7] G. Heinzinger, D. Fenwick, B. Paden, and F. Miyazaki. \Stability of Learning Control with Disturbances and Uncertain Initial Conditions". IEEE Trans. Au- tomatic Control, 37:110{114, 1992.

8] Y.J. Liang and D.P Looze. \Performance and Robustness Issues in Iterative Learn- ing Control". In Proc. 32nd CDC, pages 1990{1995, San Antonio, TX, 1993.

9] N. Amann, D.H. Owens, and E. Rogers. \Iterative Learning Control using Optimal Feedback and Feedforward Actions". Technical report, Report Number: 95/13, Centre for Systems and Control Engineering, University of Exeter, Exeter, United Kingdom, 1995.

10] N. Amann, D.H. Owens, and E. Rogers. \Iterative Learning Control for Discrete Time Systems with Exponential Rate of Convergence". Technical report, Report Number: 95/14, Centre for Systems and Control Engineering, University of Exeter, Exeter, United Kingdom, 1995.

11] S. Kawamura, F. Miyazaki, and S. Arimoto. \Realization of Robot Motion Based on a Learning Method". IEEE Trans. on Systems, Man and Cybernetics, 18:126{

134, 1988.

12] M. Togai and O. Yamano. \Analysis and Design of an Optimal Learning Control Scheme for Industrial Robots: A Discrete System Approach". In Proc. 24th CDC, pages 1399{1404, Ft. Lauderdale, Fl., 1985.

13] S. Panzieri and G. Ulivi. \Disturbance rejection of Iterative Learning Control Applied to Trajectory for a Flexible Manipulator". In Proc. ECC 1995, pages 2374{2379, Rome, Italy, 1995.

15

References

Related documents

Channell’s description of vagueness is based on the notion developed by Peirce (1902, quoted in Channell 1994: 7), in which he defines ‘intrinsic uncertainty’ as “not uncertain

46 Konkreta exempel skulle kunna vara främjandeinsatser för affärsänglar/affärsängelnätverk, skapa arenor där aktörer från utbuds- och efterfrågesidan kan mötas eller

För att uppskatta den totala effekten av reformerna måste dock hänsyn tas till såväl samt- liga priseffekter som sammansättningseffekter, till följd av ökad försäljningsandel

Från den teoretiska modellen vet vi att när det finns två budgivare på marknaden, och marknadsandelen för månadens vara ökar, så leder detta till lägre

The increasing availability of data and attention to services has increased the understanding of the contribution of services to innovation and productivity in

Generella styrmedel kan ha varit mindre verksamma än man har trott De generella styrmedlen, till skillnad från de specifika styrmedlen, har kommit att användas i större

I regleringsbrevet för 2014 uppdrog Regeringen åt Tillväxtanalys att ”föreslå mätmetoder och indikatorer som kan användas vid utvärdering av de samhällsekonomiska effekterna av

Industrial Emissions Directive, supplemented by horizontal legislation (e.g., Framework Directives on Waste and Water, Emissions Trading System, etc) and guidance on operating