• No results found

Local Models for Dynamic Processes in Image Sequences

N/A
N/A
Protected

Academic year: 2021

Share "Local Models for Dynamic Processes in Image Sequences"

Copied!
6
0
0

Loading.... (view fulltext now)

Full text

(1)

Local Models for Dynamic Processes in Image

Sequences

Hagen Spies1,2, Tobias Dierig2,3, and Christoph S. Garbe3 1

Computer Vision Laboratory

Dept. of Electrical Engineering, Link¨oping University 581 83 Link¨oping, Sweden

hspies@isy.liu.se 2

ICG-III: Phytosphere

Research Center J¨ulich, 52425 J¨ulich, Germany h.spies@fz-juelich.de 3

Interdisciplinary Center for Scientific Computing, University of Heidelberg, INF 368, 69120 Heidelberg, Germany,

{Tobias.Dierig,Christoph.Garbe}@iwr.uni-heidelberg.de

Abstract. We present a computational framework that extends classical image velocity estimation to include more general parameters of dynamic brightness changes. The introduced method allows for an extraction of these parameters, ranging from models of linear illumination changes over diffusion and decay constants to expansion rates. We illustrate the benefit of such an extension on a real image sequence with illumination changes. We also introduce a new depth estimation technique termed depth from diffusion and apply it to some real ex-amples.

1

Introduction

Classical image motion analysis relies on the assumption that all intensity changes are due to motion. This implies that the total derivative of the intensity g with respect to time vanishes, which is the brightness change constraint equation [Horn and Schunk, 1981]:

dg

dt = gxu + gyv + gt= 0 . (1)

Here we denote partial derivatives using subscripts. This concept is illustrated in Fig. 1a where the motion is along isobrightness contours. Clearly this assumption does not hold in real world situations where we encounter changes in image brightness due to variations in surface orientation or lighting conditions. An example where the intensity function also undergoes a diffusion is shown in Fig. 1b. Here the isobrightness lines will not correspond to the movement any more. The resulting velocity field computed with and without incorporation of this additional brightness change for an example image sequence is shown in Fig. 1c-f. Interestingly humans have little difficulty in perceiving the correct movement in this case.

(2)

a

f

x

g

t

t

0 0

t

x

0

x

0

x

t

b x g t f t t 0 0t x0 x0x c d e f

Fig. 1: Illustration of the brightness change equation: a with conserved brightness, b with

in-tensity changing due to diffusion. c first and d last frame of a moving Gaussian bell undergoing diffusion. e optical flow assuming conserved brightness and f estimated velocity using the ex-tended model.

To account for such variations the used conservation law has to be extended. To-wards this end the use of multiplier and offset fields have been suggested [Negah-daripour, 1998]. Below we give a more general extension that replaces (1) by a linear partial differential equation [Haußecker et al., 1999; Haußecker and Fleet, 2001]. The novel contributions of this paper are quantitative results for a sequence with motion and illumination changes and the introduction of a new depth from X algorithm.

2

Models for Dynamic Processes

To describe more general dynamic models we allow for the intensity to vary along the trajectories we are estimating. We assume that this variation can be expressed in terms of a model function f which may depend on the intensity, time and a set of model parameters a. Then the brightness change equation becomes:

[gxgy]v + gt= f (g, t, a) . (2) Here v is the geometric velocity, for instance described by an affine motion (v = t +

Ax). The concept is very general in the sense that the parameters of any dynamic

process that can be modeled by a linear partial differential equation can be quantified. Since most physical, chemical, and biological processes can be described by equations of this type, it covers many applications.

(3)

3

Total Least Squares Estimation

As all the observation data in (2) is suspect to noise it is appropriate to use a total least squares (TLS) method [Van Huffel and Vandewalle, 1991]. This method is in contrast to ordinary least squares estimation where the noise is assumed to be confined to the tem-poral domain. It has recently been pointed out that such a TLS model can successfully describe some observations made for the mammalian visual system [Langley, 2002].

To enable a total least squares solution we note that (2) can be written as the scalar product of a known data vector d with an unknown parameter vector p: dTp = 0. This

equation poses only one constraint in the unknown parameters, thus further assumptions are needed in order to solve for the parameter field. A common smoothness requirement assumes constant parameters in a small local neighborhood of N pixel. A weighted total least squares estimate is then given by the eigenvector ˆento the smallest eigenvalue λn of the so called structure tensor [Haußecker et al., 1999]:

J = B ∗ ( d dT) , (3)

where B is an integration kernel and ∗ denotes convolution. A good choice for B is a binomial filter as it is both symmetric and leads to a decreasing influence with distance from the considered pixel.

The above estimation is only optimal if the entries in the data vector d are uncorre-lated zero mean random variables with the same noise variance [M¨uhlich and Mester, 1998; Van Huffel and Vandewalle, 1991]. Depending on the model used this may not be case here. To accommodate for this we simply scale the data vector accordingly, imply-ing diagonal covariance matrices. More elaborate schemes are discussed in [M¨uhlich and Mester, 1999; Van Huffel and Vandewalle, 1991].

4

Experiments

In this section we demonstrate the application of the described technique to real image sequences containing illumination changes and diffusion caused by a small field of depth.

4.1 Brightness Changes

In Fig. 2a,b two frames of a sequence containing a translating plane with a random dot texture are shown. In addition to the movement the illumination changes smoothly during the sequence. The scene is illuminated via a fiber optic bundle which moves towards the scene and causes a gradual increase in intensity. Such illumination changes are easily modeled in (2) by a linear source term f (g, t, a) = −q and a translational velocity v = [u v]T:

(4)

a b c

d e f

Fig. 2: Sequence with illumination changes: a frame 1, b frame 20 and c correct displacement

field. d Velocity estimated using standard optical flow constraint equation, e displacement when a linear source term is modeled and f estimated brightness changes in the range of [0, 2.5] greyvalues/frame.

In this case there even is a constant (error free) term in the data vector. Here we simply use an error variance for this term that is two orders of magnitude smaller than that in the other terms in the scaling procedure. In practice this simplified approach usually gives good results. However, it is possible to take this error structure explicitly into account to achieve even better results [Garbe et al., 2002].

The scene consists of a plane which is moved using a linear positioner. In our lab-oratory setup geometric calibration information for the observing camera is available. Thus we can compute the ground truth velocity field as shown in Fig. 2c. The veloc-ity computed assuming conserved brightness is given in Fig. 2d and that using a linear source term in Fig. 2e. In the later case we also obtain an estimate of the illumination change which is given in Fig. 2f.

Comparing the velocity fields (Fig. 2c,d,e) we can clearly see an improvement when the extended model is used. However because we do have available ground truth we can even put numbers to this improvement. The following table contains the relative error in the magnitude of the velocity, the directional error and the angular error often used in optical flow evaluations [Barron et al., 1994].

method density [%] rel. error [%] dir. error [◦] ang. error [◦]

standard 92.6 7.9 ± 6.3 3.3 ± 2.7 2.5 ± 1.4

extended 94.7 1.3 ± 1.2 0.5 ± 0.5 0.4 ± 0.3

Obviously there is a dramatic increase in accuracy when the illumination change is modeled.

(5)

4.2 Depth from Diffusion

An interesting application of the presented technique allows an extension of the depth from focus procedure. In its standard form a series of images with limited depth of field is acquired and at each pixel the depth is determined by the frame where it appears in focus. This technique does require telecentric lenses as the world point viewed by each pixel changes otherwise. It is common to model the blurring caused by out of focus imaging with a Gaussian point spread function. Hence the assumed underlying process is diffusion. If we thus model the changes in the intensity as a translation plus a diffusion we can capture both the motion due to the non telecentric lens and the amount of blur. Such a model can be formulated as:

gxu + gyv + gt= −D∆g → d = [gxgy∆g gt]T ; p = [u v D 1]T , (5) where D is the diffusion constant. It can be shown that this diffusion constant is directly proportional to the distance of the observed point to the plane in focus [Dierig, 2002]. Hence D is a direct measure of depth.

In Fig. 3 two real examples are given. The displacement field is diverging as ex-pected and the estimated depth appears to be qualitatively correct. A quantitative anal-ysis of the recovered depth on real data has yet to be done. For a realistic setup and typical image noise we obtain a relative error in the depth below 5% on synthetic data [Dierig, 2002]. This shows that the presented general parameter estimation framework can be used successfully to compute depth from focus sequences using standard off the shelf lenses thus avoiding expensive telecentric setups and allowing for a much wider field of view.

5

Conclusion

We have presented a general framework to estimate the parameters of dynamic pro-cesses in image sequences where the assumption of conserved brightness does not hold. This has potentially a very wide application. Here we quantitatively investigated the increase in accuracy of the computed displacement field on one sequence where the illumination changes. Furthermore we introduced a novel algorithm termed depth from diffusion to compute depth from focus series taken with non telecentric cameras. This is achieved by modeling blur as a diffusion process.

Acknowledgements. Part of this work has been founded under the DFG research unit “Image Sequence Analysis to Investigate Dynamic Processes” (FOR240) and by a fel-lowship within the Postdoc-Programme of the German Academic Exchange Service (DAAD).

Bibliography

J. L. Barron, D. J. Fleet, and S. Beauchemin. Performance of optical flow techniques. International Journal of Computer Vision, 12(1):43–77, 1994.

(6)

a b c

d e f

Fig. 3: Depth from Diffusion. First example: a one image of the sequence, b displacement field

and c estimated depth. Second example: d original image, e displacement field and f estimated depth.

T. Dierig. Gewinnung von Tiefenkarten aus Fokusserien. PhD thesis, University of Heidelberg, Heidelberg, Germany, July 2002.

C. Garbe, H. Spies, and B. J¨ahne. Mixed ols-tls for the estimation of dynamic processes with a linear source term. In DAGM, Lecture Notes in Computer Science, Z¨urich, Switzerland, September 2002. Springer.

H. Haußecker and D. J. Fleet. Computing optical flow with physical models of bright-ness variation. PAMI, 23(6):661–673, June 2001.

H. Haußecker, C. Garbe, H. Spies, and B. J¨ahne. A total least squares framework for low-level analysis of dynamic scenes and processes. In DAGM, pages 240–249, Bonn, Germany, 1999. Springer.

B. K. P. Horn and B. Schunk. Determining optical flow. Artificial Intelligence, 17: 185–204, 1981.

K. Langley. Motion perception and motion estimation by total-least squares. Spatial Vision, 15(2):171–190, 2002.

M. M¨uhlich and R. Mester. The role of total least squares in motion analysis. In ECCV, pages 305–321, Freiburg, Germany, 1998.

M. M¨uhlich and R. Mester. Subspace methods and equilibration in computer vision. Technical Report XP-TR-C-21, Institute for Applied Physics, Goethe-Universit¨at, Frankfurt, Germany, November 1999.

S. Negahdaripour. Revised definition of optical flow: Integration of radiometric and geometric cues for dynamic scene analysis. PAMI, 20(9):961–979, September 1998. S. Van Huffel and J. Vandewalle. The Total Least Squares Problem: Computational Aspects and Analysis. Society for Industrial and Applied Mathematics, Philadelphia, 1991.

References

Related documents

For the tuning process, a set of four nodules, with one reading each form a single patient case, were used, see figure 5.1. They were chosen because of their different

Av de histogram som presenteras i Figur 17 uppvisar hastighet ett tydligt normalfördelat mönster. Även residualerna för DTC ser ut att kunna vara normalfördelade.

The aim of this work is to develop an optimisation algorithm that can at each production stage in a multi-stage scheduling problem choose to use job allocation and job sequence or

antal stater och de interoperabilitetsproblem som kommer sig av detta minskar inte sannolikheten för seger för en allian/koalition i krig. Med tanke på den stora mängd av

Instead, due to the weak internal forces the contour will be stuck in the homogenous region (regions where the image forces are close to zero) or on false edges (edges other than

Keywords: high dynamic range, super-resolution, image reconstruction, optical flow, motion analysis, inverse problem, human visual system, digital camera system, multiple

Digital Rights Management (DRM) is a technology that allows service and content providers to distribute and sell digital content in a secure way.. The content is encrypted and

På grund av resultatet kan vi inte dra någon slutsats för om utnyttjandet av redovisningen av goodwill utnyttjas olika beroende på ersättning till verkställande