• No results found

ROF model on the graph

N/A
N/A
Protected

Academic year: 2021

Share "ROF model on the graph"

Copied!
25
0
0

Loading.... (view fulltext now)

Full text

(1)

& %

Department of Mathematics

ROF model on the graph

Japhet Niyobuhungiro and Eric Setterqvist

(2)
(3)

ROF model on the graph

Japhet Niyobuhungiro

Division of Mathematics and Applied Mathematics, Department of Mathematics, Linköping University, SE–581 83 Linköping, Sweden

Department of Mathematics, School of pure and applied sciences, College of Science and Technology, University of Rwanda, P.O. Box 117 Butare, Rwanda

Eric Setterqvist

Division of Mathematics and Applied Mathematics, Department of Mathematics, Linköping University, SE–581 83 Linköping, Sweden

Abstract

In this paper we consider an analogue of the well–known in image processing, Rudin–Osher–Fatemi (ROF) denoising model on a general finite directed and connected graph. We consider the space BV on the graph and show that the unit ball of its dual space can be described as the image of the unit ball of the space

`∞on the graph by a divergence operator. Based on this result, we propose a new fast algorithm to find the exact minimizer for the ROF model. Finally we prove convergence of the algorithm and illustrate its performance on some image denoising test examples.

Keywords: ROF model, Directed graph, L–functional, Image processing, Dual BV, Regularization. 2010 MSC: 46B70, 46E35, 68U10

1. Classical ROF model and ROF model on the graph

Let us suppose that we observed noisy image fob∈L2defined on a square domainΩ= [0, 1]2inR2

fob= f∗+η,

where f∗∈ BV is the original image and η∈L2is noise. Denoising is one of the problems which appear in

image processing: "How to recover the image f∗from the noisy image fob?". Variational methods using the

total variation minimization are often employed to solve this problem. The total variation regularization technique was introduced by Rudin, Osher and Fatemi in [1] and is called the ROF model. It suggests to take as an approximation to the original image f∗ the function fopt,t ∈ BV, which is the exact minimizer

for the L2,1– functional for the couple L2, BV:

L2,1  t, fob; L2, BV  = inf g∈BV  1 2kfob−gk 2 L2+tkgkBV  , for some t>0, (1) i.e., fopt,t∈BV is such that

L2,1  t, fob; L2, BV  = 1 2 fob− fopt,t 2 L2+t fopt,t BV. (2) By BV we denote the space of functions of bounded variation defined by the seminorm

kfkBV(Ω)= Z 1 0 varxf(x, y)dy+ Z 1 0 varyf(x, y)dx, ∗Corresponding author

Email addresses: japhet.niyobuhungiro@liu.se, jniyobuhungiro@ur.ac.rw (Japhet Niyobuhungiro), eric.setterqvist@liu.se(Eric Setterqvist)

(4)

where varxf(x, y) = sup 0≤x1≤...≤xn≤1 n−1

j=1 f xj+1, y  −f xj, y is a function of y equal to variation of f on the horizontal axis for y fixed, and

varyf(x, y) = sup 0≤y1≤...≤yn≤1 n−1

i=1 |f(x, yi+1) − f(x, yi)|

is a function of x equal to variation of f on the vertical axis for x fixed.

However the problem of actual calculation of the function fopt,t(see (1) and (2)) is non-trivial. Standard

approach is connected with discretization of the functional (1), i.e. we divideΩ into N×N square cells and instead of the space L2(Ω) consider its finite dimensional subspace SN which consists of functions

that are constant on each cell.

SN= ( f ∈L2(Ω): f = N

i,j=1 fij

χ

ij, ) , (3) where

χ

ij(x, y) =      1 if j−1N ≤x< Nj and i−1N ≤y< Ni; 0 otherwise. (4) Throughout, we will consider our discretization grid as a 2D Cartesian coordinate in (computer) screen space, i.e., with the origin at the top left corner and+y axis pointing down, which is the way matrices are represented on the computer, as illustrated in Figure 1.

Figure 1: Illustration of the rectangular grid. The grid contains N2cells refered to as vertices; N(N1)horizontal edges directed

left right and(N−1)N vertical edges directed up down. Total of 2N(N−1)edges.

As the BV seminorm of a function g∈SNis equal to

kgkBV(SN) = 1 N N

i=1 N−1

j=1 gi,j+1−gij + N

j=1 N−1

i=1 gi+1,j−gij ! ,

(5)

the discrete analogue of the functional (1) can be written as L2,1  t, fob; L2, BV  (5) = inf g∈SN 1 2N2 N

i,j=1  fobij−gij 2 ! + t N N

i=1 N−1

j=1 gi,j+1−gij + N

j=1 N−1

i=1 gi+1,j−gij !! .

Now we can consider the graph G= (V, E), where the set of vertices V corresponds to cells and the set of edges E corresponds to set of pairs of cells which have common faces (see Figure 1), i.e.

V=vij; i, j=1, 2, . . . , N , and E=EH∪EV, where EH= n eijH= vij, vi,j+1 ; i=1, 2, . . . , N and j=1, 2, . . . , N−1 o are the N(N−1)horizontal edges (that we choose to be directed left right) and

EV =

n

eVij = vij, vi+1,j ; i=1, 2, . . . , N−1 and j=1, 2, . . . , N

o

are the N(N−1)vertical edges (that we choose to be directed top down). So we have that dim V= N2and dim E=2N(N−1). Let us denote by SV and SE the set of real-valued functions on V and E respectively.

Next we consider the analogue of the gradient operator on the graph, i.e., grad : SV −→SE which maps

function f ∈SV to function grad f ∈SE defined as (see Figure 2)

Figure 2: The neighbors of the vertex vij.

(grad f)eijH= f vi,j+1− f vij , i=1, 2, . . . , N and j=1, 2, . . . , N−1;

(grad f)eVij= f vi+1,j− f vij , i=1, 2, . . . , N−1 and j=1, 2, . . . , N.

In this framework, the observed image fob∈SNcan be considered as an element of SV, and the functional

(5) can be written as L2,1  t, fob; L2, BV  = inf g∈SV  1 2N2kfob−gk 2 `2(SV)+ t Nkgrad gk`1(SE)  , (6)

(6)

where kfk`2(SV)=

v∈V (f(v))2 !12 ,khk`1(SE)=

e∈E |h(e)| =

e∈EH |h(e)| +

e∈EV |h(e)|. Of course, exact minimizer of (6) coincides with exact minimizer of

L2,1  s, fob;`2(SV), BV(SV  ) = inf g∈SV  1 2kfob−gk 2 `2(SV)+skgrad gk`1(SE)  , s=Nt. (7) This leads to the following analogue of the ROF model on a general finite directed and connected graph G= (V, E).

1.1. Formulation of the problem on general graph

Let G = (V, E) be a finite directed and connected graph with N vertices V = {v1, v2, . . . , vN}and M

directed edges E={e1, e2, . . . , eM}, where each edge is determined by a pair of two vertices e= vi, vj for

some i, j∈ {1, 2, . . . , N}. Let SV ={f : f : V−→R}denote the set of of real–valued functions defined on

the vertices and SE ={g : g : E−→R}denote the set of of real–valued functions defined on the edges.

Then the analogue of ROF model on the graph G can be formulated as follows:

Problem 1.1. Suppose that we know function fob∈SV. For given t>0, find exact minimizer of the functional

L2,1  t, fob;`2(SV), BV(SV)  = inf g∈BV(SV)  1 2kfob−gk 2 `2(S V)+tkgkBV(SV)  , (8) where kfk`2(SV)=

v∈V (f(v))2 !12 ;kfkBV(S V)=kgrad fk`1(SE);khk`1(SE)=

e∈E |h(e)|, (9) and operator grad : SV −→SEis defined by the formula

(grad f) (e) = f(vj) − f(vi)if e= (vi, vj).

In this paper we will suggest a fast reiterative algorithm that constructs exact minimizer of (8). We would like to note that the case of a general graph could be particularly useful when instead of usually considered rectangular domain we have some manifold, for example map of the Earth is an image on the sphere.

2. Description of the ball of dual space to BV(SV)

It was shown in [2] that exact minimizer for the L2,1-functional

L2,1  t, fob;`2, X  = inf g∈X  1 2kfob−gk 2 `2+tkgkX 

for the couple `2, X, where space`2is defined by the standard Euclidean normk·k2, X is a Banach space onRn and t is a given positive parameter, is equal to the difference between foband the nearest element

to fobof the ball of radius t>0 of the space X∗. Note that proofs in [2] are also true when X is equipped

with seminorm. So to construct exact minimizer for the L2,1-functional for the couple `2(SV), BV(SV)

we first need to describe the ball of radius t>0 of the space BV∗(SV)with norm defined by

khkBV(S V)= sup k f kBV(SV)≤1 hh, fiS V, where hh, fiSV =

v∈V h(v)f(v). (10)

Remark 1. Note thatkfkBV(S

V)=0 if and only if f =C for some constant C∈R. Therefore

khkBV(S V)=    sup k f kBV(SV)≤1 hh, fi , if ∑ v∈V h(v) =0; +∞ , otherwise.

(7)

In order to formulate the main result of this section, let us consider divergence operator on the graph, i.e. the operator div : SE →SV defined by

(div g)(v) =

i:(vi,v)∈E

g((vi, v)) −

j:(v,vj)∈E

g (v, vj) . (11)

We can now formulate the result

Theorem 2.1. The unit ball of the space BV∗(SV)is equal to the image of the unit ball of the space`∞(SE)under

the operator div, i.e.,

BBV(S V)=div  B`∞(SE)  .

The proof of Theorem 2.1 is based on two elementary lemmas. The first one is

Lemma 1. The operatordiv : SE−→SV is conjugate to the operator grad : SV −→SE, i.e.

hdiv g, fiS V =hg, grad fiSE, wherehf1, f2iSV = ∑ v∈V f1(v)f2(v)andhg1, g2iSE = ∑ e∈E g1(e)g2(e).

Proof. The expressionhg, grad fiS

E is, by definition of the operator grad, equal to

hgrad f , giS E =

e=(vi,vj)∈E  f vj−f(vi) g(e) =

e=(vi,vj)∈E f vj g vi, vj−

e=(vi,vj)∈E f(vi)g vi, vj =

v∈Vi:(v

i,v)∈E f(v)g((vi, v)) −

v∈Vj:(v,v

j)∈E f(v)g v, vj =

v∈V f(v)  

i:(vi,v)∈E g((vi, v)) −

j:(v,vj)∈E g v, vj  . The expressionhdiv g, fiS

V is, by definition of the operator div, equal to

hf , div giS V =

v∈V f(v) (div g) (v) =

v∈V f(v)  

i:(vi,v)∈E g((vi, v)) −

j:(v,vj)∈E g v, vj  . We conclude that hdiv g, fiS V =hg, grad fiSE. 

For the statement of the second lemma, we will need the following definition

Definition 2.1. Let X be a Banach space and let V be a subspace of X. The annihilator of V denotedAnn(V)is the set of bounded linear functionals that vanish on V. That is the set defined by

Ann(V) ={x∗ ∈X∗: hx∗, vi =0, for all v∈V}, (12) where X∗is the dual space of X.

Lemma 2. Let X be a Banach space with dual space X∗and let V be a finite dimensional subspace of X and Ann(V)

be the annihilator of V. Let x0∈X. Then

inf

v∈Vkx0+vkX =x∈Bsup

X∗∩Ann(V)

hx∗, x0i, (13)

(8)

Proof. The case x0 ∈ V is obvious. From now on we suppose x0 ∈/ V. Let us take an arbitrary x∗ ∈

BX∗∩Ann(V)and v∈V. Since x∗∈Ann(V)(i.e.,hx∗, vi =0, ∀v∈V) then we have

hx∗, x0i = hx∗, x0+vi ≤ kx∗kX∗kx0+vkX, ∀v∈V.

Since x∗∈ BX∗(i.e.,kx∗kX∗ ≤1) it follows that

hx∗, x0i ≤ kx0+vkX, ∀v∈V.

Therefore since x∗∈ BX∗∩Ann(V)and v∈V were arbitrary, we have that inf

v∈Vkx0+vkX ≥x∈Bsup

X∗∩Ann(V)

hx∗, x0i. (14)

In order to prove the reverse inequality, let us consider the space W, which is the algebraic sum between the span of x0and the space V:

W={x0} +V={w∈X : w=λx0+v, v∈V and λR}, (15)

and take v0 ∈ V such that infv∈Vkx0+vkX = kx0+v0kX. The existence of such v0 follows from the

assumption that V is a finite dimensional subspace of X. Without loss of generality we can assume that

kx0+v0kX = 1. Since W is a normed vector space by itself, it is possible to consider its dual space.

Moreover V is a linear subspace of W and then by the Hahn-Banach Theorem, since x0+v0 ∈ W and

x0+v0∈/V there exists a bounded linear functional x∗0∈W∗such that

hx∗0, vi =0 for all v∈V, and hx∗0, x0+v0i =1. (16)

It follows that

x∗0 ∈Ann(V) and hx∗0, x0i =1. (17)

Let us now investigate the action of x∗0on W. Let λR, v∈V and let w =λx0+v be an element of W.

Then we have hx∗0, wi = hx∗0, λx0+vi = hx∗0, λx0+λv0+v−λv0i =λhx∗0, x0+v0i + hx∗0, v−λv0i =λ, (18) becausex∗ 0, x0+v0 =1 andx∗ 0, v−λv0

=0 since x∗0∈Ann(V)and v−λv0∈V. Let us now describe

the unit ballBW of W. Suppose that w=λx0+v∈ BW where λ6=0. We have that

1≥ kwkX =kλx0+vkX =|λ| x0+ v λ X≥ |λ| kx0+v0kX =|λ|. (19)

Therefore w=λx0+v∈ BWimplies that|λ| ≤1. The W∗- norm of the functional x∗0 is by definition

kx∗0kW∗= sup

w∈BW

hx∗0, wi.

From (18) and (19), it follows that

kx∗0kW∗= sup

w∈BW

λ= sup

|λ|≤1

λ=1. (20)

By invoking the Hahn-Banach theorem, we can extend the functional x∗0:

∃xe0∗∈X∗such that ex∗0|W=x∗0and xe ∗ 0 X∗=kx ∗ 0kW∗=1. (21) From (17) and (21) we conclude that

e

(9)

It follows that inf v∈Vkx0+vkX =kx0+v0kX =hx ∗ 0, x0+v0i = D e x∗0, x0+v0 E (22) ≤ sup x∗∈B X∗∩Ann(V) hx∗, x0+v0i = sup x∗∈B X∗∩Ann(V) hx∗, x0i.

Putting (14) and (22) together, we obtain inf

v∈Vkx0+vkX =x∈Bsup

X∗∩Ann(V)

hx∗, x0i.



Proof of Theorem 2.1. The operator grad has a kernel which is given by

ker(grad) ={f ∈SV: f =C, for some C∈R}, (23)

and the annihilator (or orthogonal complement) of the ker(grad)is given by Ann(ker(grad)) =

( F∈SV :

v∈V F(v) =0 ) . (24)

Since from Lemma 1, div = (grad)∗, where (grad)∗is the conjugate operator of grad, then it is a simple result from linear algebra that

Im(div) =Ann(ker(grad)) and Im(grad) =Ann(ker(div)). (25) From Remark 1, note that if ∑v∈Vh(v) 6= 0 then khkBV∗(S

V) = +∞, so to avoid this case we will

re-strict ourselves on functions h ∈ Ann(ker(grad)). Then from relations (24) and (25), we conclude that BV∗(SV) =Im(div). Therefore for all h∈ BV∗(SV), there exists at least one g∈SEsuch that h=div g.

So let h ∈ BV∗(SV)and fix g0∈ SE such that h =div g0. Now let us consider the infimum of kgk`∞(SE)

over all functions g in the fiber of h∈BV∗(SV)under the operator div.

inf

h=div gkgk`∞(SE)={u∈ker(div)}inf kg0+uk`∞(SE)

Put U=ker(div). Then we have Ann(U) =Im(grad). By applying Lemma 2, we have that inf h=div gkgk`∞(SE)= sup k f∗k`1(SE)≤1, f∗=grad f , f ∈SV  hf∗, g0iSE = sup kgrad f k`1(SE)≤1 hgrad f , g0iSE.

From Lemma 1, h=div g0, (9) and (10) follows that

inf h=div gkgk`∞(SE)= sup kgrad f k`1(SE)≤1 hf , div g0iSV = sup kgrad f k`1(SE)≤1 hf , hiS V = sup k f kBV( SV)≤1 hf , hiS V =khkBV(S V). It means that khkBV(S

V)≤1 if and only if h=div ginf kgk`∞(SE)≤1. (26)

Note that the infimum is attained because ker(div)is a finite dimensional subspace of SE. Therefore there

exists an element

gh∈g0+ker(div), such that kghk`∞(SE)≤1 and div gh=h.

We conclude that BBV(S V)=div  B`∞(SE)  . 

(10)

2.1. Corollary for the discrete ROF model

As a corollary of Theorem 2.1 we will have the following result (for a reminder of space SN, see

expression (3)).

Corollary 1. Let h ∈ SN. Then khkBV∗(S

N) ≤ 1 if and only if h can be decomposed as h = h1+h2, with

h1, h1∈SNsuch that sup 0≤x≤1 Z x 0 h1(s, y)ds ≤1 and Z 1 0 h1(s, y)ds=0, for all y∈ [0, 1]; and sup 0≤y≤1 Z y 0 h2(x, t)dt ≤1 and Z 1 0 h2(x, t)dt=0, for all x∈ [0, 1].

Proof. Let us take h ∈ SN such thatkhkBV∗(SN) ≤1. By interpretation of the space SN on the graph, we

have from Theorem 2.1 that

khkBV(S

N)≤1 if and only if h=div g for some g∈SEsuch that kgk`∞(SE)≤N,

where h vij  = (div g) vij  =hgeHi,j−1−geijHi+hgei−1,jV −geViji, i, j=1, . . . , N, with the assumption geHi,0=geV0,j=0. It follows that

h=h1+h2,

where functions h1, h2∈SV are defined as follows

h1 vij



=hgei,j−1H −geHiji and h2 vij



=hgeVi−1,j−geViji. (27) From (27) we deduce that

               g ei1H = −h1(vi1); g ei2H = − (h1(vi1) +h1(vi2)); .. . geH i,N−1  = − (h1(vi1) +h1(vi2) +. . .+h1(vi,N−1)); g eH iN  = − (h1(vi1) +h1(vi2) +. . .+h1(vi,N−1) +h1(viN)). ∀i=1, 2, . . . , N. (28)

From the the first N−1 equations of (28), we have that sup 1≤k≤N−1 g  eikH = sup 1≤k≤N−1 k

j=1 h1 vij  , ∀i=1, 2, . . . , N. (29) From the the last equations of (28), we deduce that

0=geiNH= −

N

j=1

h1 vij , ∀i=1, 2, . . . , N. (30)

From the correspondance between SV and space SN, the expression (29) implies that

sup 1≤k≤N−1 g  eikH = sup 1≤k≤N−1 N Z k N 0 h1(s, y)ds = sup 0≤x≤1 N Z x 0 h1(s, y)ds , ∀y∈ [0, 1]. (31) Sincekgk`∞(SE)≤N, it follows from (31) that

sup 0≤x≤1 N Z x 0 h1 (s, y)ds ≤N, ∀y∈ [0, 1].

(11)

Therefore sup 0≤x≤1 Z x 0 h1 (s, y)ds ≤1, ∀y∈ [0, 1]. From (30), we have that

N

j=1 h1 vij  =N Z NN 0 h1 (s, y)ds=0, ∀y∈ [0, 1]. Therefore Z 1 0 h1 (s, y)ds=0, ∀y∈ [0, 1]. We conclude that sup 0≤x≤1 Z x 0 h1(s, y)ds ≤1 and Z 1 0 h1(s, y)ds=0, for all y∈ [0, 1].

Similar arguments for the function h2and the function g acting on the edges eVij will yield that

sup 0≤y≤1 Z y 0 h2 (x, t)dt ≤1 and Z 1 0 h2 (x, t)dt=0, for all x∈ [0, 1]. 

3. Algorithm for construction of exact minimizer

Below we present a fast reiterative algorithm for the actual construction of the element fopt,t. This

algorithm is a specialized version of an algorithm constructed in [3] for problems concerning a weighted divergence operator on a general finite directed graph.

As was mentioned in the previous section, we know from [2] that the exact minimizer fopt,t for the

L2,1-functional L2,1  t, fob;`2(SV), BV(SV)  = inf g∈BV(SV)  1 2kfob−gk 2 `2(SV)+tkgkBV(SV)  is given by fopt,t= fob−eh, (32)

where eh is the nearest element to fob, in the metric of`2(SV), in the ball tBBV∗(S

V), i.e. inf h∈tBBV∗ (SV) kfob−hk`2(SV)= fob−eh `2(S V) . (33)

Next, from Theorem 2.1 we know that tBBV(S

V) = t div



B`∞(SE)



for t > 0. The proposed algorithm constructs eh through a sequence of elements gn ∈tB`∞(SE)such that

(12)

3.1. Algorithm

The algorithm consists of several steps outlined in this section.

Let fob ∈ SV on G= (V={v1, . . . , vN}, E={e1, . . . , eM}), a regularization parameter t and a maximum

number of iterations Niterbe given. Set

ek= vi, vj



∈E, k=1, 2, . . . , M; for some i, j∈ {1, 2, . . . , N}. Define the operator T as follows:

T=TMTM−1TM−2. . . T2T1,

where for k=1, 2, . . . , M, Tk: tB`∞(SE)−→tB`∞(SE)is defined as follows:

(Tkg) (e) =               Kg(ek) ,if Kg(ek) ∈ [−t,+t]; −t ,if Kg(ek) < −t; +t ,if Kg(ek) > +t. ,if e=ek; g(e) ,if e6=ek. where Kg(ek) =  fob(vj) − div\ekg  (vj)− fob(vi) − div\ekg  (vi) 2 . and        div\ekg(vi) = (div g) (vi) +g(ek); div\ekg  vj= (div g) vj−g(ek); and div\ekg(v`) = (div g) (v`), ∀` 6=i, j.

With these notions at our disposal, we now outline the steps of the algorithm.

Step 1. Take g0=0, or choose any g0∈tB`∞(SE)

Step 2. Calculate g = Tg0. i.e., calculate (Tg0) (ek) for k = 1, 2, . . . , M. If g = g0 then take eh = div(g0),

otherwise go to Step 3.

Step 3. Put g0=g and go to Step 2.

We continue this process applying the operator T to the new element g∈tB`∞(SE)generating the sequence

of elements g0, g1=Tg0, g2=Tg1, . . . , gn=Tgn−1with gn ∈tB`∞(SE), n=0, 1, 2, . . . until the maximum

number of iterations Niteris reached. In the next Section we will show that

div(gn) −→eh as n→ +∞ in the metric of`2(SV).

To prove the convergence of the proposed algorithm we need the following proposition

Proposition 3.1. Let eh be the minimizer defined by (33). The operator T is continuous and satisfies the following two conditions

(a) For any g∈tB`(SE), div g=eh if and only if Tg=g; (b) For any g∈tB`∞(SE), if div g6=eh and

kfob−div(Tg)k`2(SV)<kfob−div gk`2(SV). (34) Proof. It is clear from definition that small changes of g∈tB`∞(SE)leads to small changes of any operator

Tkand therefore the operator T is continuous as a product of continuous operators.

We now prove (a). Assume first that g∈ tB`(SE)is such that div g =eh. Let ek = (vi, vj) ∈E. We note that g(ek)appears only in the following two terms ofkfob−div gk2`2(S

V): fob(vj) − (div g)(vj)2+ (fob(vi) − (div g)(vi))2= (35) h fob(vj) − (div\ekg)(vj) −g(ek) i2 +hfob(vi) − (div\ekg)(vi) +g(ek) i2 .

(13)

Since div g=eh, g(ek)in particular must minimize (35) in the interval[−t, t]. We note that by the Jensen’s inequality ω(g(ek)) = h fob(vj) − (div\ekg)(vj) −g(ek) i2 +hfob(vi) − (div\ekg)(vi) +g(ek) i2 ≥ 2   h fob(vj) − (div\ekg)(vj) −g(ek) i +hfob(vi) − (div\ekg)(vi) +g(ek) i 2   2 = 2   h fob(vj) − (div\ekg)(vj) i +hfob(vi) − (div\ekg)(vi) i 2   2

with equality if and only if

fob(vj) − (div\ekg)(vj) −g(ek) = fob(vi) − (div\ekg)(vi) +g(ek), i.e. when g(ek) = h fob(vj) − (div\ekg)(vj) i −hfob(vi) − (div\ekg)(vi) i 2 = (Kg)(ek).

Moreover, ω(x)is strictly convex and therefore strictly decreasing for x< (Kg) (ek)and strictly increasing

for x > (Kg)(ek). So the minimal value of ω(x) on the interval [−t, t] is only attained at i) the point

(Kg)(ek) if (Kg)(ek) ∈ [−t, t], ii) the point −t if (Kg)(ek) < −t and iii) the point t if(Kg)(ek) > t. The

assumption div g=eh then implies that g(ek)must be the nearest point in the interval[−t, t]to(Kg)(ek), i.e. Tkg(ek) =g(ek). Since ek ∈E was arbitrary it follows that Tkg(ek) =g(ek), k=1, ..., M which implies

Tkg=g, k=1, ..., M and therefore Tg=g.

We now show the other direction of (a). Let us assume that g ∈ tB`∞(SE) and Tg = g. Then for

any edge e ∈ E, g(e) coincides with the point of the interval [−t, t] which is nearest to (Kg)(e). As

kfob−div gk`2(SV)is a convex function of g on tB`∞(SE), to show this direction it is enough to show that g

minimizeskfob−div gk`2(S

V)locally, i.e. it is enough to show that for some small ε>0 we have

kfob−div gk`2(SV)= inf

f ∈Rε

kfob−div fk`2(SV),

where Rε is the tubular set given by

Rε=

n

f ∈tB`∞(SE): kg−fk`∞(SE)≤ε

o . Note that for any f ∈Rεand e∈E we have

f(e) ∈ [−t, t] ∩ [g(e) −ε, g(e) +ε].

The set Rεis a closed convex subset of SE, therefore there exists a function fε ∈Rε such that kfob−div fεk`2(S

V)= f ∈Rinf ε

kfob−div fk`2(S V)

so we need to prove that

kfob−div gk`2(SV)=kfob−div fεk`2(SV).

We first note that it follows from the necessity direction proved above that for any edge e∈ E, fε(e)will

coincide with the point of the interval[−t, t] ∩ [g(e) −ε, g(e) +ε]which is nearest to(K fε)(e).

Let us now decompose the set of edges E into two parts, the first part which we denote byΩgconsists

of the edges for which(Kg)(e)does not belong to the interval[−t, t], i.e. Ωg={e∈E :(Kg)(e)∈ [−/ t, t]}.

As g(e)is the nearest point in the interval[−t, t]to(Kg)(e)we have g(e) =



−t ,if Kg(e) < −t;

(14)

If the number ε > 0 is small enough, it follows fromkg− fk`∞(SE) ≤ εthat on e ∈ Ωg where we have

(Kg)(e) < −t we will also have(K fε)(e) < −t and therefore fε(e) = −t= g(e). Analogously, if ε >0 is

small enough, on e∈ Ωgwhere(Kg)(e) >t we will have(K fε)(e) > t and therefore fε(e) = t=g(e). So

we have proved that

fε(e) =g(e)for all e∈Ωg. (36)

The next step is to consider the graph G0 = (V, E\Ωg), i.e. the graph G with edges in Ωg removed.

G0 is the union of several connected components (Vk, Ek), k = 1, ...,` (we have V1∪...∪V` = V and

E1∪...∪E`=E\Ωg). Note that some of the graphs(Vk, Ek)can consist just of one single vertex. For these

graphs there is nothing to prove as Ek=∅.

Let us now consider a subgraph(Vk, Ek)where Ek 6=∅. On each e∈ Ek we have(Kg)(e) ∈ [−t, t]and

therefore g(e) = (Kg)(e), i.e. if e= (vi, vj)then

g(e) = (Kg)(e) = h fob(vj) − (div\eg)(vj) i −hfob(vi) − (div\eg)(vi) i 2 , or equivalently fob(vj) − (div g)(vj) = fob(vi) − (div g)(vi).

(Here it is important to note that operators K, div and div\e are considered in the orignal setting of

G= (V, E).) Therefore, for all v∈Vkthe values of

fob(v) − (div g)(v)

are equal. So we have

v∈Vk [fob(v) − (div g)(v)]2=|Vk| ∑v∈Vk[fob(v) − (div g)(v)] |Vk| !2 . (37)

For fεwe can with Jensen’s inequality derive the corresponding inequality

v∈Vk [fob(v) − (div fε)(v)] 2≥ | Vk| ∑v∈Vk[fob(v) − (div fε)(v)] |Vk| !2 . (38)

Now, note that flow on edges in Ekare cancelled in the sums

v∈Vk

[fob(v) − (div g)(v)] and

v∈Vk

[fob(v) − (div fε)(v)]. (39)

Therefore, only flows on edges inΩgremain in these sums. It then follows from (36) that

v∈Vk

[fob(v) − (div fε)(v)] =

v∈Vk

[fob(v) − (div g)(v)].

Therefore, taking into account (37) and (38), we obtain

v∈Vk [fob(v) − (div fε)(v)] 2

v∈Vk [fob(v) − (div g)(v)]2.

Summing over all Vkthen gives

kfob−div fεk 2 `2(S V)≥ kfob−div gk 2 `2(S V).

Hence, from the definition of fεwe must have kfob−div fεk

2

`2(SV)=kfob−div gk2`2(SV).

(15)

To prove (b) note that for ∀g ∈tB`∞(SE)the operators Tk, k =1, ..., M, by definition (see also necessity

part of (a)) satisfy

kfob−div(Tkg)k`2(SV)≤ kfob−div gk`2(SV)

with equality if and only if Tkg(ek) =g(ek). This implies for the operator T=TMTM−1...T1that

kfob−div(Tg)k`2(S

V)≤ kfob−div gk`2(SV)

with equality if and only if Tg= g which in turn by condition (a) is equivalent to div g =eh. Hence for any g∈tB`(SE), if div g6=eh then

kfob−div(Tg)k`2(S

V)<kfob−div gk`2(SV),

which establishes (b) and completes the proof. 

3.2. Convergence of the algorithm

With Proposition 3.1 at hand, we can now prove convergence of the Algorithm.

Theorem 3.1. Let eh be the minimizer defined by (33), g∈tB`(SE)and let T be the operator constructed in Section 3.1. Then

div(Tng) −→eh as n→ +∞ in the metric of`2(SV). (40)

Proof. Let T be the operator constructed in Section 3.1. From Proposition 3.1, T : tB`(SE)−→tB`(SE)is continuous and satisfies the conditions(a)and(b). Then

Tng∈tB`(SE), ∀g∈tB`(SE); n=0, 1, 2, . . . . (41) Since T satisfies conditions (a) and (b), the sequence of numbers (fob−div(Tng))n∈N

`2(S

V) is

mono-tonically decreasing and is bounded below by inf

g∈tB`(SE)kfob−div gk`2(SV). Therefore it converges. Let us

now consider the sequence(Tng)n∈N⊂tB`(SE). Since tB`(SE)is a compact set, then(Tng)n∈Ncontains a convergent subsequence in tB`(SE), say(Tnkg)

k∈N:

lim

k→∞(T

nkg) =g

h∈tB`∞(SE). (42)

Since T, div andk·k`2(S

V)are continuous operators, we have

kfob−div(Tgh)k`2(S V)= fob −div  T  lim k→∞(T nkg)  `2(S V) = lim k→∞kfob−div(T(T nkg))k `2(S V) = lim k→∞ fob−div  Tnk+1g `2(SV).

From Proposition 3.1 and continuity of div andk·k`2(SV)it follows that

kfob−div(Tgh)k`2(S V)=k→∞lim fob−div  Tnk+1g `2(S V) ≥ lim k→∞kfob−div(T nk+1g)k `2(S V) = fob−div  lim k→∞(T nk+1g)  `2(S V) =kfob−div ghk`2(S V).

Therefore by Proposition 3.1 we conclude that div gh=eh. By continuity of div andk·k`2(S

V)we have that lim k→∞kfob−div(T nkg)k `2(S V)=kfob−div ghk`2(SV).

Now, since the subsequencekfob−div(Tnkg)k`2(S V)



k∈Nmust converge to the same limit as

 kfob−div(Tng)k`2(SV)  n∈N, we conclude that lim n→∞kfob−div(T ng)k `2(S V)=kfob−div ghk`2(SV)= fob−eh `2(S V) .

(16)

Therefore, as eh is the unique nearest element to fobin tBBV∗(S

V)we conclude that

div(Tng) −→div gh=h as ne → +∞ in the norm of`2(SV).



4. Tikhonov functional

In this section, we consider the Tikhonov functional in order to use it for comparaison purposes in the next section. Let us consider the Tikhonov functional for the couple of Sobolev spaces L2(Ω), ˙W1,2(Ω) defined on a square domainΩ= [0, 1]2inR2. It is given by:

L2,2  t, fob; L2(Ω), ˙W1,2(Ω)  = inf g∈ ˙W1,2(Ω)  1 2kfob−gk 2 L2(Ω)+tkgk2W˙1,2(Ω)  , (43) where kfkW˙1,2(Ω) = ∂xf(x) L2(Ω) = ∂x1 f(x1, x2) 2 L2 + ∂x2 f(x1, x2) 2 L2 !1/2 , (44) and kfkL2(Ω) = Z 1 0 Z 1 0 (f(x1, x2)) 2 dx1dx2 1/2 . (45)

We need to find the element fopt,t∈W˙1,2(Ω), exact minimizer for (43). That is fopt,t∈W˙1,2(Ω)such that

L2,2  t, fob; L2(Ω), ˙W1,2(Ω)  = 1 2 fob−fopt,t 2 L2(Ω)+t fopt,t 2 ˙ W1,2(Ω). (46)

The exact minimizer fopt,t ∈ W˙ 1,2 for (43) can be found directly by solving the corresponding optimality

condition. However functions from SN don’t have derivative (44). So we need to look for an approximate

smooth function for which (44) exists. We consider a finite dimensional subspace of ˙W1,2(Ω) which consists of functions that can be represented as trigonometric polynomials of degree≤N. To this end, we will carry the processing in the Fourier transform domain where the image is represented as a weighted sum of the complex exponentials. Afterwards, we apply the inverse Fourier transform to return to the spatial domain. We will consider

fob(n1, n2) = 1 N2 N−1

k=0 N−1

`=0 cfob(k,`)

e

i2π N (kn1+`n2), n 1, n2=0, 1, . . . , N−1, (47) and g(n1, n2) = 1 N2 N−1

k=0 N−1

`=0 b g(k,`)

e

i2πN(kn1+`n2), n 1, n2=0, 1, . . . , N−1, (48) where c fob(k,`) = N−1

n1=0 N−1

n2=0 fob(n1, n2)

e

i2π N (kn1+`n2), k,` =0, 1, . . . , N1, (49) and b g(k,`) = N−1

n1=0 N−1

n2=0 g(n1, n2)

e

i2π N(kn1+`n2), k,` =0, 1, . . . , N1. (50)

(17)

Equations (49)–(50) correspond to the discrete Fourier transforms of the images defined over the discrete finite 2D grid of size N×N and equations (47)–(48) correspond to their inverses. We use N samples in both directions and in both spatial and frequency domains. Note that the image is periodized along both dimensions with period N. In this formulations, the functional (43) becomes

inf b g(k,`) 1 2 N−1

k,`=0 cfob(k,`) −gb(k,`) 2 +t "N−1

k,`=0 i Nk 2 |bg(k,`)|2+ N−1

k,`=0 i N` 2 |gb(k,`)|2 #! . So we need to minimize each term inside the sum, i.e.,

inf b g(k,`) 1 2 cfob(k,`) −gb(k,`) 2 +t Nk 2 + N` 2! |bg(k,`)|2 ! , k,` =0, 1, . . . , N−1. (51) We will represent the complex numbers cfob(k,`)andgb(k,`)in their polar form, i.e.,

c

fob(k,`) =rk,`

e

k,` andbg(k,`) =erk,`

e

k,`. (52)

Note that rk,` and ϕk,`, k,` = 0, 1, . . . , N−1 are known because fob is given. We only need to solve for

erk,`’s. Putting (52) in (51), we need to solve inf erk,` 1 2(rk,`−erk,`) 2+ 2π N 2 tk2+ `2er2k,` ! , k,` =0, 1, . . . , N−1. Taking derivative with respect toerk,`, we see that the infimum is attained aterk,`which satisfies

− (rk,`−erk,`) +  2π N 2 tk2+ `2· 2erk,`=0, k,` =0, 1, . . . , N−1, we have erk,`= rk,` 1+2 N2 t(k2+ `2), k,` =0, 1, . . . , N−1.

Therefore the exact minimizer fopt,tis given by the formula

fopt,t(n1, n2) = 1 N2 N−1

k=0 N−1

`=0 [fopt,t(k,`)

e

i2π N (kn1+`n2), n 1, n2=0, 1, . . . , N−1, (53) where [fopt,t(k,`) = rk,` 1+2 N2 t(k2+ `2)

e

k,`. The matlab routines f f t2 and i f f t2 are employed for a fast computation of (53).

5. Experimental results

In this section we will illustrate on some test images the performance of ROF model using our algorithm and we will compare with result obtained by using the Tikhonov functional described in the previous section. All the test observed images used have width 512 and height 512 pixels and the corresponding original image is provided for comparison purposes. In view of our analogue of ROF model on the graph (see (7)), note that the regularization parameter we use for ROF model is s=Nt, where N =512 and t is the regularization parameter appearing in the original model (see (1)). The original image used in Figures 3–4 is an 8–bit gray–scale image with intensity values ranging from 0 (black) to 255 (white). The original image used in Figures 5–6 is an artificial image with 4 parts of intensity values 180, 120, 90 and 50. The original image used in Figures 7–8 is an artificial binary image of intensity values 150 and 100. In all the cases the noisy image is the corrupted version of the original image where a Gaussian additive noise of

(18)

standard deviation ε with different values of ε is added to the original image. i.e, if we denote the original image by f∗, then the observed noisy image fobis defined by

fob= f∗+ε×randn(512, 512),

where randn(N, N) is a function returning a N–by–N matrix containing pseudorandom values drawn from the standard normal distribution as buit in Matlab.

In Figures 4, 6 and 8, we have highlighted in white the parts of the reconstructed image fopt,t where

f∗i,j−fopt,ti,j

α, i, j=1, 2, . . . , N, for different values of α in both reconstructions.

Figure 3: Lenna, the original is a 512×512 image with intensity values ranging from 0 to 255. Top left: Original image f∗; Top

right: Noisy image fob=f∗+ε×randn(512,512), i.e image with Gaussian additive noise of standard deviation ε=23.6. Bottom left:

(19)

Figure 4: Top: Residual image, i.e., the difference between the observed image and the reconstructed image fob−fopt,t. Top

left: Residual from Tikhonov reconstruction. Top right: Residual from ROF reconstruction. Bottom: The white shows, matrix componentwise, parts of the recontructed image fopt,twhere

f∗i,j−fopt,ti,j

≥35, i, j=1, 2, . . . , N. Left: in Tikhonov reconstruction. Right: in ROF reconstruction.

(20)

Figure 5: Geometric features with stairs, the original is a 512×512 image with 4 regions of intensity values 180, 120, 90 and 50. Top left: Original image f∗; Top right: Noisy image fob = f∗+ε×randn(512,512), i.e image with Gaussian additive noise of standard

(21)

Figure 6: Geometric features with stairs. Top: Residual image, i.e., the difference between the observed image and the reconstructed image fob−fopt,t. Top left: Residual from Tikhonov reconstruction. Top right: Residual from ROF reconstruction. Bottom: The

white shows, matrix componentwise, parts of the recontructed image fopt,twhere

f∗i,j−fopt,ti,j

≥20, i, j=1, 2, . . . , N. Left: in Tikhonov reconstruction. Right: in ROF reconstruction.

(22)

Figure 7: Geometric features with a cusp, the original is a 512×512 image with regions of intensity values 150 and 100. Top left: Original image f∗; Top right: Noisy image fob=f∗+ε×randn(512,512), i.e image with Gaussian additive noise of standard deviation ε=30.85. Bottom left: Tikhonov reconstruction using t=4.37. Bottom right: ROF reconstruction using s=36.5.

(23)

Figure 8: Geometric features with a cusp. Top: Residual image, i.e., the difference between the observed image and the reconstructed image fob−fopt,t. Top left: Residual from Tikhonov reconstruction. Top right: Residual from ROF reconstruction. Bottom: The

white shows, matrix componentwise, parts of the recontructed image fopt,twhere

f∗i,j−fopt,ti,j

≥20, i, j=1, 2, . . . , N. Left: in Tikhonov reconstruction. Right: in ROF reconstruction.

(24)

6. Final remarks and discussions

Since its appearance in 1992, the ROF model has received a large amount of popularity for its effeciency in regularizing images without smoothing the boundaries, and it has since been applied to a multitude of other imaging problems (see for example the book [4]). Some of the earlier works for minimizing the total variation based on dual formulation include [5], where the methods presented is based on removal of some of the singularity caused by the non-differentiability of the quantity grad f appearing in the regularization term. In 2004, Chambolle (see [6]) provided an algorithm related to [5] for minimizing the total variation of an image and proved its convergence. In the last few years image decomposition models into a piecewise– smooth and oscillating components that usually people refer to as cartoon and textures (or textures +

noise) respectively, have received interest in the image processing community. For example fopt,t ∈ BV

satisfying (2) is such that

fob= fob− fopt,t+ fopt,t.

This is the decomposition of fob into the piecewise–smooth component fopt,t ∈ BV and the component

fob−fopt,t ∈ L2 wich contains textures and noise. The original theoretical model for such an image

decomposition was introduced in 2002 by Yves Meyer in [7] by using the total variation to model the piecewise–smooth component and an appropriate dual space named G which is the Banach space com-posed of the distributions f which can be written f =1g1+2g2=div(g), with g1and g2in L∞(Ω)with

the following norm

kfkG=inf 

kgkL∞(Ω;R2): f =div(g), kgkL∞(Ω;R2)=ess supx∈

q

|g1(x)|2+|g2(x)|2

 ,

to model the oscillating component. Some of the most known works proposed in the literature for numeri-cally solving the Meyer’s model or its variants include for instance the works of [8] who proposed a model that splits the image into three components, a geometrical components modeled by the total variation, a texture component modeled by a negative Sobolev norm and a noise component modeled by a negative Besov norm; the works of [6, 5, 9, 10] that proposed an efficient projection algorithm to minimize the total variation. In 2006 [11] and [12] adapted Chambolle’s algorithm in cases of presence of an operator like a convolution kernel for example. Recently, [13] showed that the Bregman iteration is a very efficient and fast way to solve TV problems among other L1–regularized optimization problems and proposed a split Bregman method which they applied to the Rudin–Osher–Fatemi functional for image denoising and [14] designed an algorithm by using the Split Bregman iterations and the duality used by Chambolle to find the minimizer of a functional based on Meyer G–norm. Other works based on the Meyer’s G–norm include for example [15] and [16].

Acknowledgement. The authors are very grateful to Professor Natan Kruglyak for his invaluable suggestions and

discussions that made a considerable improvement to this work.

References

[1] L. I. Rudin, S. Osher, E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica, North-Holland D (60) (1992) 259–268.

[2] N. Kruglyak, J. Niyobuhungiro, Characterization of Optimal Decompositions in Real Interpolation, submitted.

[3] N. Kruglyak, E. Setterqvist, Weighted Divergence on the Graph.

[4] T. F. Chan, J. Shen, Image Processing and Analysis. Variational, PDE, Wavelet, and Stochastic Methods, Siam, 2005.

[5] T. F. Chan, G. H. Golub, P. Mulet, A nonlinear primal–dual method for total variation–based image restoration, Siam J. Sc. Comput. 20 (6) (1999) 1964–1977.

[6] A. Chambolle, An algorithm for total variation minimization and applications, Journal of mathemati-cal imaging and vision 20 (2004) 89–97.

(25)

[7] Y. Meyer, Oscillating Patterns in Image Processing and Nonlinear Evolution Equations, University Lecture Series, Vol. 22, AMS Providence, 2002.

[8] J. F. Aujol, A. Chambolle, Dual norms and image decomposition, International journal of computer vision 63 (1) (2005) 85–104.

[9] J. F. Aujol, G. Aubert, L. Blanc-Feraud, A. Chambolle, Image decomposition into a bounded variation component and an oscillating component, Journal of mathematical imaging and vision 22 (2005) 71– 88.

[10] A. Chambolle, J. Darbon, Algorithms for total variation minimization, Mathematical model for multi-channel image processing, Beijing 7 (2006) 7–8.

[11] J. Gilles, Noisy image decomposition: a new structure, texture and noise model based on local adap-tivity.

[12] J. F. Aujol, G. Gilboa, T. Chan, S. Osher, Structure-texture image decomposition-modeling, algorithms, and parameter selection, International Journal of Computer Vision 67 (1).

[13] T. Goldstein, S. Osher, The split bregman method for L1–regularized problems, Siam J. Imaging Sci-ences 2 (2) (2009) 323–343.

[14] J. Gilles, S. Osher, Bregman implementation of Meyer’s G–norm for cartoon + textures decomposition. [15] D. M. Strong, J. F. Aujol, T. F. Chan, Scale recognition, regularization parameter selection and Meyer’s

G– norm in total variation regularization, Multiscale Model. Simul 5 (1) (2006) 273–303.

[16] S. Osher, O. Scherzer, G–norm properties of bounded variation regularizaton, Comm. Math. Sci 2 (2) (2004) 237–254.

References

Related documents

By means of a literature review, this article analyses how research on benchmark- ing e-government has evolved over time, with a particular focus on its main findings and the

In this experiment, we tested the impact on the sensitivity and detection delay for varying model size N and decay factor T between models, when the underlying structure of the

1) Overall survey of the trade and industry in Wuhan. 2) General description of the environmental situation of Wuhan and different environmental departments and

You suspect that the icosaeder is not fair - not uniform probability for the different outcomes in a roll - and therefore want to investigate the probability p of having 9 come up in

However, the board of the furniture company doubts that the claim of the airline regarding its punctuality is correct and asks its employees to register, during the coming month,

Ahmed Karadawi (1983:539), himself Sudanese and Assistant Commissioner for Refugees, wrote that refugees have often been portrayed wrongly as sources of political,

According to Lo (2012), in the same sense “it points to the starting point of the learning journey rather than to the end of the learning process”. In this study the object

In summary, we have in the appended papers shown that teaching problem- solving strategies could be integrated in the mathematics teaching practice to improve students