• No results found

Maximum Likelihood Estimation in the Tensor Normal Model with a Structured Mean

N/A
N/A
Protected

Academic year: 2021

Share "Maximum Likelihood Estimation in the Tensor Normal Model with a Structured Mean"

Copied!
16
0
0

Loading.... (view fulltext now)

Full text

(1)

' $

Department of Mathematics

Maximum Likelihood Estimation in the

Tensor Normal Model with a Structured Mean

Joseph Nzabanita, Dietrich von Rosen, Martin Singull

LiTH-MAT-R--2015/08--SE

(2)

Department of Mathematics Link¨oping University S-581 83 Link¨oping, Sweden.

(3)

Maximum likelihood estimation in the tensor

normal model with a structured mean

Joseph Nzabanita1,a, Dietrich von Rosen1,b, Martin Singull1

1Department of Mathematics,

Link¨oping University, SE–581 83 Link¨oping, Sweden. E-mail: joseph.nzabanita@liu.se E-mail: dietrich.von.rosen@liu.se

E-mail: martin.singull@liu.se

aDepartment of Mathematics,

University of Rwanda, PO.Box 3900 Kigali, Rwanda E-mail: j.nzabanita@ur.ac.rw

bDepartment of Energy and Technology,

Swedish University of Agricultural Sciences, SE–750 07 Uppsala, Sweden.

E-mail: dietrich.von.rosen@slu.se

Abstract

There is a growing interest in the analysis of multi-way data. In some studies the inference about the dependencies in three-way data is done using the third order tensor normal model, where the focus is on the estimation of the variance-covariance matrix which has a Kronecker product structure. Little attention is paid to the structure of the mean, though, there is a potential to improve the analysis by assuming a structured mean. In this paper, we introduce a 2-fold growth curve model by assuming a trilinear structure for the mean in the tensor normal model and propose an algorithm for estimating parameters. Also, some direct generalizations are presented.

Keywords: growth curve model, Kronecker product structure, maximum likelihood estimators, multi-way data, tensor normal model, trilinear regression

1

Introduction

The analysis of multi-way data (or tensor data or array data) have become popular in various fields such as chemometrics [1, 2, 3], psychometrics [4, 5], signal processing [6, 7, 8], imaging [9], and many others. The multi-way data consist of measurements indexed by multiple categorical factors. For example, the 3-dimensional data may be obtained when a single response is sampled in 3-D space or in 2-D space and time, or when multiple responses are recorded in 2-D space or in 1-D space and time.

(4)

Several methodologies for the analysis of multi-way data have been adopted with a pre-dominance of the tensor decomposition approach [10]. Since multi-way data generally present correlations both within and among dimensions, the tensor normal model [11, 12, 13] is often chosen as a model for variation among entries of the data tensor. Thus, the variance-covariance structure is modeled as a Kronecker product which reduces the number of parameters with considerable gain in inference tasks. This paper focuses on the third order tensor normal model or the trilinear normal model.

The trilinear normal model has been studied by several authors. For example, the prob-lem of estimating the Kronecker structured covariance matrix, the so called double separable covariance matrix, has been considered in [14]. The algorithm presented in [14] can be seen as a direct generalization of the algorithm presented in [15] for the estimation of the separable covariance matrix in the matrix normal model. In [16] the so called MLE-3D algorithm was studied as an extension of the MLE-2D algorithm presented in [17].

Most studies on the tensor normal model focused on the estimation of the variance-covariance matrix. Little attention has been paid to the structure of the mean, however, there is a potential to improve the analysis by assuming a structured mean. In this paper, we formally introduce a 2-fold growth curve model by assuming a trilinear structure for the mean in the tensor normal model and propose an algorithm for estimating parameters. Some direct extensions are also explored.

2

The tensor normal distribution with a structured mean

Let X = (xijk) : p × q × r be a random tensor of order three; see Figure 1 for illustration.

✒ k = 1, . . . , r    x11k . . . x1qk .. . . .. ... xp1k . . . xpqk    X =

Figure 1: Visualization of a third order tensor. We define a vectorization of X as vecX = p X i=1 q X j=1 r X k=1 xijke3k⊗ e2j⊗ e1i,

where e1i, e2j and e3k are the unit basis vectors of size p, q and r respectively and ⊗ denotes the Kronecker product. A matricization or unfolding or flattening of a tensor is the process of reordering its elements into a matrix [10]. This can be done in several ways and in this paper we will use the so called mode-n matricization and the notation X(n) will be used to denote

(5)

n-mode matrix from the tensor X . For some details about matricization and decomposition of tensors refer to [10].

A random tensor X : p×q×r is normally distributed with mean tensor M = (µijk) : p×q×r

and covariance matrices Σ : p × p, Ψ : q × q, Ω : r × r, here assumed to be positive definite, if vecX is distributed as a multivariate normal distribution with mean vecM and covariance matrix Ω ⊗ Ψ ⊗ Σ. We write

X ∼ Np,q,r(M , Σ, Ψ, Ω) if vecX ∼ Npqr(vecM , Ω ⊗ Ψ ⊗ Σ). (1) We refer to the model (1) as the third order tensor normal model or the trilinear normal model. The aim of this paper is to introduce a trilinear structure for the mean of the form

µijk= s X ℓ=1 t X m=1 u X n=1 bℓmnaiℓcjmdkn, i = 1, . . . , p, j = 1, . . . , q, k = 1, . . . , r, that is M = B × {A, C, D}, (2)

where B = (bℓmn) : s × t × u is the parameter given as a tensor of order three, A = (aiℓ) : p × s,

C = (cjm) : q × t and D = (dkn) : r × u are known design matrices, and × denotes the Tucker

product, see [10], and it is defined by the identity

vec(B × {A, C, D}) = (D ⊗ C ⊗ A)vecB. This structure is shown in Figure 2.

M =

A B C

D′

Figure 2: Trilinear structure for the mean. B : s × t × u is the parameter given as a tensor of order three and A : p × s, C : q × t and D : r × u are known design matrices.

The model (1) with the mean structure (2) can be written in matrix form using three different modes as

X(1) ∼ Np,qr(AB(1)(D ⊗ C)′, Σ, Ω ⊗ Ψ),

X(2) ∼ Nq,pr(CB(2)(D ⊗ A)′, Ψ, Ω ⊗ Σ),

X(3) ∼ Nr,pq(DB(3)(C ⊗ A)′, Ω, Ψ ⊗ Σ).

The following artificial example illustrates how the model with the mean structure (2) may arise.

(6)

Example 2.1 Assume that one has measured pH in r lakes from u regions at q levels of depth and for p time points. The aim is to investigate how pH varies with depth and/or time and how pH differs across regions. Thus, we have spatio-temporal measurements. Data form a random tensor X : p × q × r, where r = r1+ r2+ · · · + ru and rn is the number of lakes in the

nth region. It is assumed that measurements of each lake (a frontal slice in the tensor X ) is distributed as a matrix normal distribution with covariance matrices Σ : p × p, and Ψ : q × q, and that the measurements of different lakes are independent, that is, Ω = Ir. If the first r1

frontal slices of X represent region one, the next r2 frontal slices represent region two, and

so on, we get the between-individuals design matrix D′ = blockdiag(1′r1, . . . , 1′ru). It is also assumed that the expected trend in time is a polynomial of order s − 1 and that the expected trend in depth is a polynomial of order t − 1. Thus, we have two within-individuals design matrices A=      1 t1 · · · ts−11 1 t2 · · · ts−12 .. . ... . .. ... 1 tp · · · ts−1p      and C =      1 d1 · · · dt−11 1 d2 · · · dt−12 .. . ... . .. ... 1 dq · · · dt−1q     . Hence, the model for the data tensor X is

X = B × {A, C, D} + E , (3) where E ∼ Np,q,r(O, Σ, Ψ, Ir), and O is a tensor of zeros.

The model (3) may be thought as a generalization of the classical growth curve model by [18] in the sense that it considers growth curves in two directions whereas the classical growth curve model considers one direction. In the classical growth curve model we have two design matrices; one models the within-individuals structure whereas the other one models the between-individuals structure. In our extension we have an additional within-individuals design matrix. We may refer to the model (3) as a 2-fold Growth Curve Model and refer to the classical growth curve model as a 1-fold Growth Curve Model. Also, it is worth mentioning that the model (3) generalizes the model studied by [19] which corresponds to u = 1.

In the rest of this paper, for brevity, we assume that the matrices A, C and D are full column rank, that is r(A) = s, r(C) = t and r(D) = u.

3

Maximum likelihood estimation

3.1 Estimation of parameters in the 2-fold Growth Curve Model

The model (3) can be written in matrix form using three different modes as

X(1) ∼ Np,qr(AB(1)(D ⊗ C)′, Σ, Ir⊗ Ψ), (4)

X(2) ∼ Nq,pr(CB(2)(D ⊗ A)′, Ψ, Ir⊗ Σ), (5)

X(3) ∼ Nr,pq(DB(3)(C ⊗ A)′, Ir, Ψ ⊗ Σ). (6)

(7)

Proof. The likelihood function corresponding to (4)-(6) are respectively given by L1 = c|Σ|− qr 2 |Ψ|− pr 2 e− 1

2tr{Σ−1(X(1)−AB(1)(D⊗C)′)(Ir⊗Ψ−1)(X(1)−AB(1)(D⊗C)′)′}, L2 = c|Σ|−

qr 2 |Ψ|−

pr

2 e−12tr{Ψ−1(X(2)−AB(2)(D⊗C)′)(Ir⊗Σ−1)(X(2)−AB(2)(D⊗C)′)′}, L3= c|Σ|− qr 2|Ψ|− pr 2 e− 1 2tr{(X(3)−AB(3)(D⊗C)′)(Ψ−1⊗Σ−1)(X(3)−AB(3)(D⊗C)′)′}, where c is a proportionality constant and tr stands for the trace of a matrix.

To prove that the models (4)-(6) give the same likelihood function, it is enough to show that t1 = t2= t3, where

t1 = tr{Σ−1(X(1)− AB(1)(D ⊗ C)′)(Ir⊗ Ψ−1)(X(1)− AB(1)(D ⊗ C)′)′},

t2 = tr{Ψ−1(X(2)− CB(2)(D ⊗ A)′)(Ir⊗ Σ−1)(X(2)− CB(2)(D ⊗ A)′)′},

t3 = tr{(X(3)− DB(3)(C ⊗ A)′)(Ψ−1⊗ Σ−1)(X(3)− DB(3)(C ⊗ A)′)′}.

First, we observe that

vecX(1)= (Ir⊗ Kq,p)vecX(2) = Kr,pqvecX(3),

vecB(1)= (Iu⊗ Kt,s)vecB(2)= Ku,stvecB(2).

where Kp,q: pq × pq is the commutation matrix.

Next calculations are based on properties about the vec-operator, the trace, the commu-tation matrix and the Kronecker product. We have

vec(X(1)− AB(1)(D ⊗ C)′) = (Ir⊗ Kq,p)vec(X(2)− CB(2)(D ⊗ A)′)

= Kr,pqvec(X(3)− DB(3)(C ⊗ A)′).

and

tr{Σ−1(X(1)− AB(1)(D ⊗ C)′)(Ir⊗ Ψ−1)(X(1)− AB(1)(D ⊗ C)′)′}

= vec′(X(1)− AB(1)(D ⊗ C)′)(Ir⊗ Ψ−1⊗ Σ−1)vec(X(1)− AB(1)(D ⊗ C)′)

= vec′(X(2)− CB(2)(D ⊗ A)′)(Ir⊗ Kp,q(Ψ−1⊗ Σ−1)Kq,p)

× vec(X(2)− CB(2)(D ⊗ A)′)

= vec′(X(2)− CB(2)(D ⊗ A)′)(Ir⊗ Σ−1⊗ Ψ−1)vec(X(2)− CB(2)(D ⊗ A)′)

= tr{Ψ−1(X(2)− CB(2)(D ⊗ A)′)(Ir⊗ Σ−1)(X(2)− CB(2)(D ⊗ A)′)′}.

This proves that t1 = t2. Similar calculations lead to t1= t3 and the proof is complete.

We are going to apply the maximum likelihood approach to find estimators for B, Σ and Ψ. First, we notice that the parameters Ψ and Σ are defined up to a positive multiplicative constant because, for example, Ir ⊗ cΨ ⊗ c−1Σ = Ir ⊗ Ψ ⊗ Σ with c > 0. This issue

has been discussed by some authors. In [17] and in [16] the point have been discussed in terms of stability of the MLE-2D and the MLE-3D algorithms. The algorithm presented in [15] addressed the issue of uniqueness of estimators in the estimation of the separable covariance with an additional restriction and the result has been generalized in [14] for the

(8)

double separable covariance matrix in the third order tensor normal model. In this paper we do not impose such restrictions.

To find estimating equations for parameters we first fix Ψ in (4) and find estimating equations for Σ and B(1). Secondly, we fix Σ in (5) and find estimating equation for Ψ. This procedure is justified by the fact that the models in (4)-(6) give the same likelihood function as shown in Theorem 3.1. The idea is to transform the models (4) and (5) to the classical growth curve model given in [18] and use results in [20] to get estimators of non-fixed parameters and then replace the fixed parameters by their maximum likelihood estimators to get estimating equations. For example, the next theorem gives the estimators of B(1) and Σ for fixed Ψ. Theorem 3.2 For fixed positive definite Ψ, the maximum likelihood estimators of B(1) and

Σin the model (4) are given by

b

B(1) = (A′S−11 A)−1A′S−11 X(1)(D(D′D)−1⊗ Ψ−1C(C′Ψ−1C)−1), S1 = X(1)(Ir⊗ Ψ−1− D(D′D)−D′⊗ Ψ−1C(C′Ψ−1C)−1C′Ψ−1)X′(1),

qr bΣ = (X(1)− A bB(1)(D ⊗ C)′)(Ir⊗ Ψ−1)(X(1)− A bB(1)(D ⊗ C)′)′,

where S1 is assumed to be positive definite.

Proof. Consider the model

X(1) ∼ Np,qr(AB(1)(D ⊗ C)′, Σ, Ir⊗ Ψ).

As Ψ is positive definite, we transform the data by letting Y = X(1)(Ir ⊗ Ψ)−

1 2. Since (D ⊗ C)′(Ir⊗ Ψ)− 1 2 = D′⊗ (C′Ψ− 1

2), the model for the new data is

Y ∼ Np,qr(AB(1)H, Σ, Iqr), (7)

where H = D′⊗ (C′Ψ−12).

The model (7) is the classical growth curve model. From results in [20], the estimators of parameters are

b

B(1) = (A′S−11 A)−1A′S−11 Y H′(HH′)−1 S1 = Y (Iqr− H′(HH′)−1H)Y′,

qr bΣ = (Y − A bB(1)H)(Y − A bB(1)H)′. Coming back to the original data, we get the desired result.

Replacing Ψ with its maximum likelihood estimator bΨ in the above results, we get the estimating equations for B(1) and Σ:

b B(1) = (A′S−11 A)−1A′S−11 X(1)(D(D′D)−1⊗ bΨ−1C(C′Ψb−1C)−1), (8) S1 = X(1)(Ir⊗ bΨ −1 − D(D′D)−D′⊗ bΨ−1C(C′Ψb−1C)−1C′Ψb−1)X′(1), qr bΣ = (X(1)− A bB(1)(D ⊗ C)′)(Ir⊗ bΨ −1 )(X(1)− A bB(1)(D ⊗ C)′)′, (9)

(9)

where S1 is assumed to be positive definite.

Replacing Σ and B(2) with bΣand bB(2) (obtained from bB(1) by a proper rearrangement of

elements) in (5), we write the likelihood function as L = c| bΣ|−qr2 |Ψ|− pr 2 e− 1 2tr{Ψ−1(X(2)−C bB(2)(D⊗A)′)(I⊗ bΣ −1 )(X(2)−C bB(2)(D⊗A)′)′}. Maximizing with respect to Ψ, we get the estimating equation

pr bΨ = (X(2)− C bB(2)(D ⊗ A)′)(Ir⊗ bΣ −1

)(X(2)− C bB(2)(D ⊗ A)′)′. (10) Theorem 3.3 The estimator bB(1)given in (8) is uniquely determined regardless of the parametriza-tion of the covariance parameters.

Proof. It is straightforward to see that when bΨ is replaced with c bΨ for some c > 0, the expression (8) does not change.

The estimating equations for parameters are nested and cannot be solved explicitly. To obtain maximum likelihood estimates of Σ, Ψ and B in the 2-fold growth curve model we propose the following iterative algorithm.

Algorithm 3.4 1. Choose initial solution bΨ= bΨ0;

2. Compute bB(1) using (8); 3. Compute bΣusing (9); 4. Compute bΨusing (10);

5. Repeat steps 2–4 until the convergence criterion is met.

Usually, there is no prior information that may guide to choose the initial solution. Very often the identity matrix would be enough to get the solutions. The convergence criterion may be based on the rate of change in bΨ⊗ bΣand not separately on bΨand bΣ. As mentioned earlier the covariance matrices bΣand bΨare defined up to a scale factor. Only, the Kronecker product

b

Ψ⊗ bΣis uniquely defined. This does not affect bBwhich is uniquely defined, see Theorem 3.3. In Section 4 we will carry out empirical studies on the performance of the Algorithm 3.4 with respect to the choice of initial solution and the sample size.

3.2 Estimation of parameters in the 2-fold Growth Curve Model when Σ has a uniform structure

Now, we assume that Σ in (4)-(6) has a uniform structure, that is Σ= σ2[(1 − ρ)Ip+ ρ1p1′p],

where σ > 0 and −1/(p − 1) < ρ < 1 are unknown.

To find estimating equations for B(1), σ2, ρ and Ψ, we use similar reasoning as in Subsection

3.1. We first fix Ψ in (4) and find estimating equations for B(1), σ2 and ρ. Then we plug in b

(10)

Let Ψ be fixed in the model X(1) ∼ Np,qr(AB(1)(D ⊗ C)′, Σ, Ir⊗ Ψ) and transform the

data as Y = X(1)(Ir⊗ Ψ)−

1

2. Then, the model for the new data is

Y ∼ Np,qr(AB(1)H, Σ, Iqr), (11)

where H = D′⊗ (C′Ψ−12).

Model (11) is the classical growth curve model with a uniform covariance structure. Sup-pose that the constant term is included in the growth function and the design matrix A is of the form A = (1p, A1). Then, the uniform covariance structure is a special case of Rao’s

simple structure (RSS) [21, 22] and we can use results in [23] to get the maximum likelihood estimators of B(1), σ2 and ρ: b B(1) = (A′A)−1A′Y H′(HH′)−1, ˆ σ2 = trS/pqr, ˆ ρ = (1′pS1p− trS)/((p − 1)trS), where S = S1+ (Ip− A(A′A)−1A′)Y H′(HH′)−1HY′(Ip− A(A′A)−1A′), S1 = Y (Iqr− H′(HH′)−H)Y′.

In the original data the estimating equations become b B(1) = (A′A)−1A′X(1)(D(D′D)−1⊗ bΨ−1C(C′Ψb−1C)−1), (12) ˆ σ2 = trS/pqr, (13) ˆ ρ = (1′pS1p− trS)/((p − 1)trS), (14) where S = S1+ S2, S1 = X(1)(Ir⊗ bΨ −1 − D(D′D)−D′⊗ bΨ−1C(C′Ψb−1C)−1C′Ψb−1)X′(1), S2 = (Ip− A(A′A)−1A′) × X(1)(D(D′D)−D′⊗ bΨ−1C(C′Ψb−1C)−1C′Ψb−1)X′(1)(Ip− A(A′A)−1A′),

in which Ψ has been replaced with its maximum likelihood estimator bΨand S is assumed to be positive definite.

Finally, it is clear that the estimating equation for Ψ is given by (10) in which b

Σ= ˆσ2[(1 − ˆρ)Ip+ ˆρ1p1′p]. (15)

As in Subsection 3.1, the maximum likelihood estimators of Σ, Ψ and B are obtained using an iterative algorithm.

(11)

2. Compute bB(1) using (12);

3. Compute bΣusing (15), (13) and (14); 4. Compute bΨusing (10);

5. Repeat steps 2–4 until the convergence criterion is met.

The choice of the initial solution and the convergence criterion is the same as in Subsection 3.1.

3.3 Estimation of B, Σ, Ψ and Ω: General case

Consider the model

X ∼ Np,q,r(B × {A, C, D}, Σ, Ψ, Ω), (16) where B = (bℓmn) : s × t × u is the parameter given as a tensor of order three, and A = (aiℓ) :

p × s, C = (cjm) : q × t and D = (dkn) : r × u are known design matrices.

To find estimators of the parameters in model (16) a sample of N i.i.d. observation tensors Xi∼ Np,q,r(B × {A, C, D}, Σ, Ψ, Ω), i = 1, 2, . . . , N,

is needed. We stack the observation tensors into matrices as follows: X(1) = (X1(1), . . . , XN(1)) : p × qrN,

X(2) = (X1(2), . . . , XN(2)) : q × prN,

X(3) = (X1(3), . . . , XN(3)) : r × pqN.

The distribution of these data matrices are

X(1) ∼ Np,qrN(AB(1)(1N ⊗ D ⊗ C)′, Σ, IN ⊗ Ω ⊗ Ψ), (17)

X(2) ∼ Nq,prN(CB(2)(1N ⊗ D ⊗ A)′, Ψ, IN ⊗ Ω ⊗ Σ), (18)

X(3) ∼ Nr,pqN(DB(3)(1N ⊗ C ⊗ A)′, Ω, IN ⊗ Ψ ⊗ Σ). (19)

Similar as in Theorem 3.1 one can show that the models in (17)-(19) give the same likelihood function. To find estimating equations for parameters we use similar reasoning as in Subsection 3.1. We first fix Ψ and Ω in (17) and find estimating equations for Σ and B(1). Secondly, we fix Σ and Ω in (18) and find estimating equations for Ψ. At last, we fix Ψ and Σ in (19) and find estimating equations for Ω. To do so, transform the models (17)-(19) to the classical growth curve model and use results in [20] to get estimators as in Theorem 3.2, from which we replace the fixed parameters by their maximum likelihood estimators to get estimating equations. For example, letting Z1 = X(1)(IN⊗ Ω ⊗ Ψ)−

1

2. The model (17) becomes Z1 ∼ Np,qrN(AB(1)H1, Σ, IqrN),

(12)

where H1 = 1N⊗ D′Ω−

1

2 ⊗ C′Ψ− 1

2. This is the classical growth curve model from which we get the estimating equations for Σ and B(1):

N bB(1) = (A′S−11 A)−1A′S−11 X(1)((1N ⊗ bΩ −1 D⊗ bΨ−1C)(D′b−1D⊗ CΨb−1C)−1), (20) N qr bΣ= (X(1)− A bB(1)(1N ⊗ D ⊗ C)′)(IN ⊗ bΩ⊗ bΨ)−1(X(1)− A bB(1)(1N ⊗ D ⊗ C)′)′, (21) where S1 = X(1)((IN ⊗ bΩ⊗ bΨ)−1− P1)(X(1))′, P1 = 1 N(1N ⊗ bΩ −1 D⊗ bΨ−1C)(D′Ωb−1D⊗ C′Ψb−1C)−1(1N ⊗ bΩ −1 D⊗ bΨ−1C)′ = 1 N1N1 ′ N ⊗ bΩ −1 D(D′Ωb−1D)−1D′Ωb−1⊗ bΨ−1C(C′Ψb−1C)−1C′Ψb−1, and S1 is assumed to be positive definite.

Similarly, using (18), the estimating equations for Ψ is given by

N pr bΨ= (X(2)− C bB(2)(1N ⊗ D ⊗ A)′)(IN ⊗ bΩ⊗ bΣ)−1(X(2)− C bB(2)(1N ⊗ D ⊗ A)′)′. (22)

Finally, using (19), we get the estimating equations for Ω,

N pq bΩ= (X(3)− D bB(3)(1N⊗ C ⊗ A)′)(IN ⊗ bΨ⊗ bΣ)−1(X(3)− D bB(3)(1N⊗ C ⊗ A)′)′. (23)

Again, the estimating equations cannot be solved explicitly and the following iterative algorithm is proposed to calculate the maximum likelihood estimators of Σ, Ψ, Ω and B in the model (16).

Algorithm 3.6 1. Choose initial solutions bΨ= bΨ0 and bΩ= bΩ0;

2. Compute bB(1) using (20); 3. Compute bΣusing (21); 4. Compute bΨusing (22); 5. Compute busing (23);

6. Repeat steps 2–5 until the convergence criterion is met.

Here the identity matrices may be chosen to be initial solutions and the convergence criterion may be based on the rate of change in bΩ⊗ bΨ⊗ bΣ.

4

Simulation study

The aim of this section is to study through simulations the performance of the procedures proposed in this paper, where we restrict ourselves on the algorithm proposed in Subsection 3.1. In the first place the objective is to study the effect of the choice of initial solutions on the performance of the algorithm. In the second place the objective is to investigate empirically the bias and the dispersion of the estimators with respect to the sample size.

(13)

4.1 Procedures

To study the effect of the choice of initial solutions on the performance of the algorithm, we considered two extreme cases of initial solution, the identity matrix ( bΨ0 = Iq) and the

true parameter ( bΨ0 = Ψ). The average number (rounded to next integer) of iterations

required to achieve a nominal absolute rate of change of order ε = 10−12, ||( bΨ⊗ bΣ) old−

( bΨ⊗ bΣ)new||F < ε, was calculated over 1000 simulation runs for each sample size considered

r ∈ {20, 30, 40, 50, 100, 200, 500, 1000}. Here, || · ||F stands for the Frobenius norm, defined as

kAkF = kvecAk2 and k · k2 is the usual L2−vector norm. At the same time, the empirical bias

and empirical dispersion of the estimators were calculated. The standardized empirical bias for bΨ⊗ bΣwas calculated as

|| bΨ⊗ bΣ− Ψ ⊗ Σ||F

||Ψ ⊗ Σ||F

,

where the bar indicates that the quantity has been averaged over 1000 simulation runs, and the standardized measure of dispersion was calculated as

|| bΨ⊗ bΣ− Ψ ⊗ Σ||F ||Ψ ⊗ Σ||F

.

Similarly, the standardized empirical bias for bB was calculated as || bB− B||F

||B||F

,

and the standardized measure of dispersion was calculated as || bB− B||

F

||B||F

.

In all cases, data was generated randomly from the distribution X ∼ N4,3,r(B × {A, C, D}, Σ, Ψ, Ir), with design matrices

A′=  1 1 1 1 2 3 4 5  ; C′ =  1 1 1 0.5 5.5 10.5  ; D′ = blockdiag(1′r1, 1′r2), r1= r2= r/2; and parameters B(1)=  1 1 3 2 1 2 4 5  ; Σ =     2 1 0.5 2 1 3 −2 0.4 0.5 −2 4 −1 2 0.4 −1 5     ;Ψ=   3 0.5 0.6 0.5 2 0.4 0.6 0.4 1   .

Calculations were performed using MATLAB R2015a, tensor manipulation were done using the Tensor Toolbox [24, 25].

(14)

Table 1: Average number of iterations (ITER), rounded to the next integer, required to achieve a nominal absolute rate of change of order ε = 10−12 in the Kronecker product estimates, the standardized empirical bias and the standardized measure of dispersion of estimators as a function of sample size (r). The upper and lower blocks correspond to the initial solutions taken as identity matrix and true parameter matrix, respectively.

r ITER || bB−B||F ||B||F || bB−B|| F ||B||F || bΨ⊗ bΣ−Ψ⊗Σ||F ||Ψ⊗Σ||F || bΨ⊗ bΣ−Ψ⊗Σ||2 ||Ψ⊗Σ||F 20 101 0.0071 0.1552 0.0371 0.3108 30 80 0.0096 0.1214 0.0258 0.2493 40 72 0.0063 0.1057 0.0162 0.2117 50 69 0.0052 0.0948 0.0121 0.1923 100 66 0.0011 0.0669 0.0091 0.1327 200 64 0.0016 0.0471 0.0041 0.0932 500 63 0.0009 0.0295 0.0026 0.0589 1000 63 0.0001 0.0209 0.0011 0.0420 20 100 0.0025 0.1495 0.0352 0.3107 30 72 0.0051 0.1196 0.0199 0.2491 40 65 0.0056 0.1085 0.0186 0.2161 50 63 0.0015 0.0955 0.0126 0.1920 100 58 0.0020 0.0658 0.0074 0.1346 200 56 0.0033 0.0465 0.0036 0.0937 500 54 0.0011 0.0299 0.0030 0.0592 1000 53 0.0006 0.0214 0.0012 0.0416 4.2 Results

With no surprise, the number of iterations required to attained a nominal convergence criterion decreased as the sample size increased (see, Table 1) whether the initial solution was the identity matrix or the true parameter matrix. The choice of the true parameter matrix slightly decreased the number of iterations required to meet the convergence criterion.

Regarding, the standardized empirical bias and the standardized measure of dispersion for both bBand bΨ⊗ bΣ, there was no remarkable difference between the use of the true parameter matrix as initial solution and the use of the identity matrix. The standardized empirical bias for bB or bΨ⊗ bΣ was small for all sample sizes considered and the standardized measures of dispersion for both bB and bΨ⊗ bΣ exhibited a clear pattern of decrease towards zero as the sample size increased. These results are in agreement with properties of the maximum likelihood estimators and the conclusion is that our estimators are consistent.

Acknowledgements

The research of Joseph Nzabanita has been supported through the UR-Sweden Program for Research, Higher Learning and Institution Advancement and all persons and institutions in-volved are hereby acknowledged.

(15)

References

[1] C. J. Appellof and E. Davidson, “Strategies for analyzing data from video fluorometric monitoring of liquid chromatographic effluents,” Analytical Chemistry, vol. 53, no. 13, pp. 2053–2056, 1981.

[2] R. Bro, “Review on multiway analysis in chemistry-2000–2005,” Critical reviews in

ana-lytical chemistry, vol. 36, no. 3-4, pp. 279–293, 2006.

[3] A. Smilde, R. Bro, and P. Geladi, Multi-way analysis: applications in the chemical

sci-ences. John Wiley & Sons, 2005.

[4] J. D. Carroll and J.-J. Chang, “Analysis of individual differences in multidimensional scaling via an n-way generalization of ”eckart-young” decomposition,” Psychometrika, vol. 35, no. 3, pp. 283–319, 1970.

[5] H. A. Kiers and I. V. Mechelen, “Three-way component analysis: Principles and illustra-tive application,” Psychological Methods, vol. 6, no. 1, p. 84, 2001.

[6] A. Cichocki, D. Mandic, A. Phan, C. Caiafa, G. Zhou, Q. Zhao, and L. De Lathauwer, “Tensor decompositions for signal processing applications from two-way to multiway com-ponent analysis,” arXiv preprint arXiv:1403.4462, 2014.

[7] N. D. Sidiropoulos, R. Bro, and G. B. Giannakis, “Parallel factor analysis in sensor array processing,” Signal Processing, IEEE Transactions on, vol. 48, no. 8, pp. 2377–2388, 2000. [8] L. De Lathauwer and B. De Moor, “From matrix to tensor: Multilinear algebra and signal processing,” in Institute of Mathematics and its Applications Conference Series, vol. 67, pp. 1–16, Citeseer, 1998.

[9] M. A. O. Vasilescu and D. Terzopoulos, “Multilinear subspace analysis of image en-sembles,” in Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE

Computer Society Conference on, vol. 2, pp. II–93, IEEE, 2003.

[10] T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” SIAM review, vol. 51, no. 3, pp. 455–500, 2009.

[11] D. Akdemir and A. K. Gupta, “Array variate random variables with multiway Kronecker delta covariance matrix structure,” Journal of Algebraic Statistics, vol. 2, no. 1, pp. 98– 113, 2011.

[12] P. D. Hoff, “Separable covariance arrays via the tucker product, with applications to multivariate relational data,” Bayesian Analysis, vol. 6, no. 2, pp. 179–196, 2011.

[13] M. Ohlson, M. R. Ahmad, and D. von Rosen, “The multilinear normal distribution: Intro-duction and some basic properties,” Journal of Multivariate Analysis, vol. 113, pp. 37–47, 2013.

(16)

[14] M. Singull, M. R. Ahmad, and D. von Rosen, “More on the Kronecker structured co-variance matrix,” Communications in Statistics-Theory and Methods, vol. 41, no. 13-14, pp. 2512–2523, 2012.

[15] M. S. Srivastava, T. von Rosen, and D. von Rosen, “Models with a Kronecker product covariance structure: estimation and testing,” Mathematical Methods of Statistics, vol. 17, no. 4, pp. 357–370, 2008.

[16] A. M. Manceur and P. Dutilleul, “Maximum likelihood estimation for the tensor nor-mal distribution: Algorithm, minimum sample size, and empirical bias and dispersion,”

Journal of Computational and Applied Mathematics, vol. 239, pp. 37–49, 2013.

[17] P. Dutilleul, “The MLE algorithm for the matrix normal distribution,” Journal of

Statis-tical Computation and Simulation, vol. 64, no. 2, pp. 105–123, 1999.

[18] R. F. Pothoff and S. N. Roy, “A generalized multivariate analysis of variance model useful especially for growth curve problems,” Biometrika, vol. 51, pp. 313–326, 1964.

[19] M. S. Srivastava, T. von Rosen, and D. von Rosen, “Estimation and testing in general multivariate linear models with Kronecker product covariance structure,” Sankhy¯a: The Indian Journal of Statistics, vol. 71-A, pp. 137–163, 2009.

[20] T. Kollo and D. von Rosen, Advanced Multivariate Statistics with Matrices. Dordrecht, The Netherlands: Springer, 2005.

[21] J. C. Lee and S. Geisser, “Growth curve prediction,” Sankhy¯a: The Indian Journal of Statistics, Series A, vol. 34, no. 4, pp. 393–412, 1972.

[22] C. R. Rao, “Least squares theory using an estimated dispersion matrix and its application to measurement of signals,” in Proceedings of the Fifth Berkeley Symposium on

Mathe-matical Statistics and Probability, vol. 1, pp. 355–372, Univ. of California Press Berkeley,

1967.

[23] J. C. Lee, “Prediction and estimation of growth curves with special covariance structures,”

Journal of the American Statistical Association, vol. 83, no. 402, pp. 432–440, 1988.

[24] B. W. Bader, T. G. Kolda, et al., “Matlab tensor toolbox version 2.6.” Available online: http://www.sandia.gov/ tgkolda/TensorToolbox/, February 2015.

[25] B. W. Bader and T. G. Kolda, “Algorithm 862: MATLAB tensor classes for fast algo-rithm prototyping,” ACM Transactions on Mathematical Software, vol. 32, pp. 635–653, December 2006.

References

Related documents

Using the Bode’s integral theorem it is shown that if the system has relative degree greater or equal to one, then a CITE or traditional causal ILC algorithm cannot be designed to

The Standard 2 is about specifying the expected learning outcomes, in terms of knowledge and skills, of an education to prepare the students for the intended professional role,

Against this background, this thesis has utilized a visual and textual discourse analysis to investigate memory books published by the Swedish Armed Forces, which detail the pres-

Vi skulle dock även kunna argumentera för att risken kan uppstå vid tolkningen av den Formaliserade su- metoden då respondent 3 nämner att Scrum enligt skaparna ska

Public data stream services provided by different entities to support route optimization, in- cluding weather, ice conditions, Maritime Safety Information (MSI), Maritime Spatially

A method is presented that models a shock tube with respect to pressure step amplitudes, maximum dwell-time and also including thin boundary layer theory. It is a part of a

Det är genom feminismens fokus på genus och makt som det blir möjligt att förstå hur feministisk aktivism på sociala medier förhåller sig till hegemoniska diskurser i samhället

Vår studie uppmärksammar hur elever i läs- och skrivsvårigheter och dyslexi upplever motivation som en del i det egna lärandet och ambitionen är att kunskapen ska leda till