• No results found

Minimax Confidence Intervals for Pointwise Nonparametric Regression Estimation

N/A
N/A
Protected

Academic year: 2021

Share "Minimax Confidence Intervals for Pointwise Nonparametric Regression Estimation"

Copied!
8
0
0

Loading.... (view fulltext now)

Full text

(1)

Technical report from Automatic Control at Linköpings universitet

Minimax Confidence Intervals for

Pointwise Nonparametric Regression

Estimation

Anatoli Judisky, Alexander Nazin, Lennart Ljung, Arkadi

Nemirovski

Division of Automatic Control

E-mail: Anatoli.Juditski@imag.fr, nazin@ipu.rssi.ru,

ljung@isy.liu.se, nemirovs@isye.gatech.edu@isy.liu.se

13th May 2009

Report no.: LiTH-ISY-R-2908

Accepted for publication in 15th IFAC Symposium on System

Identifi-cation, Saint-Malo, France

Address:

Department of Electrical Engineering Linköpings universitet

SE-581 83 Linköping, Sweden

WWW: http://www.control.isy.liu.se

AUTOMATIC CONTROL REGLERTEKNIK LINKÖPINGS UNIVERSITET

Technical reports from the Automatic Control group in Linköping are available from http://www.control.isy.liu.se/publications.

(2)

Abstract

We address a problem of estimation of an unknown regression function f at a given point x0from noisy observations yi= f (xi) + ei, ; i = 1, ..., n.Here

xi∈ Rk are observable regressors and (ei) are normal i.i.d. (unobservable)

disturbances. The problem is analyzed in the minimax framework, namely, we suppose that f belongs to some functional class F , such that its nite-dimensional cut Fn= {f (xi), f ∈ F, i = 0, ..., n, }is a convex compact set.

For an arbitrary xed regression plan Xn= (x1; ...; xn)we study minimax

on Fn condence intervals of ane estimators and constructan estimator

which attains the minimax performance on the class of arbitrary estimators when the condence level approaches 1.

Keywords: non-parametric regression, linear estimation, non-parametric identication, convex programming, identication algorithms

(3)

Minimax Confidence Intervals for Pointwize

Nonparametric Regression Estimation

Anatoli Juditsky∗ Alexander Nazin∗∗ Lennart Ljung∗∗∗

Arkadi Nemirovski∗∗∗∗ ∗

LJK, BP 53, F-38041 Grenoble Cedex 9, France Email: anatoli.juditsky@imag.fr

∗∗

Institute of Control Sciences, Russian Acad. Sci., Profsoyuznaya str., 65, 117997 Moscow, Russia

Email: nazine@ipu.rssi.ru

∗∗∗Div. of Automatic Control, Linkoping University

SE-58183 Linkoping, Sweden Email: ljung@isy.liu.se

∗∗∗∗School of ISyE, Georgia Institute of Technology

Atlanta, Georgia 30332, USA Email: nemirovs@isye.gatech.edu

Abstract:We address a problem of estimation of an unknown regression function f at a given point x0 from noisy observations yi = f (xi) + ei, ; i = 1, ..., n. Here xi ∈ Rk are observable

regressors and (ei) are normal i.i.d. (unobservable) disturbances. The problem is analyzed in

the minimax framework, namely, we suppose that f belongs to some functional class F, such that its finite-dimensional cut Fn = {f(xi), f ∈ F, i = 0, ..., n, } is a convex compact set.

For an arbitrary fixed regression plan Xn = (x1; ...; xn) we study minimax on Fn confidence

intervals of affine estimators and construct an estimator which attains the minimax performance on the class of arbitrary estimators when the confidence level approaches 1.

Keywords: non-parametric regression, linear estimation, non-parametric identification, convex programming, identification algorithms

1. INTRODUCTION

Efficient non-parametric estimation of a regression func-tion f : Rk → R based on its noisy observations

yi= f (xi) + ei, i = 1, . . . , n (1)

at some given design points xi ∈ Rk is one of the

basic problems for many applications including non-linear system identification, see, e.g., Ljung [1999]. Here the random noises ei are supposed to be normal i.i.d. with

Eei = 0, Ee2i = σ2, σ > 0. We consider the problem of

estimating f at a given point x0∈ Rk.

Recently, the Direct Weight Optimization (DWO) method has been proposed to solve a problem of such kind and its properties has been studied in the case when the unknown function f is continuously differentiable with Lipschitz continuous derivative having Roll et al. [2002, 2005]. Then in Juditsky et al. [2004] an adaptive version of the DWO procedure has be proposed which uses a data-driven choice of the corresponding Lipschitz constant. To be more precise, the DWO estimator bfn is a linear

estimator of the form b fn(x0) = n X i=1 ϕiyi,

where the vector ϕ = (ϕ1; ...; ϕn) of weights is computed

to minimize the Maximal Mean Square Error

Rn( bfn, F) = sup f ∈F M SE( bfn, f ) = sup f ∈F Ef  b fn(x0) − f(x0) 2 |Xn 

over a given functional class F; Xn= (x1, ..., xn).

Observe that the value f (x0) to be estimated can be

seen as a linear functional of unknown “signal” f . The classical problem of estimation of linear functionals has received much attention in the statistical literature – some important references here are, for instance, Ibragimov, Khas’minskij [1984], Donoho, Liu [1987], Donoho, Low [1992], Donoho [1995], Cai, Low [2003]. Some important results for this problem have been obtained in the white-noise model. In particular, Donoho [1995] has shown that when it is known a priori that unknown signal belongs to a convex compact class, for several risk measures, the

minimax affine estimator attains the risk bounds which

coinside (up to an absolute constant) with the minimax risk. Recently in Juditsky, Nemirovski [2008] analogous results for different observation models were obtained. Our objective here is to study the implications of the theory of

minimax affine estimation in the regression problem (1).

In the current paper the precision of an estimator bf of f (x0) is measured with the length of the confidence

interval for f (x0) at a given level of confidence. I.e., given

(4)

estimate at f by ℓǫ( bf ; f ) = inf h δ : Pf n | bf − f(x0)| > δ o < ǫi, where, with some abuse of notation, we denote by Pf the

distribution of observations y = (y1; ...yn) associated with

signal f and regressors Xn = (x1, ..., xn).

We consider the estimation problem in the minimax con-text, i.e. we suppose that the a priori information about unknown f amounts to the fact that f belongs to some functional class F.

The maximal risk of the estimator bf is given with Rǫ( bf ; F) = inf " δ : sup f ∈F Pf n | bf − f(x0)| > δ o < ǫ # . Then the minimax optimal ǫ-risk is

R∗

ǫ(F) = inf

b

f

Rǫ( bf ; F),

where the infimum on the RHS is taken over all measur-able estimation functions bf = bf (Xn, y). We say that an

estimator bf = bfϕ is affine if it is of the form

b fϕ= n X i=1 ϕiyi+ ϕ0= ϕTy + ϕ0,

where ϕ = (ϕ1; ...; ϕn). We denote with RAǫ the best

ǫ-risk which can be attained by an affine estimator: RAǫ(F) = inf

ϕ,ϕ0R

ǫ( bfϕ; F).

As we shall see later (cf Section 4), in the case when the functional class is symmetric with respect to the origin, the affine estimator reduces to the linear one.

The rest of the paper is organized as follows: we formulate the problem statement in the next section, then we present the main result on the properties of affine regression estimators in Section 3 and give the construction of the minimax affine estimator in Section 4. Finally, the theorem proof is provided in the appendix.

2. PROBLEM STATEMENT

Observe that our estimation problem is essentially (n + 1)-dimensional. Indeed, given the fixed design Xn =

(x1, ..., xn) ∈ Rk×n and x0∈ Rk, denote u = u(f, Xn) ≡ (f(x1); ...; f (xn)) and ¯ u = ¯u(f, Xn, x 0) ≡ (f(x0); u).

The following assumption is of primary importance here:

Assumption 1. We suppose that the functional class F

is such that its cross-section

Fn = {¯u ∈ E| f ∈ F}

is a convex and compact set in E, E being some finite-dimensional Euclidean space.

Note that Assumption 1 holds true for a wide range of nonparametric regression problems. For instance, it clearly holds for the functional classes in the examples below.

Holder class Fx0(ℓ + 1, L) . Let ℓ ∈ N, and F be the set of

function f : Rk → R such that for any x ∈ Rk,

|f(x) − Pℓ(x0, x)| ≤

L

(ℓ + 1)!|x − x0|

ℓ+1,

where Pℓ(x0, ·) is the Taylor polynomial of order ℓ (here

|a| stands for the Euclidean norm of a), with the extra constraints

f(j)(x0) ∈ SL(j), j = 0, ..., ℓ,

where SL(j) are some convex compact sets of Rkj

.

The simplest example of such a class is the Lipschitz class Fx0(1, L) such that

|f(x) − f(x0)| ≤ L|x − x0|, |f(x0)| ≤ L.

Another example is supplied by the class Fx0(2, L)

of bounded regular functions with Lipschitz-continuous derivative: |f(x) − f(x0) − f′(x0)T(x − x0)| ≤ L 2|x − x0| 2, |f(x0)| ≤ L, |f′(x0)| ≤ L. (2) It is obvious that any cross-section Fn with finite xi,

i = 1, ..., n, is a convex compact set of Rn+1. Further, this

set is efficiently computationally tractable, conditioned that so are the sets S(j)L , j = 0, ..., ℓ1.

Finite-dimensional Sobolev class WM(s, L) . Suppose that

for some family of M bounded functions ψ0, ..., ψM, ψj :

Rk → R, M < ∞, and any x ∈ Rk we have the representation f (x) = M X j=0 ψj(x)βj,

where β = (β0; ...; βM) ∈ BM, BM being a convex and

compact subset of RM+1. Then, clearly,

Fn=   u ∈ R¯ n+1 | ui= M X j=0 ψj(xi)βj, i = 0, ..., nβ ∈ BM    is a convex, compact set.

Class M (L) of bounded monotone functions . Let F be a

class of functions which are monotone increasing on R such that |f(x)| ≤ L for any x ∈ R. Let x(0); ...; x(n) be an

increasing rearrangement of Xn. Then

Fn= {u ∈ Rn+1| u(0) ≤ u(1)≤ ... ≤ u(n); |ui| ≤ L}

is a compact convex set.

A reader can easily produce other examples of functional classes which ensures that the corresponding Fn are

con-vex and compact subsets of Rn for any finite (n + 1)-uple

x0, ..., xn. Observe also that more sets of the sort can be

obtained as intersections of such sets, etc. 3. MAIN RESULT We have the following result:

Theorem 1. Suppose that Assumption 1 holds. Then for

any ǫ ∈ (0, 1/2] one can point out an affine estimator bfǫ

satisfying the relation

1 For details on computational tractability and complexity issues in

Convex Optimization, see, e.g., [Ben-Tal, Nemirovski, 2001, Chapter 4]. A reader not familiar with this area will not lose much when interpreting a computationally tractable convex set as a set given by a finite system of inequalities pi(x) ≤ 0, i = 1, ..., m, where pi(x) are

(5)

RAǫ( bfǫ; F) ≤

erfinv(ǫ/2) erfinv(ǫ) R

ǫ(F), (3)

where erfinv(ǫ) is the inverse of the error function erf(s) = √1 2π ∞ Z s exp{−s2/2}ds. Therefore, erfinv(ǫ/2)/ erfinv(ǫ) → 1 as ǫ → +0.

In other words, for small tolerance levels ǫ, the risk of the best affine estimator becomes arbitrary close to the exact minimax risk in our setup:

lim ǫ→+0 RAǫ( bfϕ; F) R∗ ǫ(F) = 1. (4)

4. CONSTRUCTION OF THE MINIMAX AFFINE ESTIMATOR

In order to compute the affine estimator bfǫ = ˜ϕTy + ˜ϕ0

which attains the bound (3) one can act as follows:

1) Let ¯u = (u0; u) and ¯v = (v0; v). Solve the optimization

problem max ¯ u,¯v∈Fn 1 2(v0− u0) subject to g(u, v) ≡ |u − v| − 2σ erfinv(ǫ/2) ≤ 0. (5) By theory of convex programming an optimal solution (˜u, ˜v) to this problem can be augmented by Lagrange multiplier ν ≥ 0 such that the vectors

eu˜= ∂ ∂ ¯u (¯u,¯v)=(˜u,˜v) 1 2(u0− v0) + νg(u, v)  =  1 2; ν u − v |u − v|  , ev˜= ∂ ∂¯v (¯u,¯v)=(˜u,˜v) 1 2(u0− v0) + νg(u, v)  =  −12; ν u − v |u − v|  ,

belong to normal cones of Fn at the points ˜u, ˜v,

respec-tively: for any ¯u, ¯v ∈ Fn

eTu¯(¯u − ˜u) ≥ 0, eTv¯(¯v − ˜v) ≥ 0, (6)

and

νg(˜u, ˜v) = 0.

2) There are two possible cases: (a) ν = 0 and (b) ν > 0.

In the case of (a), (6) implies that (˜u, ˜v) is an optimal solution to the problem obtained from (5) by eliminating the constraint g(¯u, ¯v) ≤ 0, that is, ˜v0is the maximum, and

˜

u0 is the minimum of u0 on Fn. In this case, an estimate

which reproduces u0, ¯u ∈ Fn, with the risk 12[˜v0− ˜u0] is

trivial – this is the constant estimate b

fǫ= ˜ϕ0=12[˜v0+ ˜u0].

Note that this case indeed takes place when ǫ is small enough, that is, given the number of observations, our re-liability requirement is too strong to allow for a nontrivial estimation.

Now assume that (b) is realized, i.e. ν > 0. In this case, our estimation is b fǫ= ˜ϕTy + ˜ϕ0 where ˜ ϕ = ν σ erfinv(ǫ/2)(˜v − ˜u) = 2ν (˜v − ˜u) |(˜v − ˜u)|, (7) and ˜ ϕ0= 12[˜v0+ ˜u0] −12ϕ˜T(˜v + ˜u). (8)

(note that we can use the definitions (7) and (8) for ˜ϕ and ˜

ϕ0for ν ≥ 0).

Example. Let us consider the simple case when unknown

function f belongs to the class Fx0(2, L) of regular

func-tions with Lipschitz-continuous gradient (2). In this case the set Fnis defined by the system of n + 1 affine and one

quadratic inequalities. Indeed, ¯u = (u0; u1; ...; un) ∈ Fn iff

there is w ∈ Rk such that (¯u, w) ∈ ∆, where

∆ ≡ 

|ui− u0− (xi− x0)Tw| ≤ 12L |xi− x0|2,

for i = 1, .., n, and |u0| ≤ L, |w| ≤ L.

The set Fn is symmetric, thus the constant term ˜ϕ0

vanishes and the constraint ¯u, ¯v ∈ Fn in the problem (5)

is equivalent to ¯d ≡ 12(¯u − ¯v) ∈ Fn. Further, in the case

in question (5) is a convex conic problem with a quadratic constraint:

max

¯

d,w d0 subject to

|d| ≤ 2σ erfinv(ǫ/2), ( ¯d, w) ∈ ∆. Let f : R → R be the quadratic function

f (x) = 4x2+ 5x + 5.

We are to recover f on the regular grid X = {0 : 0.1 : 2} ⊂ [0, 2] given n = 20 observations

yi= f (xi) + ei, i = 1, . . . , n,

where xi ≥ 0 i = 1, ..., n are from the standard normal

distribution, and ei ∼ N(0, 2). A typical recovery for

ǫ = 0.01 is presented on Fig. 1. On Fig. 2 we present the optimal weights ϕi, i = 1, ..., 20 for estimation of f (0); on

Fig. 3 the corresponding weights for estimation of f (1) are plotted. As it was already noticed in Roll [2003], in this case the weights ϕi depend piece-wise quadratically on x.

Let us consider now a simple bi-variate quadratic function f : R2→ R as follows:

f (x) = 5|x|2+ (5; 10)Tx + 15

(c.f, Roll [2003], p. 98). Suppose that we want to recover f on the regular 6 × 6 grid X = {0 : 0.4 : 2}2 ⊂ [0, 2]2

given n = 20 noisy observations taken at the points xi

i = 1, ..., n in the positive quadrant, sampled from the standard normal distribution on R2, and the independent

disturbances ei satisfy ei∼ N(0, 2).

A typical result of minimax linear recovery tuned for ǫ = 0.01 is presented on Fig. 4.

(6)

0 0.5 1 1.5 2 5 10 15 20 25 30

Fig. 1. Recovery of f from n = 20 observations: solid line – true signal, + represent the observations, dashed line – minimax linear recovery, dotted lines – upper and lower bounds of confidence intervals for f (x).

0 0.5 1 1.5 2 2.5 −0.1 −0.05 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35

Fig. 2. Optimal weights ˜ϕi, i = 1, ..., 20 for estimation of

f (0).

REFERENCES

Ben-Tal, A., Nemirovski, A., Lectures on Modern

Con-vex Optimization: Analysis, Algorithms and Engineer-ing Applications, MPS-SIAM Series on Optimization, SIAM, Philadelphia, 2001.

Cai, T., Low, M., A note on nonparametric estimation of linear functionals, The Annals of Statistics 31, 1140– 1153, 2003.

Donoho, D., Liu, R., Geometrizing rates of convergence,

I, Technical Report 137a, Dept. of Stat., University of

California, Berkeley, 1987.

Donoho, D., Low, M., Renormalization Exponents and Optimal Pointwise Rates of Convergence, The Annals

of Statistics 20:2, 944–970, 1992.

Donoho, D., Statistical estimation and optimal recovery.

The Annals of Statistics 22:1, 238–270, 1995.

0 0.5 1 1.5 2 2.5 0.01 0.02 0.03 0.04 0.05 0.06 0.07

Fig. 3. Optimal weights ˜ϕi, i = 1, ..., 20 for estimation of

f (1). 0 1 2 0 0.5 1 1.5 2 2.5 3 10 20 30 40 50 60 70 80 90

Fig. 4. Recovery of bivariate f from n = 20 observations: ·-plot – true signal, +-plot – observations, ⋆-plot – minimax linear recovery.

Ibragimov, I.A., Khas’minskij, R.Z., On the nonparametric estimation of a value of a linear functional in Gaussian white noise. (Russian. English summary) Teor.

Veroy-atn. Primen. 29:1, 19–32, 1984.

Juditsky, A., Nazin, A., Roll, J., and Ljung, L., Adaptive DWO estimator of a regression function. In: NOLCOS

2004, Stuttgart, September 2004.

Juditsky, A.B., Nemirovski, A.S. Nonparametric estima-tion by convex programming, Annals of Stat., to appear, 2008.

Ljung, L., System Identification. Theory For the User. 2nd

ed. Prentice Hall, Upper Saddle River, N.J., 1999.

Roll, J., Nazin, A., and Ljung, L., A non-asymptotic ap-proach to local modelling. In: The 41st IEEE Conference

on Decision and Control, 638–643, 2002.

Roll, J., Nazin, A., and Ljung, L., Nonlinear system identi-fication via direct weight optimization. Automatica:

(7)

Spe-cial Issue on Data-Based Modelling and System Identi-fication 41(3), 475–490, 2005.

Stark, P.B., Affine minimax confidence intervals for a bounded normal mean. Statist. Probab. Lett., 13, 39– 44, 1992.

Roll, J., Local and Piecewise Affine Approaches to System

Identification. Linkoping studies in science and

technol-ogy. PhD thesis no 802, April 2003. 5. APPENDIX We present here the proof of Theorem 1.

For the sake of conciseness we use the notation u0 = f (x0) and

ui= f (xi), i = 1, ..., n. Then for an affine estimatorbu = ϕTy + ϕ0

of u0,

b

u − u0= ϕTy − u0+ ϕ0+ σϕTe

= ϕTu − u0+ ϕ0+ σ|ϕ|η,

where η ∼ N (0, 1), and we have for any ¯u = (u0; u) ∈ Fn:

Pu¯  b u − u0≥ max ¯ u∈Fn [ϕTu − u0] + ϕ0+ σ|ϕ| erfinv(ǫ/2)  ≤ ǫ/2. In the same way we get

Pu¯  b u − u0≤ min ¯ u∈Fn [ϕTu − u0] + ϕ0− σ|ϕ| erfinv(ǫ/2)  ≤ ǫ/2. Let us denote Ψ(ϕ; u, v; ǫ) = ϕT(u − v) + (v0− u0) +2σ|ϕ| erfinv(ǫ/2). (9) If we set ϕ0=1 2  max ¯ v∈Fn [v0− ϕTv] − max ¯ u∈Fn [ϕTu − u0]  ,

then for any ϕ ∈ Rnthe R-risk of estimationbu,

Rǫ(bu; Fn) = inf  δ : sup ¯ u∈Fn Pu¯  |bu − u0| > δ < ǫ  , is bounded with S(ϕ; ǫ) where

2S(ϕ; ǫ) = max ¯ u,¯v∈Fn Ψ(ϕ; ¯u, ¯v; ǫ). Let now S(ǫ) = inf

ϕ S(ϕ; ǫ) = infϕ u,¯¯maxv∈Fn

1

2Ψ(ϕ; ¯u, ¯v; ǫ). (10) Observe that Ψ(ϕ; ¯u, ¯v; ǫ) is linear in ¯u, ¯v and convex in ϕ. Further, Fnbeing compact, one can exchange min and max in (10), so that

S(ǫ) = max ¯ u,¯v∈Fn inf ϕ 1 2Ψ(ϕ; ¯u, ¯v; ǫ).

The minimization with respect to ϕ is immediate: the infimum is −∞ if |u − v| > 2σ erfinv(ǫ/2), and is 12(v0 − u0) when |u −

v| ≤ 2σ erfinv(ǫ/2). Thus S(ǫ) = max ¯ u,¯v∈Fn 1 2(v0− u0) subject to g(¯u, ¯v) ≡ |u − v| − 2σ erfinv(ǫ/2) ≤ 0. (11) Note that this problem is solvable. Let us denote ˜u = (˜u0; ...; ˜un)

and ˜v = (˜v0; ...; ˜vn) an optimal solution to (11). The lower bound for

the ǫ-risk for this one-dimensional subproblem can be easily obtained using a simple 2-point test problem (cf., for instance, Stark [1992]). We present its proof here for the sake of completeness.

Suppose that for some κ > 0 and ǫ ≤ 1/2 Risk∗ǫ(F ) + κ ≤

erfinv(ǫ)

erfinv(ǫ/2)S(ǫ). (12) Let ε = 2ǫ. Observe that S(ε) is a positive and concave function of γ = erfinv(ε/2) (by (9) it is a minimum of affine in γ functions). As S(1) = 0, for any 0 < ǫ ≤ υ ≤ 1,

S(υ) erfinv(υ/2)≥

S(ǫ) erfinv(ǫ/2), and S(ε) ≥ erfinverfinv(ǫ)

(ǫ/2)S(ǫ).

Let now ˜u and ˜v be an optimal solution to (11). Note that (12) implies, by definition of ǫ-risk, the existence of ǫ′< ǫ and of estimate

b u such that sup ¯ u∈{˜u,˜v} Pu¯(|bu − u0| ≥ Risk∗ǫ(F ) + κ/2) ≤ ǫ′.

Consider the test

ψ(y) = 1l bu ≥1

2[˜u0+ ˜v0]



of the hypothesis ¯u = ˜u against the alternative ¯u = ˜v. For this test we have

Pu˜(ψ(y) = 1) + P˜v(ψ(y) = 0) ≤ 2ǫ′< 2ǫ. (13)

On the other hand, ˜u and ˜v satisfy

|˜u − ˜v| ≤ 2σ erfinv(ε/2).

It is obvious that the optimal risk of (the sum of probabilities of errors) any test of ˜u against ˜v is not less than ε = 2ǫ, what contradicts (13).

(8)

Avdelning, Institution Division, Department

Division of Automatic Control Department of Electrical Engineering

Datum Date 2009-05-13 Språk Language  Svenska/Swedish  Engelska/English   Rapporttyp Report category  Licentiatavhandling  Examensarbete  C-uppsats  D-uppsats  Övrig rapport  

URL för elektronisk version http://www.control.isy.liu.se

ISBN  ISRN



Serietitel och serienummer

Title of series, numbering ISSN1400-3902

LiTH-ISY-R-2908

Titel

Title Minimax Condence Intervals for Pointwise Nonparametric Regression Estimation

Författare

Author Anatoli Judisky, Alexander Nazin, Lennart Ljung, Arkadi Nemirovski Sammanfattning

Abstract

We address a problem of estimation of an unknown regression function f at a given point x0

from noisy observations yi= f (xi) + ei, ; i = 1, ..., n.Here xi∈ Rkare observable regressors

and (ei) are normal i.i.d. (unobservable) disturbances. The problem is analyzed in the

minimax framework, namely, we suppose that f belongs to some functional class F , such that its nite-dimensional cut Fn= {f (xi), f ∈ F, i = 0, ..., n, }is a convex compact set. For an

arbitrary xed regression plan Xn= (x1; ...; xn)we study minimax on Fncondence intervals

of ane estimators and constructan estimator which attains the minimax performance on the class of arbitrary estimators when the condence level approaches 1.

Nyckelord

Keywords non-parametric regression, linear estimation, non-parametric identication, convex program-ming, identication algorithms

References

Related documents

Omvendt er projektet ikke blevet forsinket af klager mv., som det potentielt kunne have været, fordi det danske plan- og reguleringssystem er indrettet til at afværge

I Team Finlands nätverksliknande struktur betonas strävan till samarbete mellan den nationella och lokala nivån och sektorexpertis för att locka investeringar till Finland.. För

Generally, a transition from primary raw materials to recycled materials, along with a change to renewable energy, are the most important actions to reduce greenhouse gas emissions

För att uppskatta den totala effekten av reformerna måste dock hänsyn tas till såväl samt- liga priseffekter som sammansättningseffekter, till följd av ökad försäljningsandel

Från den teoretiska modellen vet vi att när det finns två budgivare på marknaden, och marknadsandelen för månadens vara ökar, så leder detta till lägre

40 Så kallad gold- plating, att gå längre än vad EU-lagstiftningen egentligen kräver, förkommer i viss utsträckning enligt underökningen Regelindikator som genomförts

Regioner med en omfattande varuproduktion hade också en tydlig tendens att ha den starkaste nedgången i bruttoregionproduktionen (BRP) under krisåret 2009. De

Generella styrmedel kan ha varit mindre verksamma än man har trott De generella styrmedlen, till skillnad från de specifika styrmedlen, har kommit att användas i större