• Non ci sono risultati.

Large Covariance Matrix Estimation by Composite Minimization

N/A
N/A
Protected

Academic year: 2021

Condividi "Large Covariance Matrix Estimation by Composite Minimization"

Copied!
179
0
0

Testo completo

(1)

A

A

l

l

m

m

a

a

M

M

a

a

t

t

e

e

r

r

S

S

t

t

u

u

d

d

i

i

o

o

r

r

u

u

m

m

U

U

n

n

i

i

v

v

e

e

r

r

s

s

i

i

t

t

à

à

d

d

i

i

B

B

o

o

l

l

o

o

g

g

n

n

a

a

DOTTORATO DI RICERCA IN

Scienze Statistiche

Ciclo XXVIII

Settore Concorsuale di afferenza: 13/D1

Settore Scientifico disciplinare: SECS/S-01

TITOLO TESI

Large covariance matrix estimation

by composite minimization

Presentata da:

Matteo Farnè

Coordinatore Dottorato

Relatore

Prof.ssa Alessandra Luati

Prof.ssa Angela Montanari

(2)
(3)

by omposite minimization

Matteo Farnè

Dipartimentodi S ienze Statisti he Università diBologna

(4)

"Illi autem o to ursus, in quibus eadem visest duorum, septem e iunt distin tos intervallissonos, qui numerus rerum omniumfere nodusest; quod do ti homines nervis imitatiatque antibusaperuerunt sibi reditum in hun lo um, si utalii, qui praestantibus ingeniis in vita humanadivina studia oluerunt." Mar us Tullius Ci ero SomniumS ipionis, 6.18

(5)

At the end of a parti ularly ex iting and laborious work, I feel the need to address my thanks to the people who a ompanied and helped me to omplete su ha huge eort.

First ofall,I desireto expressmygratitude toProfessor Angela Monta-nari, mysupervisorand mentor, for the ontinuous supportand en ourage-ment, for the valuable advi es and suggestions, for her passionand for the unique han esofgrowthandmaturation sheoeredmea rossallmystudy path.

IwanttothankProfessorTrevorHastie,forthebriefbutde isiveresear h period spent at Stanford, whi h gave me the ne essary insights to develop my own theoryand to ee tively provide original solutions to my resear h problem.

I have to thank Dr Yuan Liao for helpful dis ussions and pre ious ref-eren es, Professor Roberto Ro i for his pun tual annotations, Professor Enri oBernardi forhisusefulideas onthemathemati al partand Professor Ni olaGuglielmifor his strongdire tions onthealgorithmi part.

A parti ularly warm a knowledgment goes to theSupervisory Statisti s Division of the European Central Bank, where I spent a semester as PhD trainee: they wel omed me and allowed me to omplete my PhD studies developing ru ial skills for real data analysis in large dimensions. A spe- i thanksgivinggoestoAngelosVouldis,wonderfulresear hsupervisor,to Mi healFedesin,for histrust and enthusiasm, andto Patri k Hoganfor his valuable managing.

Finally, I want to thank my family, for the long and patient guide, aid andsupport,myfriends,fortheyoftenalleviatedthis labour,andSilvia,for her ontinuous, passionate andenthusiasti listeningand in itement.

(6)
(7)

Thepresentthesis on ernslarge ovarian ematrixestimationvia omposite minimizationundertheassumptionoflowrankplussparsestru ture. Exist-ing methods like POET (Prin ipal Orthogonal omplEment Thresholding) perform estimationbyextra tingprin ipal omponentsandthenapplyinga soft thresholding algorithm. In ontrast, our method re overs the low rank plus sparse de omposition of the ovarian e matrix by least squares mini-mization under nu lear norm plus

l

1

norm penalization. This non-smooth onvex minimization pro edure is based on semidenite programming and subdierentialmethods,resultingintwoseparableproblemssolvedbya sin-gular valuethresholding plussoftthresholding algorithm.

Themostre entestimatorinliteratureis alledLOREC (LowRankand sparsECovarian eestimator)andprovidesnon-asymptoti errorratesaswell asidentiability onditions inthe ontext of algebrai geometry. Our work showsthattheunshrinkageoftheestimatedeigenvaluesofthelowrank om-ponentimprovestheperforman eofLOREC onsiderably. Thesamemethod also re overs ovarian e stru tures with very spiked latent eigenvalues like in the POET setting, thus over oming the ne essary ondition

p ≤ n

. In addition,itisproved thatourmethodre oversstru tureswithintermediate degrees of spikiness,obtainingalosswhi h isbounded a ordingly.

Then, an ad ho model sele tion riterion whi h dete ts the optimal point in terms of omposite penalty is proposed. Empiri al results oming from a wide original simulation study where various low rank plus sparse settings aresimulated a ordingto dierent parameter valuesaredes ribed outliningindetailthe improvements uponexistingmethods. Two real data-setsarenallyexploredhighlightingtheusefulnessofourmethodinpra ti al appli ations.

Keywords: ovarian e matrix, nu lear norm, thresholding, low rank plussparsede omposition, unshrinkage.

(8)
(9)

A knowledgments i

Abstra t iii

1 Introdu tion 1

2 State of the art 5

2.1 Sample ovarian e matrix estimators . . . 7

2.1.1 TheMaximum Likelihood ovarian e estimator . . . . 8

2.1.2 Theunbiased ovarian e estimator: xed

n

ontext . . 10

2.1.3 Covarian e matrix estimation: the IID data ontext . 10 2.2 Conditioningproperties . . . 11

2.2.1 Matrix onditioning asan ill-posedinverse problem . . 14

2.3 Ledoit andWolf's approa h . . . 16

2.3.1 General Asymptoti s . . . 17

2.4 Sparse ovarian e matrix estimation . . . 21

2.5 Fa tor analysisbasedestimators . . . 24

2.5.1 Stri tfa tor model . . . 26

2.5.2 PCAand fa toranalysis . . . 27

2.5.3 Approximate fa tormodel . . . 28

2.5.4 POET estimator . . . 30

3 Numeri al and omputational aspe ts 37 3.1 Anhistori al review . . . 40

3.1.1

l

1

norm heuristi s. . . 40

3.1.2 Nu learnormheuristi s . . . 45

3.1.3

l

1

norm plusnu lear norm . . . 49

3.2 Analyti aland algorithmi aspe ts . . . 51

3.2.1 Numeri al ontext: a semideniteprogram. . . 51

3.2.2 Solutionmethods . . . 55

4 Low rank plus sparse de omposition 63 4.1 Identi ation and re overy . . . 64

(10)

4.1.2 Approximate re overy: a fun tionalapproa h . . . 71

4.1.3 Approximate re overy: an extendedalgebrai approa h 78 4.1.4 Approximate re overy: LOREC approa h . . . 92

5 ImprovingLOREC 103 5.1 Theoreti al advan es . . . 104

5.2 Simulationsetting . . . 117

5.2.1 Simulationalgorithm . . . 117

5.2.2 Simulated settings and omparison quantities . . . 119

5.2.3 Anew modelsele tion riterion . . . 121

5.3 Data analysisresults . . . 123

5.3.1 Simulation results . . . 124

5.3.2 Realdataresults . . . 148

(11)

Introdu tion

Thepresentthesis on erns largedimensional ovarian e matrixestimation. Estimation of population ovarian e matri es from samples of multivariate data is of interest in many high-dimensional inferen e problems - prin i-pal omponents analysis, lassi ation by dis riminant analysis, inferring a graphi al modelstru ture, and others. Depending onthe dierent goalthe interest is sometimes in inferring the eigenstru ture of the ovarian e ma-trix(asinPCA) andsometimesinestimatingitsinverse(asindis riminant analysisor ingraphi al models). Examples ofappli ation areas where these problemsarisein ludegenearrays,fMRI,textretrieval, image lassi ation, spe tros opy, limate studies,nan e andma ro-e onomi analysis.

The theory of multivariate analysis for normal variables has been well worked out,see, for example,Anderson ([2℄). However, itbe ame apparent that exa t expressions were umbersome, and that multivariate data were rarely Gaussian. The remedy was asymptoti theory for large samples and xedrelatively small dimensions.

Inre entyears,datasetsthatdonot tintothisframeworkhavebe ome very ommon, the data are very high-dimensional and sample sizes an be verysmallrelativetodimension. Themosttraditional ovarian eestimator, the sample ovarian e matrix, is shown to be dramati ally ill- onditioned in a large dimensional ontext, where the pro ess dimension

p

is loser to or even larger than thesample dimension

n

,even inthe ase thatthe true ovarian e matrixiswell- onditioned. Somesolutionstothis drawba khave been proposed in the asymptoti ontext (for example [75℄ [15 ℄ [45℄). An alternative re ent approa h isbynumeri al optimization,whi h providesin thenon-asymptoti ontext, some solutions improving upon thementioned ones.

As des ribed intheexisting literature, two key propertiesof thematrix estimation pro ess assumea parti ular relevan einlarge dimensions:

1. well onditioning,i.e. numeri al stability; 2. identiability.

(12)

Bothpropertiesare ru ialfor thetheoreti alre overy andthepra ti aluse of the estimate. A bad onditioned estimate suers from ollinearity and auses its inverse, the pre ision matrix, to amplify dramati ally any error inthe data. A large dimension may ause theimpossibility to identify the unknown ovarian e stru ture andthedi ultyto interprettheresults.

The rst property is strongly related to regularization te hniques. A basi referen einthisrespe tisTibshirani(1996)([108 ℄),wheretheLASSO estimationalgorithm inthe ontext ofregressionmodelswasrstproposed. The se ond property an be ensured by dimensionality redu tion methods, whi h an beusedto redu e theparameterspa e dimensionality.

Regularization approa hes to large ovarian e matri es estimation have therefore started to be presented in the literature, both from theoreti al and pra ti al points of view. Some authors propose shrinkage towards the identitymatrix([75℄),others onsidertaperingthesample ovarian ematrix, thatis, gradually shrinkingtheo-diagonal elementstoward zero([54 ℄). At the same time, a ommon approa h is to en ourage sparsity, either by a penalizedlikelihoodapproa h([53 ℄)orbythresholdingthesample ovarian e matrix([100 ℄).

Forthisreason,ourresear hstudiesaspe i regularizationproblem un-dertheassumption oflowrankplussparsede omposition forthe ovarian e matrix. Su haproblemissolvedexploitingnon-smooth onvexoptimization methods. Thisapproa h allows toproperlyaddressbothre onditioningand dimensionalityredu tion issuesand is proved to be ee tive even in alarge dimensional ontext.

Ourdissertationmovesfromadetailedoutlineofasymptoti approa hes. In Chapter 2, we provide a thorough des ription of the motivation to our workand areviewofsome relevantasymptoti methodsfor ovarian e esti-mation. Maximumlikelihood estimators and unbiased nite estimators are des ribed ([2℄). Spe i treatment to the onditioning problem for ovari-an ematrix estimates is given. The ovarian e shrinkage estimator derived byLedoitand Wolfinthegeneral asymptoti frameworkisdes ribed([75 ℄). Sparse ovarian e estimators are shown together with the underlying as-sumptions and the estimation error rates, with parti ular referen e to the thresholding estimator of [15 ℄. POET (Prin ipal Orthogonal omplEment Thresholding) estimator([45 ℄), whi h ombines Prin ipal Component Anal-ysisand thresholding algorithms, isanalyzed indetail.

InChapter 3,we dene the regularization problemabovementioned. It isanu learnormplus

l

1

normapproximationproblem,andworksunderthe assumption of low rank plus sparse stru ture for the ovarian e matrix. It is omposed by a least squares loss and a omposite non-smooth penalty, whi h isthe sum ofthe nu learnorm ofthelow rank omponent and the

l

1

normof thesparse omponent.

The numeri al rationale behindthe problemformulation is provided. It isshownhowthis problem anbere astfromthepoint ofviewofnumeri al

(13)

analysisasasemi-deniteprogram(SDP).Nonstandardoptimizationtools, as subgradient minimization methods, are needed to solve it. We des ribe themost re ent solution algorithm andpoint out its ee tiveness.

InChapter4,weprovideawidereviewofexistingnon-asymptoti meth-ods. Theevolution path ofthe most re ent worksis guredout. Themost re ent developmentsofthenumeri al approa hunder theassumption oflow rank plus sparsestru ture for the ovarian e matrix are des ribed, starting fromthebasi ontributionbyChandraskeranetal. ([30℄)whi hrstproves the exa t re overy of the ovarian e matrix in the noiseless ontext. This resultisa hieved minimizing aspe i onvexnon-smooth obje tive, whi h isthe sumof the nu lear norm ofthelowrank omponent and the

l

1

norm ofthesparse omponent.

Then, therstapproximatesolutiontore overyandidentiabilityinthe noisy ontext, oming from[1℄,is des ribed. In the following,theextension of [30℄ providing the rst exa t solution of the numeri al problem in the noisygraphi al model setting([31 ℄) is shownindetail. In that ontext, the obje tive isa leastsquarelosspenalizedbythe above mentioned omposite penalty,anditsoptimizationallowsto re overtheinverse ovarian ematrix. In on lusion, theextension of this framework to the ovarian e matrix es-timation ontext, oming from [77℄,is explained. Theresulting estimator is alledLOREC (LOwRankand sparsE Covarian e estimator).

In the last hapter (Chapter 5), an improvement over the solution de-s ribedin[77℄isproposed,basedontheunshrinkage oftheestimated eigen-valuesof thelow rank omponent. Luo'sapproa h is ompleted by deriving therates of the sparse omponent estimate, and the onditions for its posi-tivedenitenessandinvertibility. Inaddition,theratesofLORECunderthe onditions of POET, and, more importantly, in a ontext where the eigen-values of the low rank omponent are allowed to grow with

p

α

, α ∈ [0, 1]

(generalizedspikiness ontext)areprovided.

Inthefollowing, weshowtheresults ofourpro edureon both simulated and real data sets. We illustrate a new model sele tion riterion whi h is proved to be ee tive in our ontext. An original simulation study is presentedwhere extensive simulation results arepointedout, aswell asthe simulationalgorithm and theestimationassessment framework.

In the end,theperforman e of our newproposedestimator is ompared totheoneof LORECandPOET undervarioussettings. Tworealexamples areprovidedwhere ourmodelisee tive respe ttothe ompetitors. In par-ti ular, the se ondexample isabanking supervisory dataset whi h olle ts supervisory reporting indi ators of themost relevant Euro Area banks. We expli itlythanktheSupervisoryStatisti s DivisionoftheEuropeanCentral Bank,where theauthorspentasemesterasaPhDtrainee,fortheallowan e to usethese datainanonymous formfor resear h purposes.

(14)
(15)

Covarian e matrix estimation:

state of the art

In this hapter, a short review of existing solutions to the problem of o-varian e matrix estimation is provided. Parti ular attention isgiven to the two properties displayed intheIntrodu tion(well onditioning and identi-ability)andtothe performan eofexistingmethods inthelarge dimensional ontext. An exhaustive review an be foundin Pourhamadi (2013)([95 ℄).

ThisChaptershowsapath a rossexistingestimators aimedat outlining the two mentioned features (well onditioning and identiability) for ea h estimation setting, espe ially when

p

is very large ompared to the sample size

n

orevenlarger. Thisiswhy,forea hestimator,adetaileddis ussionof theasymptoti frameworkandtheassumptionsneededtoensure onsisten y (i.e. the onvergen e to the theoreti al ovarian e matrix)is provided.

Existing approa hes to the estimation problem are des ribed in this Chapter, while non-asymptoti approa hes will be the obje tof next hap-ters. The des ription ofpast approa hes isintended to displaythe main is-suesen ountered byexistingmethods,withparti ular referen eto thelarge dimensional ontext,andthe reasonswhyweneed todevelopanalternative numeri al approa hto the ovarian e estimation problem.

Therst paragraph(2.1) isdevotedto ovarian e matrix estimation un-der the assumption of normality for the data. The maximum likelihood estimator, i.e. the sample ovarian e matrix, is introdu ed and justied. Theunbiasedsample ovarian e matrix,undertheassumption ofxed

n

,is thenoutlined. A spe i remark ontheasymptoti distribution ofthe sam-ple ovarian e matrix under the assumption of independen e and identi al distributionfor thedata on ludes these tion.

In the se ond paragraph (2.2) the onditioning properties of the sam-ple ovarian e matrix are explored. The reason why the sample ovarian e matrix is bad- onditioned when the dimension is lose to the sample size is deeply explained and analyzed, as well as the reason why the inverse

(16)

ovarian e matrixdramati allyampliestheestimationerrorin aseof bad- onditioning.

The third paragraph (2.3) widely des ribes a su essful attempt to ad-dress theproblemof re onditioning thesample ovarian e matrix when the dimension islargerthan thesample size: theshrinkage estimator byLedoit &Wolf([75 ℄). Theirmotivations, theirresults andtheir asymptoti ontext areproperlyhighlighted, tryingtoretainthekeyelementsoftheirapproa h. The fourth paragraph (2.4) briey outlines existingsparsity estimators, withparti ular referen e to the thresholding estimator by Bi kel & Levina ([15 ℄), whi h is des ribed in detail with respe t to model assumptions and onvergen e rates. There we point out the strong link between sparsity assumptions and shrinkage thresholding. That family of estimators shows howit ispossible to usesparsityto re ondition the ovarian e estimateand to signi antly redu e thenumber ofparameters.

The fth paragraph (2.5) des ribes ovarian e matri esestimator based onfa tormodelassumptions. Abriefoverviewoffa tormodelspe i ations and underlying assumptions a ross history is provided, dis ussing the dif-ferent asymptoti ontexts. The relationship between Prin ipalComponent Analysis(PCA,[72 ℄)andfa tormodelling(see[59 ℄)is ru ialinthisrespe t. Finally,POETestimator([45 ℄),basedontheassumptionofapproximate fa -tor model witha sparseresidual matrix, is widely illustrated, pointing out the ru ialassumptions for onsisten y and identiability.

In[45 ℄, the population ovarian e matrix is assumedto be thesum of a low rank and a sparse omponent. POET works under the assumption of sparseresidual ovarian e matrix and pervasiveeigenvalues of thelow rank omponent (as

p → ∞

). Thisstru ture is parti ularly onvenient in alarge dimensional ontext, and ta kles both the issues mentioned above, as we willwidelyexplain. For the same reasons,the fa toranalysis assumption is a key to approa h ovarian e estimation in large dimensions. The asymp-toti orresponden ebetweenPCAandfa torestimationisthereestablished a ordingto the underlying assumptionsand thenexploited.

Before starting, we des ribe the basi matrix terminology. We restri t our analysis to the real ase. The spe traltheorem ensures that, when

M

is a positive semidenite squared

p

- dimensional real matrix with rank

r

, thereexistsanorthogonal

p × r

matrix

U

andadiagonal

r × r

matrix

Λ

su h that

M = U ΛU

=

r

X

i=1

λ

i

u

i

u

i

,

(2.1)

whi histheeigenvaluede ompositionof

M

. S alars

λ

1

, . . . , λ

r

are alled theeigenvaluesof

M

andarestri tlylarger than

0

. The

r

olumnsof

U

are theeigenve torsof

M

. If

M

issymmetri ,theeigenvalues oin idewiththe singularvalues

σ

1,...,r

,whi harethesquarerootsoftheeigenvaluesof

M

M

(17)

i.e. the absolute values of the eigenvalues of

M

. A fortiori, this happensif

M

isa ovarian e matrix,whi h issymmetri and positive denite.

The relevant norms we aregoing to usethroughout theentire thesisare (see also[62 ℄):

• ||M||

2

=

max

(M

M )

isthespe tralnormof

M

,whi hisits largest

singularvalue.

• ||M||

= max

i,j

|m

ij

|

is the innity norm of

M

, whi h is the largest

entry inmagnitude.

• ||M||

F

= trace(M

M ) =

q

P

i

P

j

m

2

ij

is the Frobenius norm of

M

, whi his the squareroot ofthesum oftheentriesof

M

.

• ||M||

= trace(

M

M ) =

P

p

i=1

σ

i

, sum of the singular values of

M

.

||M||

is alled nu lear norm. If

M

is a Positive SemiDenite

ma-trix(PSD),

||M||

= tr(M )

,be ause theeigenvalues and the singular

valuesexa ly oin ide.

• ||M||

1

=

P

i

P

j

|m

ij

|

: sum oftheabsolute values of theentries of

M

.

For a

p

-dimensional ve tor

x

,therelevant norms for our purpose are:

• ||x||

2

=

q

P

i

x

2

i

,the Eu lidean norm of

x

.

• ||x||

1

=

P

p

i=1

|x

i

|

,the

l

1

norm of

x

.

• ||x||

= max

i

|x

i

|

,themaximumnorm of

x

.

2.1 Sample ovarian e matrix estimators

In this paragraph we fo us on the most used estimator of the ovarian e matrix: the sample ovarian e matrix. First, we will derive it asthe maxi-mumlikelihoodestimatorof the ovarian e matrix under theassumption of multivariatenormalityfor our data(2.1.1). Maximum likelihood estimators are onsistentwhen

n → ∞

. Thisiswhywethenderivetheunbiased ov ari-an e estimator under the assumption of

n

nite (2.1.2), whi h is a slightly modied version of the sample ovarian e matrix. These two estimators asymptoti ally onverge when

n → ∞

,under the assumption of

p

xed. In theend of this paragraph, we give a ash about the behaviour of this esti-mator under the assumption of independen e and identi al distribution for

our datawhen

n → ∞

(2.1.3).

Our main referen e for this argument is the famous book by Anderson ([2℄).

(18)

2.1.1 The Maximum Likelihood ovarian e estimator

Suppose we have a sample

(x

1

, . . . x

n

)

, from a real-valued

p−

dimensional normal random variable

x ∼ N

p

, Σ

)

, with

p ≤ n

. The

p × p

matrix

Σ

= E((x − µ

)(x − µ

)

)

is real positive denite and symmetri , while

µ

= E(x)

isa

p × 1

ve tor.

Thedensity of

x

isthefollowing:

f (x|µ

, Σ

) = (2π)

1

2

p

|

1

2

exp



1

2

(x − µ

)

Σ

∗−1

(x − µ

)



.

where

µ

is a

p × 1

ve tor and

Σ

is a

p × p

invertible (positive denite)

matrix.

Thelikelihood fun tionis

L(µ

, Σ

) =

n

Y

i=1

N (x

i

, Σ

) =

= (2π)

1

2

pn

|

1

2

n

exp

"

−1/2

n

X

i=1

(x

i

− µ

)

Σ

∗−1

(x

i

− µ

)

#

.

Thelog-likelihood isthen

log L(µ

, Σ

) = −

1

2

pn log 2π −

1

2

n log |Σ

| −

1

2

n

X

i=1

(x

i

− µ

)

Σ

∗−1

(x

i

− µ

).

We denoteby

µ

ˆ

M L

and

Σ

ˆ

M L

theve torand thepositivedenite matrix maximizing

log L

. They are the maximum likelihood estimators of

µ

and

Σ

. Sin e

log L

is an in reasing fun tion of

L

,

log L

and

L

share thesame maximum respe tto our parameter estimates.

Thefollowing important theoremholds:

Theorem2.1.1. If

x

1

, . . . x

n

onstituteasamplefrom

N (µ

, Σ

)

with

p < n

, the maximum likelihood estimators of

µ

and

Σ

are

µ

ˆ

M L

= ¯

x =

1

n

P

n

i=1

x

i

and

Σ

ˆ

M L

=

1

n

P

n

i=1

(x

i

− ¯x)(x

i

− ¯x)

respe tively.

Theproof anbefoundinAnderson(1958),page67andfollowing. It ex-ploitsthepropertiesofthearithmeti meanandofpositivedenitematri es. Thekeyargument isthat

log L

an berewritten inthefollowingway:

1

2

pn log 2π −

1

2

log |Σ

| −

1

2

trΣ

∗−1

D −

1

2

n(x

i

− µ

∗−1

(x

i

− µ

)

,

where

D =

P

n

i=1

(x

i

− ¯x)(x

i

− ¯x)

.

In order to perform maximization, the ne essaryassumption is that

Σ

isa positive denite matrix. This ondition is ne essaryto ensure that the

(19)

term

n(x

i

− µ

∗−1

(x

i

− µ

)

a hievesa maximumfor

µ

= ¯

x

andtheterm

log |Σ

| − tr(Σ

∗−1

D)

a hieves amaximumfor

Σ

=

1

n

D

.

ML estimators show a number of interesting optimality properties. In parti ular,theyare onsistentandasymptoti allye ient([34℄). Atheorem by Cramer ensures that

µ

ˆ

M L

and

Σ

ˆ

M L

are minimum varian e (asymptoti- ally) unbiased estimators. Theseproperties holdifand onlyif

n → ∞

.

Note that also the ondition

p < n

is ne essary in order to perform maximization. In orderto seethis point, we need to re all a basi theorem ([2℄,p.77):

Theorem2.1.2. Themaximum likelihood estimator

µ

ˆ

M L

= ¯

x =

1

n

P

n

i=1

x

i

, from

N (µ

, Σ

)

, is distributed a ording to

N (µ

,

1

n

Σ

)

and independently of

Σ

ˆ

M L

= ˆ

Σ =

1

n

P

n

i=1

(x

i

− ¯x)(x

i

− ¯x)

.

n ˆ

Σ

is distributed a ording to

P

n−1

i=1

z

i

z

i

, where

z

i

∼ N(0, Σ

)

, and

z

1

, . . . , z

n−1

are independent.

This theorem states that under the multivariate normality assumption for thedata,

n ˆ

Σ

is the sumof

n − 1

squared

p

dimensional matri eshaving

rank

1

. If

p ≥ n

,

n ˆ

Σ

will never have full rank

p

.

In addition, it has been shown by Wishart ([113 ℄) that

D = n ˆ

Σ

is a matrix-valued sto hasti pro ess having the following distribution:

f (D|Σ

) =

|D|

1

2

(n−p−1)

exp

1

2

tr(Σ

∗−1

D)



2

1

2

np

π

p(p−1)

4

|

1

2

n

Q

p

i=1

Γ[

1

2

(n + 1 − i)]

whi hisaWishart distributionwith

ν = n − 1

degrees offreedom,where

Γ(t) =

R

0

x

t−1

e

−x

dx

is the usual Gamma fun tion. The proof is reported in [2℄ (p.252 and following). It exploits massively the linear transforms of randomvariables,andis basedonthepropertiesofGram-S hmidt orthogo-nalization algorithm.

Thisresultswasrstderivedforabi-variatedistributionbyFisher([51 ℄) where the distribution of the orrelation oe ient (rst dened by Karl Pearson in[91℄)was alsoderived.

We an now understand why

p < n

is a ne essary ondition. If

n ≤ p

,

f (D|Σ

)

isnolongeradensity,su hthatitisnolongerpossibletoderivethe

asymptoti distributionfor

ˆ

Σ

(i.e.,alltheusualoptimalitypropertiesofML estimators arelost). Infa t,

|D|

would be zero, andthe distribution would thusbedegenerate, having nullmeasurein

R

p×p

everywhere. Notealsothat

if

n = p + 1 f (D|Σ

)

has not a mode, analogously to the

χ

2

distribution withtwo degrees offreedom.

Inthe sameway,denotingby

T

thequantity

T = (¯

x − µ

)

W

−1

x − µ

)

, where

W =

D

n−1

,ithasbeen shownbyHotelling ([64℄)that

ν − p − 1

vp

T

2

∼ F

(20)

where

F

isFisher'sdistribution with

p

and

ν − p + 1

degrees offreedom

(

ν = n−1

).

T

2

is alledHotelling'sT-squareddistribution. Itisnon-singular if and only if both

µ

ˆ

and

Σ

ˆ

are non-singular, i.e. if

Σ

is positive denite

and

ν − p + 1 > 0

(equivalent to

n > p

).

So, both the sample mean and the sample ovarian e matrix are ML estimators of the true mean and the true ovarian e matrix if and only if thetrue ovarian e matrixispositive deniteandthedimension

p

isstri tly smallerthanthesamplesize

n

. Inparti ular,thedistribution ofthesample ovarian ematrixis

n

n−1

W ishart(Σ

, n−1)

. Thismeansthat

Σ

ˆ

isbiasedif

n

isnite. Notethatthisdistributiondoesnot hangeevenwhenthetruemean

µ

is known, unless

x

¯

is repla ed by the true

µ

. In that ase, the degrees offreedom are

n

and theresulting estimator (

1

n

P

n

i=1

(x

i

− µ

)(x

i

− µ

)

) is unbiased.

2.1.2 The unbiased ovarian e estimator: xed

n

ontext In order to derive the nite sample unbiased estimator of the ovarian e matrix,the keyresultisTheorem2.1.2 aboutthedistribution of

D = n ˆ

Σ =

P

n

i=1

(x

i

− ¯x)(x

i

− ¯x)

shownabove.

A orollaryofthat theoremstates:

Corollary 2.1.1. Let

x

1

, . . . , x

n

(n > p)

be independently distributed, ea h a ording to

N (µ

, Σ

)

. The distribution of

ˆ

Σ

ν

=

1

ν

P

n

i=1

(x

i

− ¯x)(x

i

− ¯x)

is

W ishart(Σ

, ν)

, where

ν = n − 1

.

Thisresultmeansthat

Σ

ˆ

n−1

= (

n−1

1

)

P

n

i=1

(x

i

−¯x)(x

i

−¯x)

istheunbiased estimator of the ovarian e matrix when the dimension

n

is nite. This estimator will be theinput of our new estimation pro edure in Chapter 4. Clearly,

Σ

ˆ

n−1

and

Σ

ˆ

n

onverge asymptoti allyto thesame estimator. We are now going to derive the asymptoti (normal) distribution of the sample ovarian e matrix inthemore general ase ofIIDdata.

2.1.3 Covarian e matrix estimation: the IID data ontext Let us suppose

x

i

∼ IID(µ

, Σ

)

,

i = 1 . . . , n

. We want to derive the asymptoti distribution of

ˆ

Σ

n

=

1

n

P

n

i=1

(x

i

− ¯x)

(x

i

− ¯x).

Under the IID

hypothesis,wehave:

E(x

i

x

i

) = E(x

i

)E(x

i

) = Σ

+ µ

µ

,

V (x

i

x

i

) = V (x

i

) + V (x

i

) = Σ

+ Σ

= 2Σ

.

Ourtarget an berewritten asthesumof three omponents:

1

n

n

X

i=1

(x

i

− ¯x)(x

i

− ¯x)

=

n

X

i=1

x

i

x

i

n

− 2

n

X

i=1

¯

x

x

i

n

+

n

X

i=1

¯

x

n

(21)

Sin e

P

n

i=1

x

n

i

prob

→ µ

,we have that

−2¯x

n

X

i=1

x

i

n

+

n

X

i=1

¯

x

n

= −2¯x¯x

+ ¯

x

= −¯x¯x

.

onverges inprobability asfollows:

−¯x¯x

′ prob

→ −µ

µ

(2.2)

Now, the rst omponent

P

n

i=1

x

i

x

i

n

anbe rewrittenas

1

n

n

X

i=1

(x

i

x

i

)

n

So,for the Central Limittheorem, we have

1

n

n

X

i=1

x

i

x

i

− (Σ

+ µ

µ

)

n

CLT

n

1

N (µ

µ

+ Σ

, 2Σ

).

Re alling (2.2),wehave that

ˆ

Σ

n

distrib

1

n

N (Σ

, 2Σ

).

(2.3) Theseresults nd onrmation in[58 ℄.

2.2 Thesample ovarian ematrix: onditioning prop-erties

We arenowgoing tobrieytalkaboutmatrix onditioning. Letus suppose

p

and

n

are xed. If

n > p,

the expe ted value of

Σ

ν=n−1

is

Σ

, and the entries of its ovarian e matrix are

V (ˆ

σ

n,ij

) =

2

ij

ii

σ

jj

)

(n−1)

. This highlights whythevarian eof

Σ

ˆ

n

in reasesasthetrue onditionnumberof

Σ

in reases. If the ondition number

c = σ

max

min

in reases, the orrelation between the omponents

x

i

and

x

j

in reases, be ause

Σ

is loser to ollinearity. Consequently,

V (ˆ

σ

n,ij

)

in reases,be ause

σ

∗2

ij

is losertoitsmaximum,whi h is

σ

ii

σ

jj

(fortheCau hy-S hwartz inequality).

Coming ba k to the main point, it is ru ial to study the behaviour of thesampleeigenvalues. In thematrixestimation ontext thereisa relevant issueabout numeri al onditioning, i.e. the behaviour of samplemaximum andminimumsingular values,ofa

p × n

datamatrix

X

.

Theorem2.2.1(Theorem([39 ℄)). Givennaturalnumbers

n, p

with

p < n+1

(22)

andvarian e

1

n

. Thenthe largest and smallest singular values

σ

min

(X)

and

σ

max

(X)

are su h that

max



P r



λ

max

≥ 1 +

r p

n

+ t



, P r



λ

min

≤ 1 −

r p

n

− t



≤ exp

 −nt

2

2



,

for any

t > 0

.

Thistheoremwasprovedbyusingargumentsfromrandommatrixtheory and the geometry of Bana h spa es. It is an essential result to provide a probabilisti boundfortheerror distan e

||ˆ

Σ

n

− Σ

||

2

,where

Σ

ˆ

n

=

1

n

X

X =

1

n

P

n

i=1

x

i

x

i

.

Infa t, thefollowing Lemma holds:

Lemma 2.2.1. Let

ψ = ||Σ

||

2

. Given any

δ > 0

and

φ > 0

with

ψ ≤ 8φ

, letthe number of samples

n

be su h that

n ≥

64pφ

2

δ

2

. Then we have that

P r[||Σ

n

− Σ

||

2

≥ δ] ≤ 2 exp



2

128ψ

2



.

ThisTheorem isbased ona spe i assumption on

ψ

,thelargest eigen-value of

Σ

. By appropriately setting the parameter

ψ

, we an obtain the probabilisti bound a ordingly.

ThisLemma relieson the fa tthatthespe tralnorm isunitarily invari-ant, su h that it is possible to assume a diagonal stru ture for

Σ

ˆ

without lossofgeneralityand thenapply theprevioustheorem 2.2.1.

Itisremarkablethatwithoutfurtherassumptions,

Σ

ˆ

n

isnot invertibleif

p > n

(sin eit is perfe tly ollinear, having learly at most rank

n

,and for

therestnulleigenvalues). Evenif

p ≤ n

,inthe asetheratio

p/n

islessthan 1but not negligible, theestimated(maximumand minimum) eigenvalues are numeri ally unstable, sin e the probabilisti bound is too large. This may result in bad onditioning (i.e. too large ondition number) for

ˆ

Σ

n

. Thisis why inthe Big Data ontext, when

p

isvery large,it is frequent to have anill- onditioned sample ovarian e matrix,sin eit isdi ultto have enough observationto keep the ratio

p/n

negligible([75 ℄).

Theexampleingure(2.1) learlyoutlinesthedes ribeddrawba k. The eigenvalues of the ovarian e matrix of a simulated

n × p

pro ess

ǫ

i

=

N

p

(0,

n

1

I)

,

p = 100

,

n = [10, 50, 100, 500, 1000, 10000]

areplotted. The

g-uredisplayshowthedispersionoftheeigenvaluesde reasesas

p/n

de reases. Alldistributions tendto theMar enko-Pastur distribution, whi h isproved tobethe limitingdistributionoftheeigenvalues ofIIDrandomvariables(in the Kolmogorov asymptoti framework, see [79 ℄). The rank is always equal

to

min(p, n − 1)

. If

p = n

,the matrix isthus singular.

We have provided this simple example to state that without further as-sumption on the eigen-stru ture (values and ve tors) of

Σ

(23)

0

50

100

0

1

2

3

4

Eigenvalues

0

50

100

0

0.5

1

1.5

2

2.5

Eigenvalues

0

50

100

0

0.5

1

1.5

2

2.5

Eigenvalues

0

50

100

0.5

1

1.5

Eigenvalues

0

50

100

0.8

1

1.2

1.4

Eigenvalues

0

50

100

0.9

1

1.1

1.2

Eigenvalues

10

50

100

500

1000

10000

Figure 2.1: Eigenvalues of the sample ovarian e matrix of

ǫ

i

= N

p

(0,

1

n

I)

,

p = 100

,

n

varying

p ≤ n

is unavoidable in order to guarantee the positive deniteness (and

thus theinvertibility) of our ovarian e estimate. Anyway, the re overy of theeigen-stru ture ofa ovarian ematrix isstronglyrelatedtothe underly-ingassumptionsand to the asymptoti ontext.

Wenowenumeratethreeparametersettingsrelevantforourdissertation: 1.

p

and

n

xed: this is the ase of

Σ

ˆ

n−1

, and all numeri al estimators we will analyzeinnext hapters ([31 ℄, [1 ℄,[77℄,[15 ℄)

2.

p

xed,

n → ∞

: this isthe ase of

Σ

ˆ

M L

,or of theapproximate fa tor model([29℄)

3.

p

n

n

→ c

when

n → ∞

: herewendtheGeneralasymptoti framework,

used by Ledoit and Wolf to ensure the onsisten y of their estimator ([75 ℄),andtheKolmogorovasymptoti framework(wherealso

p → ∞

). Also onsisten y propertiesofthethresholding estimator([15 ℄)and of POETestimator([45 ℄)arederivedunderasimilarframework,where a fun tionof

p

and

n

tendsto0while

n → ∞

. Seeformoreexplanations se tions(2.4) and (2.5).

In the se ond ontext, with xed

p

and

n

,the outlined results on ern-ing numeri al onditioning for the sample ovarian e matrix hold, and the ondition

p ≤ n

is unavoidable without further assumptionsto derive nite sample bounds. This is why one of the aims of the present work is trying to exploit results from the third asymptoti framework (in terms of model assumptions) to establishboundsunder the nite sample ontext dropping the ondition

p ≤ n

.

(24)

2.2.1 Matrix onditioning as an ill-posed inverse problem We are nowexplaining in detail why a bad- onditioned sample matrix is a fataldrawba k for us. Thereason stands inthe onsequen esderiving from the inversionof abad- onditioned matrix.

Letus now onsiderthestandard linearsystem

Ax = b

,where

A

is

p × p

,

and

x

,

b

are

p × 1

. If our aim is to derive

b

(the output), we are solving

the dire tproblem. Ifour aim isto derive

x

(the input),we aresolving the inverse problem. If

A

is full rank, Cramer's theorem is ensuring that the inverse problem has exa t solution

x

= A

−1

b

. Otherwise, if

A

has rank

r < p

,weneed to solve theleastsquares problem

min

x∈R

p

||Ax − b||

2

,

andwe have

x

=

r

X

i=1

|u

i

b|

λ

i

u

i

(2.4)

||Ax

− b||

2

=

p

X

i=r+1

||u

i

b||

2

.

Thisfundamentalresult wasproved in[40℄. Howmu hissolutionthe

x

reliable? Hadamard([57℄)outlinedthethree hara teristi s ofa well-posed problem:

existen e: the problemadmits one solution

uniqueness: the problemhasat most one solution

stability: theproblem isnot sensitive to data perturbation.

Inour ontext,if

A

isfullrank,theinverseproblemmaybeill-posedsin e itviolatesthestability ondition. If

A

isnotfullrank,theinverseproblemis ill-posed sin e it violates theexisten e and the uniqueness ondition (there are only approximate solutions, no exa t ones). The least squares system servesfor identifying inany ase asolution even iftherewouldbe none.

Anyway,(2.1) and(2.4)enableus tounderstand whytheinverseof bad- onditioned matri es are numeri ally unstable. The solution of the dire t problemis

Ax = U ΛU

x =

P

p

i=1

λ

i

(u

i

x)u

i

,whi h dampensthe omponents orrespondingto thesmallesteigenvaluesof

A

. Onthe ontrary,(2.4)shows us thatthesolution ofthe inverse problemamplies theee ts ofthesame omponents. If we assumethat

b

isperturbed, i.e.

b

ǫ

= b + ǫ

,we note that

x

ǫ

= x

+

P

r

i=1

|u

i

ǫ|

λ

i

u

i

. So, if

A

is bad onditioned (i.e. we have very small eigenvalues), the ee t of data perturbation is amplied, and the solution maynot be ee tively usable inappli ations.

(25)

ThisiswhyPi ard([93 ℄)elaborateda onditionunderwhi htheinverse solution is reliable. It states that

x

=

P

r

i=1

|u

i

b|

λ

i

u

i

< ∞

if and only if

|u

i

b|

de ays morerapidlythan the orresponding

λ

i

forall

i

,whi h o ursif

λ

i

> τ ∀i

, where

τ

is the threshold at whi h thesingularvalues arelevelled

bythe noise.

If this ondition isviolated, a regularization method, like thetrun ated singularvaluede omposition(TSVD,see[55℄)orTikhonov'sregressionmethod ([109 ℄) or otherregressionmethods (like the ridgeone), are needed. Thisis whythenonasymptoti approa hfor ovarian ematrixestimationessentially onsistsinspe ifyingappropriateregularizationproblemsundersuitable on-ditions for deriving improved error rates, as we will widely des ribe in the following hapters.

Notethatthereisahugeliteraturedealingwiththedistributionof eigen-values. We mention again Mar enko-Pastur law, whi h des ribes the be-haviourof the singularvaluesof are tangular randommatrix having Gaus-sianentries([79 ℄). Tra y andWidom([107 ℄) foundthelimiting distribution ofthesingularvaluesofalargedimensionalrandomHermitianmatrix. John-stone ([70 ℄) found out the limiting distribution of the largest eigenvalue in prin ipal omponent analysis(for

n ≤ p

, underthe assumption of indepen-dentnormalityfor the olumnsofthe data matrix)whi hisproportionalto a Wishart of order

1

. A re ent work by Chiani([33 ℄) derived theexa t dis-tribution of the largest eigenvalues for real Wishart matri es and Gaussian Orthogonal Ensembles.

The work in[70℄,inparti ular, outlinedthat for large

p

it an be easier to re over the top

r

eigenvalues ifthey are parti ularly spiked, be ause the distribution of the

(r + 1)

-th eigenvalue isbounded bya Tra y-Widom law of lower dimensions (

n × (p − r)

respe t to

n × p

). Thus, the

(r + 1)

-th eigenvalueofasetof

p

eigenvalueswhere

r

arespikedissto hasti allysmaller than the largesteigenvalue ofa settingof

(p − r) < p

variables non-spiked. This fa tsuggests that large dimensions (

p → ∞

) an helpthe re overy of strong eigenvalues and somehow justies the use of "s ree-plot" to hoose thenumberof eigenvalues.

Therearealsosomeresultsonthedistributionofthesmallesteigenvalues. We referto [8 ℄ for ageneral review.

All in all,the problem of re onditioning our ovarian e matrix estimate is approa hed dierently a ording to the related asymptoti ontext. In Chapter 4 we will fo us on the non-asymptoti ontext, outlining various solutions re ently provided. Now, we will fo us on the des ription of key ovarian e estimators in the asymptoti ontext where both

p

and

n

are allowedtotendto

. Theestimatorweareabouttodes ribebelongstothe lassof shrinkage estimators ([68 ℄) whi h represent a widely usedapproa h in this ontext as an ee tive regularization method. It is relevant to note that the distributional assumption of normality is no longer needed, sin e

(26)

the approa h we aregoingto des ribe isdistribution-free.

2.3 Shrinkagetowardstheidentity: LedoitandWolf's approa h

Ledoitand Wolfweretherst toderive in[75 ℄a onsistentestimator ofthe ovarian ematrixinanewasymptoti framework, alledgeneralasymptoti framework. Theyproposedawaytotemperthenumeri alinstabilityof sam-ple eigenvalues, expli itly re onditioning them byshrinkage. The adoption of a new asymptoti framework was needed to ensure the shrinkage inten-sity to be positive, avoiding it to vanish in the limit. Their estimator is also Bayesian in nature, sin e it is a ombination of a priori and sample information. They all itEmpiri al Bayesian estimator.

Themotivatingresult oftheir analysisit reportedbelow.

Theorem 2.3.1. The eigenvalues are the most dispersed diagonal elements that an be obtained by rotation of a symmetri matrix.

Theproof exploitstheinvarian e by rotationoftra e.

This auses that the largest sample eigenvalues are positively biased, while the smallest are negatively biased, and the bias in reases in

p/n

(re- all Theorem 2.2.1). The pattern of sample eigenvalues depends on the Mar enko-Pastur distribution, whi h holds in the Kolmogorov asymptoti framework. Asdes ribed,underKolmogorovasymptoti stheratio

p/n

tends to aspe i onstant, whileboth

p

and

n

tendto innity.

HerewereportthesolutionproposedbyLedoitandWolftothedes ribed problem. Their idea is to shrink thesample ovarian e matrix towards the identity matrix, solving the following optimization problem (thus re ondi-tioningtheeigenvalues):

min

ρ

1

2

E[||Σ − Σ

||

2

]

s.t.

Σ = ρ

1

I

p

+ ρ

2

ˆ

Σ

n

.

where

ρ

1

and

ρ

2

arenonrandom oe ients.

Thetheoreti al solutionto this problemis theoptimal linear shrink-ageestimator

Σ

LW

=

β

2

γ

2

µI +

α

2

γ

2

Σ

ˆ

n

(2.5) with

E[||Σ

LW

− Σ

||

2

] =

α

2

β

2

γ

2

,where:

(27)

µ =< Σ, I >;

α

2

= ||Σ

− µI||

2

;

β

2

= E[||ˆ

Σ

n

− Σ

||

2

];

γ

2

= E[||ˆ

Σ

n

− µI||

2

].

Their derivation exploitsthe natural Pythagorean relationship

α

2

+ β

2

= γ

2

.

(2.6)

Inthis view,theratio

β

2

γ

2

is alledoptimal shrinkage intensity.

The most important interpretation of this approa h for our purposes is thefollowing. It iswell known(Theorem 2.2.1) thatthesampleeigenvalues ofIID datahave bounded error respe t to thetrueones, sothat, underthe ondition

p ≤ n

(

p

and

n

xed),

1

p

E(

P

p

i=1

λ

ˆ

i

) =

p

1

P

p

i=1

λ

i

,i.e. thetra e of

Σ

is unbiasedly estimated.

At the same time, theorem 2.3.1 shows that sample eigenvalues have a largerdispersionaroundtheirgrandmeanrespe ttothetrueones(assuming thattheeigenve torsarereliable). From (2.6)we an arguethat

1

p

E

"

p

X

i=1

λ

i

− µ)

2

#

=

1

p

p

X

i=1

i

− µ)

2

+ E[||ˆ

Σ

n

− Σ||

2

],

i.e. theex essdispersionofthesampleeigenvaluesistheerrorofthesample ovarian e matrix. This is why here the authors bound

[||ˆ

Σ

n

− Σ||

2

]

by bounding

1

p

E

h

P

p

i=1

λ

i

− µ)

2

i

,where

µ = 1

.

So,

Σ

LW

impli itly doesthere onditioning ofeigenvalues,sin e

λ

i,LW

=

β

2

γ

2

µ +

α

2

γ

2

λ

ˆ

i

, ∀i = 1, . . . , p.

1

p

E[

P

p

i=1

λ

i,LW

−µ)

2

]

isequalto

α

2

γ

,andisevensmallerthanthedispersion of thetrue ones, for the reasonsdes ribed above. Note thatthis method is very similarin its meaning to the

max log − det

heuristi s for nu lear norm minimization(see [49 ℄).

2.3.1 General Asymptoti s

In order to derive a feasible estimator, we now need to get into a new asymptoti framework, sin e theoptimal shrinkage intensity

β

2

vanishesas

||ˆ

Σ

n

− Σ

||

2

vanishes when

n → ∞

in the standard asymptoti framework (as proved in paragraph 2.1.3, see onvergen e (2.3)). This fa t, when

p

is loserto

n

or even larger,is in onsistent withreality. So,anew asymptoti framework, alledGeneralAsymptoti s,isneeded,where

β

2

(28)

Consider

n = 1, 2, . . .

indexing a sequen e of statisti al models, and for every

n

,

X

n

isa

p

n

×n

matrixof

n

iidobservationsonasystemof

p

n

random zeromeanvariables with ovarian e matrix

Σ

n

.

Thefollowing assumption hara terizes this ontext:

A1. There existsa onstant

K

1

independent of

n

su hthat

p

n

/n ≤ K

1

. Itisremarkablethatinthissetting

p

an hange andevengoto innity, but it is not required. Dierently from the Kolmogorov asymptoti frame-work (the one of Mar enko-Pastur Law), it is not even ne essary this ratio tendsto a nite onstant.

Two further assumptions are needed to derive a onsistent estimator of

Σ

LW

. If

Σ

n

= Γ

n

Λ

n

Γ

n

, the produ t

Y

n

= Γ

n

X

n

is a set of un orrelated variables spanning the same spa e as the original variables. The following restri tionson thehighermomentsof

Y

n

areimposed:

A2. There existsa onstant

K

2

independent of

n

su hthat

1

p

n

p

n

X

i=1

E[(y

n

i1

)

8

] ≤ K

2

,

A3.

lim

n→∞

p

2

n

2

P

i,j,k,l

∈ Q

n

Cov(y

i1

y

j1

, y

k1

y

l1

)

Cardinal of

Q

n

= 0.

where

Q

n

denotes the set of all the quadruples that are made of four distin tintegers between

1

and

p

n

.

Assumption2statesthattheeighthmomentof

y

isbounded(onaverage). Assumption 3 states that produ ts of un orrelated random variables are themselvesun orrelated(on average,inthelimit). Inthe asewhengeneral asymptoti s degenerate into standard asymptoti s (

p/n → 0

); Assumption 3istrivially veried asa onsequen eof Assumption2.

For what previously stated,Assumption3 isveried when random v ari-ablesarenormallyorevenellipti allydistributed,sin ethesample ovarian e of(un orrelated) normalvariables is asymptoti allyunbiased. Anyway,A3 ismu hweakerthan thatsituation.

Theseassumptions arespe i allyneeded to derive thesample ounter-partsof

µ

,

γ

2

,

β

2

.

Notethatthesetwoassumptionsheavilyinvolvetheeigenstru ture (eigen-values and eigenve tors) of the true ovarian e matrix. Here we need to imposerestri tionson eighthmoments,for theparti ular natureoftheir op-timal weights. Anyway, theneed to ontrol the pervasiveness of thelatent stru ture in the ovarian e matrix is ru ial for model re overy. We also underline how mu h latent fa torial assumptions an impa t on ovarian e estimation. Thisis whywe aregoing to spe i allydis uss therelationship between fa tor modelling and ovarian e estimationin paragraph(2.5).

Under these assumptions, Ledoit and Wolf approa h the study on the onsisten yoftheirestimator. Intheir ontext,thereferen enormis

||A||

n

=

(29)

1

p

n

tr(AA

)

,su hthattheidentitymatrixhasalwaysnormone,andthe refer-en e rossprodu tis

< A

1

, A

2

>

n

=

1

p

n

tr(A

1

A

2

)

. The problemof obtaining meaningful absolute rates in high dimensions is another relevant issue. As we willsee, in[45 ℄ the authors deriveasymptoti rates for therelative error matrix(andnotthe ovarian ematrixitself). Instead,underthe nonasymp-toti setting(Chapter4),wewillobtainniteabsoluterates,evenunderthe same assumptionsof [45℄.

We are now going to show why the sample ovarian e matrix is not onsistent in this ontext, dierently from the nite

p

ontext, where the ovarian e matrix isasymptoti ally onsistentundertheassumption of nor-mality. Theauthorsshowthatquantities

µ

n

=< Σ

n

, I >

,

α

2

n

= ||Σ

n

−µ

n

I||

2

,

β

n

2

= E[||ˆ

Σ

n

−Σ

n

||

2

]

,

γ

2

n

= E[||ˆ

Σ

n

−µI||

2

]

areboundedinthegeneral

asymp-toti framework when

n → ∞

. Then, they prove the following important Theorem: Theorem 2.3.2. Dene

θ

2

n

= V ar(

p

1

n

P

p

n

i=1

E[(y

n

i1

)

2

])

.

θ

2

n

is bounded as

n → ∞

,and we have:

lim

n→∞

E[||ˆ

Σ

n

− Σ

n

||

2

] =

p

n

n

2

n

+ θ

n

2

).

This result states that the sample ovarian e matrix is not onsistent under the general asymptoti framework, sin e its expe ted loss is lower bounded by

p

n

n

2

n

)

,whi hdoesnotusuallyvanish. (Re allthat

θ

2

n

vanishes asymptoti allyunderthe assumption ofnormality,for onvergen e (2.2)).

There aretwo interesting ex eptions:

when

p

n

n

→ 0

, we fall into the standard asymptoti ontext, where thesample ovarian e matrixis onsistent. Theonly dieren eisthat moregeneral ase

p = o(n)

isallowed,i.e.

p

isallowedtobeunbounded andgrow towardsinnity;

• µ

2

n

→ 0

and

θ

2

n

→ 0

.

µ

2

n

implies that most of the random variables havevanishingvarian es,i.e. thereare

O(n)

asymptoti allydegenerate variables. So, if the number of nondegenerate random variables is NOTnegligiblewithrespe ttothenumberofobservations,thesample ovarian e matrix is not onsistent.

In onsisten y is due to thedisequilibrium between the number of data-points

np

n

andthe number ofparameters

p

n

(p

n

+ 1)/2

. Thisis akey point in our analysis, whi h is unsolved by the approa h of Ledoit and Wolf. In fa t, they write there is no DIRECT onsistent estimator of the ovarian e matrixunderthegeneralasymptoti s. Theirstrategyistoderivea onsistent estimatoroftheirtheoreti alestimator,whi hisprovedtohavetheminimum riskamongallthelinear ombinationsof

I

p

and

Σ

n

andisshowntobebetter onditioned than thesample ovarian e matrix.

(30)

So, shrinkage matters unless

p

n

n

is negligible respe t

γ

2

µ

2

, i.e. if the dis-persionof sampleeigenvalues is mu h larger than

p

n

n

.

To on lude this se tion, we are now going to explain how Ledoit and Wolfderive a onsistent estimator for

Σ

LW

.

Theyintrodu e sample ounterparts of their keyquantities:

m

n

=< ˆ

Σ

n

, I >

n

,

d

2

n

= ||ˆ

Σ

n

− m

n

I||

2

,

¯b

2

n

=

n

X

k=1

||x

n

.k

x

n

.k

− ˆ

Σ

n

||,

b

2

n

= min(¯b

2

n

, d

2

n

),

a

2

n

= d

2

n

− b

2

n

,

where

x

n

.k

denotethe

k − th

olumn of

X

n

.

All these sample ounterparts are onsistent in the general asymptoti framework, i.e. they onverge to

µ

2

n

,

α

2

n

,

β

2

n

,

γ

2

n

respe tively in quadrati mean.

Then, their feasible onsistent estimator is

ˆ

Σ

LW

=

b

2

n

d

2

n

m

n

I

n

+

a

2

n

d

2

n

ˆ

Σ

n

(2.7)

Thisestimatoris onsistentinthe generalasymptoti frameworkrespe t to

Σ

LW

, i.e. they share the same asymptoti expe ted loss. Thus, the expe tedquadrati loss

α

2

β

2

γ

2

anbe onsistentlyestimatedinquadrati mean by

a

2

n

b

2

n

d

2

n

.

ˆ

Σ

LW

isshowntohaveanimportant optimalityproperty: ithasthesame asymptoti risk asthe theoreti al optimal linear ombination of

Σ

ˆ

n

and

I

n

withrandom oe ients. Inaddition, its ondition number isproved to be bounded inprobability,whi h isvery important for pra ti al use.

The approa h by Ledoit and Wolf is undoubtedly very elegant. How-ever, there is still one main di ulty: their estimator is ex essively better onditioned than the true ovarian e matrix, i.e. it is often too biased, for the presen e of the identity matrix in the estimator. This is why another major pointof ourdissertation will dealwiththeneedof "unshrinking"the estimatedeigenvalues.

Infa t, the numeri al issueis nottheonly relevantreason for desiringa well onditionedestimate ofthe ovarian e matrix. Deep statisti alreasons lie behind this need: we supposethat thetrue ovarian e matrix

Σ

is well onditioned, that isthere is nomulti- ollinearity amongour

p

variables. In this respe t, a well onditioned estimate is ru ial also for tting purposes, i.e. toimprove the statisti al propertiesof theestimate.

(31)

We re all the previous skrinkage estimator by the same authors ([74 ℄) in the market portfolio ontext. There, the authors spe ify for the ov ari-an ematrixasingle-indexmodel( onsistently withthebasi theoryofasset pri es, see [103 ℄), whi h is essentially a one-fa tor latent model, and then estimatethe ovarian e matrix deriving theoptimal shrinkage intensity to-wards the single-index as des ribed. This single index ovarian e matrix estimatorisaninteresting onta tpointbetween latent variablemodels and shrinkage methods.

Before passing to the analysisof fa tor-based ovarian e matrix estima-tors(paragraph(2.5)),wenowbrieyoutlinethe ovarian eestimatorsbased onpuresparsityassumptions, withparti ularreferen eto theuseof shrink-age thresholding. In this ontext, sparsity means that our true ovarian e matrix hasaprevalen eof zeros.

2.4 Sparse ovarian e matrix estimation

Inthis se tionwe listthe mostrelevant estimatorsbased ona puresparsity assumption, whi h an be ee tive for redu ing the number of parameters andre onditioning theestimate, removing unne essaryo-diagonal orrela-tions. If

p/n → c ∈ (0, 1)

(general asymptoti framework) the eigenvalues of

Σ

ˆ

n

follow the Mar enko-Pastur law, supported on

(1 −

c)

2

, (1 +

c)

2

. If

p/n

doesnot tendto a onstant, we do not have any guarantee. For this reason,enfor ing sparsity an bea key for obtaininga fullrank estimatein highdimensions, even when

n < p + 1

. However, there are lots of dierent typesofsparsityassumptions,methodsandasymptoti frameworkstoprove onsisten y.

The natural ontext whi h gave rise to the on ept of sparsity lies in a data-set showing a lear index ordering among variables. This ondition arises easily for spatial data, when the variables are geographi al areas for whi h a proximity matrix is naturally dened. Appli ations in lude spe -tros opyand limate data.

For this kind of data, several methods have been developed. Banding the ovarian e matrix, by appropriately dening a banding parameter, is one ee tive solution. In that approa h ([14 ℄), thematrix referen e lass is

Σ

∈ U(ǫ

0

),

where

U (ǫ

0

) =

∈ R

p×p

: 0 < ǫ

0

≤ Λ

i

) ≤ ǫ

−1

0

< +∞,

max

j

{

X

i

ij

| : |i − j| > k} ≤ Ck

−α

,

(2.8)

whi h is the lass of matri es having uniformly bounded eigenvalues and banded ovarian e.

For any

Σ

∈ U(ǫ

0

)

Figura

Figure 2.1: Eigenvalues of the sample 
ovarian
e matrix of ǫ i = N p (0, n 1 I) , p = 100 , n v arying
Figure 5.2: Sorted diagonal elements of L and S
Figure 5.3: Eigenvalues of Σ n and Σ
Figure 5.4: Model sele
tion 
riterion -
+7

Riferimenti

Documenti correlati

Say if the following statements are unambiguously true (TRUE), unambiguously false (FALSE) or impossible to classify the way they are stated (CAN’T SAY).. Write the mo- tivations

, n, has all entries equal to zero except those whose indeces are in ν(i) which are set equal to 1/µ i.. Let us explicit these formulas... Assume now that at the same time a new site

In the case of a matrix in reduced row echelon form this is even easier: the pivot columns are

[r]

Let us first look for a class of transformations (q, p) = C(Q, P ) satisfying the following Condition 1: to every Hamiltonian function H(q, p) one can associate another func- tion

 “A big data architecture is designed to handle the ingestion, processing, and analysis of data that is too large or complex for traditional

The reader should not be surprised by the fact that, when the rank of the unknown matrix increases, the algorithm needs more entries in order to recover correctly the entire matrix.

Figure 3.22: Experimental time series showing the presence of a stable upright position together with a stable large amplitude sub-harmonically resonant roll motion for ITACA in