simulation in heavy-tailed settings
Thorbjörn Gudmundsson
InthisthesisamethodbasedonaMarkov hainMonteCarlo(MCMC)
algorithmisproposedto omputetheprobabilityofarareevent.The on-
ditionaldistributionoftheunderlyingpro essgiventhattherareevento -
urshastheprobabilityoftherareeventasitsnormalising onstant. Us-
ingtheMCMCmethodologyaMarkov hainissimulated,withthat on-
ditionaldistributionas itsinvariantdistribution,and informationabout
thenormalising onstantisextra tedfromitstraje tory.
Thealgorithmisdes ribedinfullgeneralityandappliedtofourdier-
entproblemsof omputingrare-eventprobability. Therstproblem on-
sidersarandomwalk
Y 1 +· · ·+Y n
ex eedingahighthreshold,wherethein-rements
Y
areindependentandidenti allydistributedandheavy-tailed.These ondproblemisanextensionoftherstonetoaheavy-tailedran-
dom sum
Y 1 + · · · + Y N
ex eeding a high threshold, where the numberof in rements
N
is random and independent ofY 1 , . . . , Y n
. The thirdproblem onsidersasto hasti re urren eequation
X n = A n X n−1 + B n
ex eedingahighthreshold,wheretheinnovations
B
areindependentand identi allydistributedandheavy-tailed. Thenalproblem onsiderstheruinprobabilityforaninsuran e ompanywithriskyinvestments.
An unbiased estimator of the re ipro al probability for ea h orre-
spondingproblemis onstru tedwhosenormalisedvarian evanishesasymp-
toti ally. Thealgorithmisillustratednumeri allyand omparedtoexist-
ingimportan esamplingalgorithms.
Idennaavhandlingpresenterasenmetodbaseradpå MCMC(Markov
hain Monte Carlo) för att beräkna sannolikheten av en sällsynt hän-
delse. Den betingadefördelningenfördenunderliggande pro essengivet
att den sällsynta händelsen inträar har den sökta sannolikheten som
sin normaliseringskonstant. Med hjälp av MCMC-metodiken skapasen
Markovkedja med betingade fördelningen som sin invarianta fördelning
o henskattningavnormaliseringskonstantenbaseraspå densimulerade
kedjan.
Algoritmenbeskrivsifull generaliteto htillämpas på fyraexempel-
problem. Förstaproblemethandlaromenslumpvandring
Y 1 + · · · + Y n
som överskriderenhög tröskel,då stegen
Y
är oberoende,likafödelade medtungsvansadfördelning. Andraproblemet är enutvidgning av detförsta till summa av ett stokastisktantal termer. Tredjeproblemetbe-
handlar sannolikhetenattlösningen
X n
tillenstokastiskrekurrensekva- tionX n = A n X n−1 + B n
överskrider enhög tröskel då innovationernaB
är oberoende, likafördelade medtungsvansadfördelning. Sista prob- lemethandlaromruinsannolikhetförettförsäkringsbolagmedriskfylldainvesteringar.
Förvarjeexempelproblemkonstruerasenväntevärdesriktigskattning
av denre iproka sannolikheten. Skattningarnaär eektivai meningen
attderasnormaliseradevariansgårmotnoll. Vidare ärdekonstruerade
Markovkedjornalikformigtergodiska. Algoritmernaillustrerasnumeriskt
o hjämfösmedexisterandeimportan esamplingalgoritmer.
Iwanttoexpressmydeepestappre iationforthesupportandhelpthatIhave
re eived frommy supervisorHenrikHult. I amtrulygrateful forbegiven the
opportunitytoworkunder hisguidan e.
Iwanttooermyspe ialthanksto olleaguesatthefa ultyfortheiradvi e
andhelp,inparti ularFilip LindskogandTobias Rydén. Alsowantthankmy
fellow Ph.D. students, Björn, Johan and Pierre, for ountless dis ussions and
pra ti esessionsonthebla kboard.
Finally, I wantto thankmytwospe ial ones Rannveig and Gyðafor their
immensesupport andlove.
1 Introdu tion 1
1.1 Sto hasti simulation. . . 2
1.1.1 Samplingarandomvariable. . . 2
1.1.2 Markov hainMonteCarlo . . . 4
1.1.3 Rare-eventsimulation . . . 5
1.1.4 Importan esampling . . . 6
1.1.5 Heavy-taileddistributions . . . 6
1.2 Markov hainMonteCarloinrare-eventsimulation. . . 7
1.2.1 Formulation. . . 7
1.2.2 Controllingthenormalisedvarian e . . . 8
1.2.3 Ergodi properties . . . 10
1.2.4 E ien yoftheMCMCalgorithm . . . 11
1.3 Outlineand ontributionofthisthesis . . . 11
2 General Markov hain MonteCarlo formulation 13 2.1 Asymptoti e ien y riteria . . . 14
3 Heavy-tailedRandomWalk 15 3.1 AGibbssamplerfor omputing
P(S n > a n )
. . . . . . . . . . . . 153.2 Constru tingane ientestimator . . . 18
3.3 Numeri alexperiments. . . 19
4 Heavy-tailedRandomSum 21 4.1 AGibbssamplerfor omputing
P(S N n > a n )
. . . . . . . . . . . 224.2 Constru tingane ientestimator . . . 25
4.3 Numeri alexperiments. . . 27
5 Sto hasti Re urren e Equations 28 5.1 AGibbssamplerfor omputing
P(X m > c n )
. . . . . . . . . . . 285.2 Constru tingane ientestimator . . . 31
5.3 Numeri alexperiments. . . 33
6 Ruin probability in an Insuran e Model with Risky Invest- ments 35 6.1 AGibbssamplerfor omputingtheruinprobability . . . 36
6.2 Constru tingane ientestimatorofthere ipro alruinprobability 36
Mathemati almodellingofsystems,inforinstan enaturals ien eshasbeenone
ofthekeybuildingblo ksofs ienti understanding. Thesystemofinterestmay
bethemotionoftheplanets,thedynami owinaliquid, hangesinsto kpri es
orthetotalamountofinsuran e laimsmadeinayear. Oftenthemodelinvolves
thesystem'sdynami laws,long-timebehavioranddierentpossibles enarios.
Su hmodels nearlyalwaysin lude aparameter,oraset ofparameters,whi h,
thoughunknown in advan e are still needed to alibratethe model to reality.
Thus in order tohaveafully spe iedmodel apable offore astingthefuture
properties or value,then oneneedsto measure thevaluesof thetheunknown
parametersandtherebymostlikelyintrodu ingsomemeasurementerror. This
errorisassumedtoberandomandthustheresultingfore astistheout omeof
asto hasti mathemati almodel.
Withtheeverin reasing omputational apa ityinre entde adesthemod-
elsarebe omingmoreand more omplex. Minoraspe tsthat wereignored in
thesimpler models an now bein luded in the omputations, within reasing
omplexity. Resear hersandpra titionersalikestrivetoenhan e urrentmodels
andintrodu emoreandmoredetails toit,inthehopeofin reasingtheirfore-
astingability. Weathersystemsandnan epro essesareexamplesofmodels
that today are so involvedthat it is be oming di ult to give analyti aland
losedform answersto propertyand fore astingquestions. This hasgivenrise
toalternativeapproa hestohandlingsu h omplexsto hasti models,namely
sto hasti simulation.
Briey, simulation is the pro ess of sampling the underlying random fa -
tors of a model to generate many instan es of it, in order to make inferen es
about its properties. This has proved to be a powerfultool for omputation
in many a ademi elds su h as physi s, hemistry, e onomi s, nan e, in-
suran e. Generatinginstan es of even thehighly advan ed sto hasti models,
multi-dimensional, non-linear and highly sto hasti models an be done in a
few millise onds. Sto hasti simulationhas thus playedits partin the s ien-
ti progressofre entde adesandthesimulationthemselveshasgrownintoan
a ademi eldinitsownright.
In physi s,hypothesisare oftentested and veried viaanumberof exper-
iments. Oneexperiment is arried outafter another, and if su iently many
of the experiments support the hypothesis then it a quires a ertain validity
andbe omesatheory. Thiswasforinstan e the aseatCERN inthesummer
of2012, whenthe existen eoftheHiggs boson was onrmed throughexperi-
mentswhi hsupportedthe oldand well known hypothesis. However,one an
notalways arryoutexperimentstovalidatehypotheses. Sometimesit issim-
ply impossibleto repli ate the model in reality, asis the ase when studying
the ee ts of global warming. Obviously, sin e we anonly generate a single
physi al instan e of theEarth, any simulationsneedto bedone via omputer
modelling. To better ree treality, theresolution needsto behigh and many
dierentphysi alandmeteorologi alfa torsneedtobetakenintoa ount. The
surfa e of the Earth is broken into 10km times 10km squares, ea h with its
temperature, airpressure,moistureand more. Thedynami sof theseweather
fa torsneed to besimulatedwith small times steps,perhapsmany years into
thefuture. TheMathemati s andClimate Resear h Network (MCRN) arries
outextensivesto hasti simulations,repli atingtheEarthusingdierenttypes
ti simulationis immensely omputationally ostly. This s ienti work alone
justiestheimportan e of ontinuingresear handimprovementin theeld of
sto hasti simulation.
Asubeldofsto hasti simulationwhi hdealswithunlikelyeventsofsmall
probabilityis alled rare-eventsimulation. Examples of rare-eventsimulation
is when al ulating apital requirements of a nan ing rm subje t to Basel
III regulations, orof a insuran e ompany subje t to Solven y II regulations.
Natural atastrophessu h asavalan hes,vol ani eruptions,to name but few,
are also typesrare-eventsfor whi h we areinterested in analysing. This is of
parti ular importan e when it omes to omputationally heavy models. That
isbe ause,ifaneventisrarea omputerneedsmanysimulationsto getafair
pi ture of its frequen y and the ir umstan es in whi h it o urred. And if
everysimulationtakes upalot of omputationaltime, then athoroughstudy
wouldrequireaprohibitiveamountof omputertimewouldindeedberequired.
Thereforetheimprovementofe ientrare-eventsto hasti simulationisofhigh
importan e.
Theee tofheavy-tailsinsto hasti modellingisanimportantfa tornotto
beoverlooked. Byheavytailswemeanessentiallythatthereisanon-negligible
probabilityofextremeout omesthatdiersigni antlyfromtheaverage.Su h
extremeout omesmayhavea onsiderableimpa tonasto hasti system. For
instan e,large laimsduetoa atastrophi eventarriveataninsuran e ompany
ausing seriousnan ialdistressforthe ompany. Similarly,largeu tuations
onthenan ialmarketmayleadtoinsolven yofnan ialinstitutions. Indata
networksthearrivalofhugelesmay auseseriousdelaysinthenetwork,and
soon.
This thesispresentsa newmethodology in rare-event simulationbased on
the theory of Markov hain Monte Carlo. The general method presented in
Se tion 2makesverymodestprobabilisti assumptionsand in subsequentse -
tions(randomwalkinSe tion3,randomsuminSe tion4,sto hasti re urrent
equationsin Se tion 5,ruinprobabilityin Se tion6)is applied tofew on rete
examplesandshowntobee ient.
1.1 Sto hasti simulation
In this se tion we introdu e the basi tools in sto hasti simulation, su h as
pseudo random number, the inversionmethod and Monte Carlo. We present
theMarkov hain MonteCarlomethodologyanddis ussbrieyergodi ity.
1.1.1 Samplinga randomvariable
Inthisse tionwepresentthefoundationsof sto hasti simulation,namelythe
generationofapseudorandomnumberbya omputerandhowit anbeused
tosamplearandomvariable viatheinversionmethod.
Most statisti al software programs provide methods for generating a uni-
formly distributed pseudo random number on the interval, say,
[0, 1]
. Thesealgorithms are deterministi , at its ore, and an only imitate the properties
and behaviour of auniformly distributed randomvariable. The early designs
ofsu halgorithmsshowedawsin thesense thatthepseudorandomnumbers
generated followed a pattern whi h ould easily be identied and predi ted.
randomnumbers,mimi kingatruerandomnumberquitewell. Forthepurposes
of this thesis we assume the existen eof an algorithm produ ing auniformly
distributedpseudorandomnumber,andignoreanyde ien iesanderrorsaris-
ing from the algorithm. Inshort, we assume that we an sample a perfe tly
uniformlydistributed randomvariableinsome omputerprogram. Foramore
thoroughanddetaileddis ussionwereferto[48℄.
Now onsider arandomvariable
X
and denote byF
itsprobabilitydistri- bution. Saywewould like,viasome omputersoftware,to sampletherandomvariable
X
.One approa h is the inversionmethod. The inversionmethod in-volvesonlyapplyingthequantilefun tion touniformlyrandomvariable. More
formallythealgorithmisasfollows.
1. Sample
U
from thestandarduniformdistribution.2. Compute
Z = F −1 (U )
,where
F −1 = min{x | F (x) ≥ p}
. TherandomvariableZ
hasthesamedistri-butionas
X
asthefollowingdisplayshows.P(Z ≤ x) = P(F −1 {U } ≤ x) = P(U ≤ F (x)) = F (x)
.Themethod aneasilybeextendedtosampling
X
onditionedonbeinglarger than some onstantc
. Meaning that wewantto samplefrom the onditional distributionP(X ∈ · | X > c)
.Thealgorithmisformallyasfollows.
1. Sample
U
from thestandarduniformdistribution.2. Compute
Z = F −1
1 − F (c)
U + F (c)
.
Thedistributionof
Z
isgivenby,P(Z ≤ x) = P (1 − F (c))U + F (c) ≤ F (x)
= P
U ≤ F (x) − F (c) 1 − F (c)
= F (x) − F (c)
1 − F (c) = P(c ≤ X ≤ x)
P(X > c) = P(X ≤ x | X > c)
.Thustheinversionmethodprovidesasimplewayofsamplingarandomvariable,
onditionedonbeinglargerthan
c
,basedsolelyonthegenerationofauniformlydistributedrandomnumber.
The moststandardtoolfor sto hasti simulationis the Monte Carlote h-
nique. ThepowerofMonteCarloisitssimpli ity. Let
X
bearandomvariableandassumewewantto omputetheprobabilitythat
{X ∈ A}
forsomeBorelset
A
. The ideaof Monte Carlois to sample independentand identi ally dis- tributed opies of random variable, sayX 1 , . . . , X n
and simply ompute thefrequen y of hitting the set
A
. More formally, the Monte Carlo estimatorofP(X ∈ A)
isgivenbyb p = 1
n X n i=1
I{X i ∈ A}
.Whilethe pro edure is easyand simplethere are drawba ksthat will be dis-
ussedinSe tion1.1.3.
In this se tion we present a simulation te hnique alled Markov hain Monte
Carlo(MCMC) forsampling arandomvariable
X
despiteonlyhavinglimitedinformationaboutitsdistribution.
MCMC is typi ally useful when sampling a random variable
X
having adensity
f
that isonlyknownuptoa onstant,sayf (x) = π(x)
c
,where
π
isknown butc = R
π(x)dx
isunknown. This mayseemstrangesetupatrstbuton enotedthatthenormalising onstant
c
maybedi ulttodeter-mine,saythereisnoknown losedformfor
c
,thenthisisanaturalformulation.Anexampleofthistypeofsetup anbefoundinBayesianstatisti sandhidden
Markov hains.
Inshort,thebasi ideaofsamplingviaMCMCistogenerateaMarkov hain
(Y t ) t≥0
whose invariant density is the sameas ofX
, namelyf
. There existsplentiful of MCMC algorithms but weshall only name twoin this thesis, the
Metropolis-HastingsalgorithmandtheGibbsalgorithm.
ThemethodrstlaidoutbyMetropolis[41℄andthenextendedbyHastings
[26℄ is based on a proposal density, whi h we shall denote by
g
. Firstly theMarkov hain
(Y t ) t≥0
is initialised with someY 0 = y 0
. The idea behind theMetropolis-Hastingsalgorithmis togenerateaproposalstate
Z
usingthepro-posaldensity
g
. ThenextstateoftheMarkov hain isthenassignedthevalueZ
with the a eptan eprobabilityα
, otherwise the nextstate of the Markovhain staysun hanged (i.e. retains the same value as before). More formally
thealgorithmisasfollows.
Algorithm 1.1. Set
Y 0 = y 0
. ForagivenstateY k
, forsomek = 0, 1, . . .
, thenextstate
Y k+1
issampled asfollows1. Sample
Z
fromtheproposaldensityg
.2. Let
Y k+1 =
Z
with probabilityα(Y k , Z) Y k
otherwisewhere
α(y, z) = min{1, r(y, z)}
,r(y, z) = π(z)g(z,y) π(y)g(y,z)
.Thisalgorithmprodu esaMarkov hain
(Y k ) k≥1
whoseinvariantdensityisgivenby
f
. Foremoredetails ontheMetropolis-Hastingsalgorithmwereferto [3℄and[23℄.Anothermethodof MCMCsamplingistheGibbssampler,whi hwasorig-
inally introdu ed by Gemanand Geman in [22℄. If the random variable
X
ismulti-dimensional
X = (X 1 , . . . , X d )
, the Gibbs sampler updates ea h om-ponent at the time by sampling from the onditional marginal distributions.
Let
f k|6k (x k | x 1 , . . . , x k−1 , x k+1 , . . . , x d ), k = 1, . . . , d,
denote the onditional density ofX k
givenX 1 , . . . , X k−1 , X k+1 , . . . , X d
. The Gibbs sampler an beviewed as a spe ial ase of the Metropolis-Hastings algorithm where, given
Y k = (Y k,1 , . . . , Y k,d )
,onerstupdatesY k,1
fromthe onditional densityf 1|61 (· |
Y k,2 , . . . , Y k,d )
,thenY k,2
fromthe onditionaldensityf 2|62 (· | Y k+1,1 , Y k,3 , . . . , Y k,d )
,waysequalto
1
,sonoa eptan estepisneeded.An importantproperty of aMarkov hainis its ergodi ity. Informally, er-
godi itymeasuresthehowqui klytheMarkov hainmixesand thus howsoon
thedependen y of the hain dies out. This is ahighly desired property sin e
goodmixing speedsupthe onvergen eoftheMarkov hain.
1.1.3 Rare-eventsimulation
Insomespe i asesweareinterestedin omputingtheprobabilityof arare
event. Thismaybetheprobabilityofruinofanan ial ompanyduetorandom-
nessinthefuturevalueofassetsandliabilities. Themultidimensionalsystemof
investmentsandbondsmaybeso omplexthatasimulationofthe atastrophi
eventofaruinmaybefeasible. Foranotherexample, onsideragraphofsome
sort and say wesend out a parti le on arandom walkalong the graphgiven
somestartingposition. Computingthesmall,and qui klyde reasingprobabil-
ity,of thatparti lereturningto itsstartingposition maybeofinterestasitis
anindi atorofthat graph'sdimension. Forthese reasonsand manyother, the
omputationoftheprobabilityforarare-eventisrelevant.
Consideranunbiasedestimator
p b
oftheprobabilityp
andinvestigateitsper- forman eastheprobabilitygetssmallerp → 0
. Ausefulperforman emeasure istherelativeerror:RE
(b p) =
Std(b p) p
.An estimatoris said to have vanishing relative error if RE
(b p) → 0
asp → 0
andboundedrelative error ifRE
(b p) < ∞
asp → 0
.Itiswellknownthat theMonteCarloestimatorisine ientfor omputing
rare-event probabilities as the following argument shows. Let
X
be a givenrandomvariablewithdistributionfun tion
F
andsaywewouldliketo omputep = P(X ∈ A)
. Wesamplenumberofi.i.d. opiesofX
,denotedbyX 1 , . . . , X n
and ompute
b p = 1 n
X n i=1
I{X i ∈ A}
.Thevarian eoftheestimatoris
Var(b p) = n 1 p(1 − p)
,whi h learlytendstozeroas
n → ∞
but that isnot main on ernhere. Whatis moreinteresting is its relativeerrorastheprobabilityp
tendstozero. ItsrelativeerrorisgivenbyStd
(b p)
p =
r 1 n
1 p − 1
.
The relative error tends to innity as
p → 0
. Thus making the Monte Carloalgorithmvery ostlywhenit omestorare-eventsimulation. Forexample,ifa
relativeerrorat
1%
isdesiredandtheprobabilityisoforder10 −6
thenweneedtotake
n
su hthatp
(10 6 − 1)/n ≤ 0.01
. This impliesthatn ≈ 10 10
whi hisinfeasibleonmost omputersystems.
Toimproveon standardMonte Carloa ontrol me hanismneedsto bein-
trodu ed that steer the samples towardsthe relevant partof the statespa e,
therebyin reasing therelevan eof ea h sample. There are severalwaysto do
this,forinstan ebyimportan esamplingdes ribedbrieybelow,orbysplitting
[14℄.
1.1.4 Importan esampling
Thesimulationmethodofimportan esampling omesasaremedytotheprob-
lemarisinginrare-eventsimulation. TheunderlyingproblemoftheMonteCarlo
simulationforrare-eventstudiesis thefa t thatwegettoofew samples in the
importantpartofthe outputspa e,meaningthatwegettoofewsampleswhere
{X ∈ A}
. The basi idea of importan e sampling is that insteadof samplingfrom the original distribution
F
theX 1 , . . . , X n
are sampled from a so- alledsamplingdistribution,say
G
. Thesampling distributionG
is hosensu hthatweobtainmoresampleswhere
{X ∈ A}
. Theimportan esamplingisthensim-plytheaverageofhittingtheevent,weightedwiththerelevantRadon-Nikodym
derivative,
b p
IS= 1
n X n i=1
dF
dG I{X i ∈ A}
.Thisisaunbiasedand onsistentestimatorsin e
E G [b p
IS] = Z
A
dF
dG dG = P(X ∈ A)
.Themaindi ultyinimportan esamplingistodesignthesamplingdistri-
bution. Traditionally the fun tionality and reliabilityof new sto hasti simu-
lation algorithmsisproved byrunningextensivenumeri alexperiments. But
numeri al eviden e alone is insu ient. There are numerous exampleswhere
the standard heuristi s fail and the numeri al eviden e indi ates that the al-
gorithm has onverged when, in fa t, it is severely biased [24℄. The limited
eviden e providedbysimplyrunningnumeri alexperimentshasgenerated the
need for a deeper theoreti al understanding and analysis of the performan e
of sto hasti simulationalgorithms. Overthe last de ade mathemati al tools
from stability theoryand ontrol theory havebeen developed withthe aimto
theoreti ally quantify the performan e of sto hasti simulation algorithms for
omputingprobabilitiesof rareevents. Inthe ontext ofimportan esampling
two main approa hes have been studied; the subsolution approa h, based on
ontrol theory, by Dupuis, Wang, and ollaborators, see e.g. [18, 19, 17℄, and
the approa h based on Lyapunov fun tions and stability theory by Blan het,
Glynn, andothers,see[5,6,7,10℄.
Inthetheoreti alworkone ientimportan esamplinganalgorithmissaid
to bee ient ifrelativeerrorpersample,Std
(b p)/p
doesnotgrowtoo rapidlyas
p ↓ 0
.1.1.5 Heavy-taileddistributions
Inthisthesiswe onsider inparti ular probabilitydistributions
F
withheavy-tails. Thenotionofheavytailsreferstotherateofde ayofthetail
F = 1 − F
ofadistributionfun tion
F
. Apopular lassofheavy-taileddistributionsisthe lass of subexponentialdistributions. A distribution fun tionF
supported onthepositiveaxisissaidto belongtothesubexponentialdistributions if
x→∞ lim
P(X 1 + X 2 > x)
P(X 1 > x) = 2
,forindependentrandom variables
X 1
andX 2
with distributionF
. A sub lassofthe subexponentialdistributions is theregularlyvarying distributions.
F
isalledregularlyvarying(at
∞
)withindex−α ≤ 0
ift→∞ lim F (tx)
F (t) = x −α ,
forallx > 0
.Theheavy-taileddistributions are oftendes ribed withtheonebigjump
analogy,meaningthattheeventofasumofheavy-tailedrandomvariablesbeing
largeis dominated by the aseof one of the variables being very largewhilst
therestarerelativelysmall. Thisisin sharp ontrastto the aseoflight-tails,
where thesame event is dominated by the aseof everyvariable ontributing
equallyto the total. As areferen eto theonebig jump analogywerefer the
readerto[28,30,15℄.
This one big jump phenomena has been observed in empiri al data. For
instan e, when we onsider sto k market indi es su h as Nasdaq, Dow Jones
et . itturnsoutthatthedistributionofdailylogreturnstypi allyhasaheavy
lefttail,see Hultetal.in[29℄. AnotherexampleisthewellstudiedDanishre
insuran e data, whi h onsists of real-life laims aused by industrial res in
Denmark. Whilethearrivalsof laimsisshowedtobenotfarfromPoisson,the
laimsizedistributionshows learheavy-tailbehavior.Thedatasetisanalysed
byMikos h in [43℄and thetailof the laimsize is shown tobet wellwith a
Pareto distribution.
Sto hasti simulationin thepresen eofheavy-taileddistributionshasbeen
studiedwith mu h interestin re entyears. The onditionalMonte Carlote h-
niquewasappliedonthissettingbyAsmussenetal.[2,4℄. Dupuisetal.[16℄used
importan e sampling algorithm in a heavy-tailed setting. Finally we mention
theworkofBlan hetetal. onsideringheavy-taileddistributions in[11,8℄.
1.2 Markov hain Monte Carlo in rare-event simulation
Inthis se tion wedes ribeanewmethodologybasedon Markov hain Monte
Carlo (MCMC), for omputing probabilities of rare events. A more general
versionof the algorithm, for omputingexpe tations, is provided in Se tion 2
alongwithapre iseasymptoti e ien y riteria.
1.2.1 Formulation
Let
X
beareal-valuedrandomvariablewithdistributionF
anddensityf
withrespe ttotheLebesguemeasure. Theproblem isto omputetheprobability
p = P(X ∈ A) = Z
A
dF
. (1.1)Theevent
{X ∈ A}
isthoughtofasrareinthesensethatp
issmall. LetF A
bethe onditionaldistributionof
X
givenX ∈ A
. ThedensityofF A
isgivenbydF A
dx (x) = f (x)I{x ∈ A}
p .
(1.2)ConsideraMarkov hain
(X t ) t≥0
withinvariantdensitygivenby(1.2). Su haMarkov hain anbe onstru ted byimplementing anMCMC algorithm su h
asaGibbssampleroraMetropolis-Hastingsalgorithm, seee.g.[3,23℄.
To onstru tan estimator for the normalising onstant
p
, onsider a non-negativefun tion
v
, whi h isnormalisedin thesense thatR
A v(x)dx = 1
. Thefun tion
v
will be hosenlater aspartof thedesignof theestimator. Foranyhoi e of
v
thesamplemean,1 T
T−1 X
t=0
v(X t )I{X t ∈ A}
f (X t ) ,
anbeviewedasanestimateof
E F A
v(X)I{X ∈ A}
f (X)
= Z
A
v(x) f (x)
f (x) p dx = 1
p Z
A
v(x)dx = 1 p .
Thus,
b q T = 1
T
T−1 X
t=0
u(X t ),
whereu(X t ) = v(X t )I{X t ∈ A}
f (X t )
, (1.3)isanunbiasedestimatorof
q = p −1
. Thenp b T = b q T −1
isanestimatorofp
.The expe tedvalueaboveis omputed undertheinvariantdistribution
F A
of the Markov hain. It is impli itly assumed that the sample size
T
is su-iently largethat theburn-in period, the timeuntil the Markov hain rea hes
stationarity,is negligibleoralternativelythat theburn-in period is dis arded.
Anotherremarkisthatitistheoreti allypossiblethatallthetermsin thesum
in (1.3) arezero, leadingto the estimate
q b T = 0
andthenp b T = ∞
. Toavoidsu hnonsenseone ansimplytake
p b T
astheminimumofq b −1 T
andone.Therearetwoessentialdesign hoi esthatdeterminetheperforman eofthe
algorithm: the hoi e ofthefun tion
v
and thedesignoftheMCMC sampler.Thefun tion
v
inuen esthevarian eofu(X t )
in(1.3)andisthereforeofmainon ernfor ontrollingtherare-eventpropertiesofthealgorithm. Itisdesirable
totake
v
su hthatthenormalisedvarian eoftheestimator,givenbyp 2 Var(b q T )
,isnottoolarge. ThedesignoftheMCMCsampler,ontheotherhand,is ru ial
to ontrolthedependen eoftheMarkov hainandtherebythe onvergen erate
ofthealgorithm asafun tionof thesamplesize. Tospeedupsimulationit is
desirable that the Markov hain mixes fast so that the dependen e dies out
qui kly.
1.2.2 Controllingthe normalisedvarian e
This se tion ontains a dis ussion on how to ontrol the performan e of the
estimator
q b T
by ontrollingitsnormalisedvarian e.Fortheestimator
q b T
tobeusefulitisof ourseimportantthatitsvarian eisnottoolarge. Whentheprobability
p
tobeestimatedissmallitisreasonabletoask that
Var(b q T )
isofsize omparabletoq 2 = p −2
,orequivalently,that the standarddeviationoftheestimatoris roughlyofthesamesizeasp −1
. Tothisendthenormalisedvarian e
p 2 Var(b q T )
isstudied.Letus onsider
Var(b q T )
. Withu(x) = v(x)I{x ∈ A}
f (x) ,
p 2 Var F A (b q T ) = p 2 Var F A
1 T
T X −1 t=0
u(X t )
= p 2 1
T Var F A (u(X 0 )) + 2 T 2
T−1 X
t=0 T X −1 s=t+1
Cov F A (u(X s ), u(X t ))
, (1.4)
Letusforthemomentfo usourattentionontherstterm. It anbewritten
as
p 2
T Var F A u(X 0 )
= p 2 T
E F A
u(X 0 ) 2
− E F A
u(X 0 ) 2
= p 2 T
Z v(x)
f (x) I{x ∈ A} 2
F A (dx) − 1 p 2
= p 2 T
Z v 2 (x)
f 2 (x) I{x ∈ A} f (x) p dx − 1
p 2
= 1
T
Z
A
v 2 (x)p
f (x) dx − 1
.
Therefore,inorderto ontrolthenormalisedvarian ethefun tion
v
mustbehosensothat
R
A v 2 (x)
f(x) dx
is losetop −1
. An importantobservationisthat the onditionaldensity(1.2)playsakeyrolein ndingagood hoi eofv
. Lettingv
bethe onditionaldensityin (1.2)leadstoZ
A
v 2 (x) f (x) dx =
Z
A
f 2 (x)I{x ∈ A}
p 2 f (x) dx = 1 p 2
Z
A
f (x)dx = 1 p
,whi h implies,
p 2
T Var F A u(X)
= 0
.This motivatestaking
v
as an approximationof the onditional density (1.2).Thisissimilartotheideologybehind hoosingane ientimportan esampling
estimator.
Ifforsomeset
B ⊂ A
theprobabilityP(X ∈ B)
anbe omputedexpli itly,thena andidatefor
v
isv(x) = f (x)I{x ∈ B}
P(X ∈ B)
,the onditionaldensityof
X
givenX ∈ B
. This andidateislikelyto performwellif
P(X ∈ B)
isagoodapproximationofp
. Indeed,inthis aseZ
A
v 2 (x) f (x) dx =
Z
A
f 2 (x)I{x ∈ B}
P(X ∈ B) 2 f (x) dx = 1 P(X ∈ B) 2
Z
B
f (x)dx = 1 P(X ∈ B)
,whi h willbe loseto
p −1
.Now,letus shiftemphasisto the ovarian etermin (1.4). Asthe samples
(X t ) T t=0 −1
formaMarkov haintheX t
'saredependent. Thereforethe ovarian etermin(1.4)isnon-zeroandmaynotbeignored. The rudeupperbound
Cov F A (u(X s ), u(X t )) ≤ Var F A (u(X 0 )),
2p 2 T 2
T X −1 t=0
T X −1 s=t+1
Cov F A (u(X s ), u(X t )) ≤ p 2 1 − 1
T
Var F A (u(X 0 ))
forthe ovarian eterm. Thisisavery rudeupperbound asitdoesnotde ay
to zero as
T → ∞
. But, at the moment, the emphasis is on smallp
so wewillpro eedwiththisupperboundanyway. Asindi atedabovethe hoi eof
v
ontrols theterm
p 2 Var F A (u(X 0 ))
. We on ludethat thenormalisedvarian e(1.4)oftheestimator
q b T
is ontrolledbythe hoi eofv
whenp
issmall.1.2.3 Ergodi properties
Aswehavejustseenthe hoi eofthefun tion
v
ontrolsthenormalisedvarian eof theestimator forsmall
p
. Thedesign of theMCMC sampler, onthe otherhand,determinesthestrengthof thedependen ein theMarkov hain. Strong
dependen eimpliesslow onvergen ewhi hresultsinahigh omputational ost.
The onvergen e rate of MCMC samplers an be analysed within the theory
of
ϕ
-irredu ible Markov hains. Fundamentalresults forϕ
-irredu ibleMarkov hainsaregivenin[42,44℄. Wewillfo uson onditionsthatimplyageometrionvergen erate. The onditionsgivenbelowarewellstudiedinthe ontextof
MCMC samplers. Conditionsforgeometri ergodi ityin the ontext ofGibbs
samplers have been studied by e.g. [12, 51, 52℄, and for Metropolis-Hastings
algorithmsby[40℄.
A Markov hain
(X t ) t≥0
with transitionkernelp(x, ·) = P(X t+1 ∈ · | X t = x)
isϕ
-irredu ible if there exists a measureϕ
su h thatP
t p (t) (x, ·) ≪ ϕ(·)
,where
p (t) (x, ·) = P(X t ∈ · | X 0 = x)
denotes thet
-steptransition kerneland≪
denotesabsolute ontinuity. AMarkov hainwithinvariantdistributionπ
isalledgeometri allyergodi ifthereexistsapositivefun tion
M
anda onstantr ∈ (0, 1)
su hthatkp (t) (x, ·) − π(·)k
TV≤ M (x)r t ,
(1.5)where
k · k
TVdenotesthetotal-variationnorm. This onditionensuresthatthe
distributionoftheMarkov hain onvergesatageometri ratetotheinvariant
distribution. Ifthefun tion
M
isbounded,thentheMarkov hainissaidtobeuniformlyergodi . Conditionssu has(1.5)maybedi ulttoestablishdire tly
and are therefore substituted by suitable minorisation or drift onditions. A
minorisation onditionholdsonaset
C
ifthereexist aprobabilitymeasureν
,apositiveinteger
t 0
,andδ > 0
su hthatp (t 0 ) (x, B) ≥ δν(B),
for all
x ∈ C
and Borel setsB
. In this aseC
is said to be a small set.Minorisation onditions havebeen used for obtaining rigorous bounds on the
onvergen eofMCMCsamplers,seee.g.[49℄.
If the entire state spa e is small, then the Markov hain is uniformly er-
godi . Uniformergodi itydoestypi allynothold forMetropolissamplers, see
Mengersen and Tweedie in [40℄ Theorem 3.1. Therefore useful su ient on-
ditions for geometri ergodi ityare often givenin theform of drift onditions
[12, 40℄. Drift onditions,establishedthrough the onstru tionof appropriate
Lyapunovfun tions, are alsouseful for establishing entral limit theorems for
MCMCalgorithms,see[34, 42℄andthereferen estherein.
Roughlyspeaking,theargumentsgivenaboveleadtothefollowingdesiredprop-
ertiesoftheestimator.
1. Rareevent e ien y: Constru tanunbiasedestimator
q b T
ofp −1
a ord-ing to (1.3)by nding a fun tion
v
whi h approximates the onditional density (1.2). The hoi e ofv
ontrols the normalised varian e of theestimator.
2. Large sample e ien y: Design the MCMC sampler, by nding an ap-
propriateGibbssampleroraproposaldensityintheMetropolis-Hastings
algorithm,su hthattheresultingMarkov hainisgeometri allyergodi .
1.3 Outline and ontribution of this thesis
Theoutlineand ontributionofthethesisareasfollows.
a. General formulation of the algorithm in Se tion 2. In this se tion we
presenttheformalmethodology in howto set uptheMCMC simulation
fore ientrare-event omputation. Theprobabilisti assumptionsmade
aremildandthesettingis forinstan e notrestri tedtoheavy-tails. The
twoessentialdesign hoi esarehighlighted. Correspondingtorare-event
e ien yand largesamplee ien y.
b. Appli ationtoheavy-tailedrandomwalksinSe tion3. Inthisse tionthe
MCMCmethodologyisappliedto theproblem of omputing
p n = P(Y 1 + · · · + Y n > a n )
,where
a n → ∞
su ientlyfastsothattheprobabilitytendstozero. The in rementsY
areassumedtobeheavy-tailed. WepresentaGibbssampler toprodu eaMarkov hainwhoseinvariantdistributionisthe onditionaldistribution
P (Y 1 , . . . , Y n ) ∈ · | Y 1 + · · · + Y n > a n
.
TheMarkov hainisshowntopreservestationarityanduniformlyergodi ,
ensuring the largesample e ien y. Inaddition we designan estimator
for
1/p n
having vanishing normalised varian e. Numeri al experiments performed and omparison made between MCMC and best-performingexistingimportan esamplingestimatorsaswellasstandardMonteCarlo.
. Appli ationtoheavy-tailedrandomsumsinSe tion4. Inthisse tionthe
MCMCmethodologyisappliedto theproblem of omputing
p n = P(Y 1 + · · · + Y N n > a N n )
,where
N
isa randomvariable anda N → ∞
su iently fastso that theprobability tends to zero. The in rements
Y
are assumed to be heavy-tailed. We present a Gibbs sampler to produ e a Markov hain whose
invariantdistributionisthe onditionaldistribution
P (N, Y 1 , . . . , Y N ) ∈ · | Y 1 + · · · + Y N > a N
.
ensuring the largesample e ien y. Inaddition wedesignan estimator
for
1/p n
having vanishing normalised varian e. Numeri al experiments performed and omparison made between MCMC and best-performingexistingimportan esamplingestimatorsaswellasstandardMonteCarlo.
d. Appli ationtosto hasti re urrentequationsinSe tion5. Inthisse tion
the MCMC methodology is applied to the problem of omputing
p n = P(X n > a n )
,whereX n = A n X n−1 + B n
,X 0 = 0
,and
a n → ∞
su ientlyfastsothattheprobabilitytendstozero. Thein- rementsB
areassumedtoberegularlyvaryingofindexα
andE[A α+ǫ ] <
∞
forsomeǫ > 0
. WepresentaGibbssamplertoprodu eaMarkov hainwhoseinvariantdistribution isthe onditional distribution
P (A 2 , . . . , A n , B 1 , . . . , B n ) ∈ · | X n > a n
.
TheMarkov hainisshowntopreservestationarityanduniformlyergodi ,
ensuring the largesample e ien y. Inaddition wedesignan estimator
for
1/p n
having vanishing normalised varian e. Numeri al experiments performed and omparison made between MCMC and best-performingexistingimportan esamplingestimatorsaswellasstandardMonteCarlo.
e. Appli ation to omputingprobabilityofruininaninsuran emodel with
riskyinvestmentsinSe tion6...
A papertitled Markov hain Monte Carlo for omputing rare-event proba-
bilities for a heavy-tailed random walk by Gudmundssonand Hult [25℄ based
on Se tions 2, 3, and4 in the thesis hasbeena epted for publi ation in the
JournalofAppliedProbabilityinJune2014.
tion
Inthisse tiontheMarkov hainMonteCarloideasareappliedto theproblem
of omputinganexpe tation. Here thesettingisgeneral,forinstan e, thereis
noassumptionthatdensitieswithrespe ttoLebesguemeasureexist.
Let
X
be a randomvariable with distributionF
andh
be anon-negativeF
-integrablefun tion. Theproblemisto omputetheexpe tationθ = E h(X)
= Z
h(x)dF (x)
.Inthe spe ial asewhen
F
hasdensityf
andh(x) = I{x ∈ A}
this problemredu estothesimplerproblemof omputingtheprobabilityin(1.1). illustrated
inSe tion 1.2.
The analogueof the onditional distribution in (1.2)is thedistribution
F h
givenby
F h (B) = 1 θ
Z
B
h(x)dF (x),
formeasurablesetsB
.ConsideraMarkov hain
(X t ) t≥0
havingF h
asitsinvariantdistribution. To deneanestimatorofθ −1
, onsideraprobabilitydistributionV
withV ≪ F h
.Thenitfollowsthat
V ≪ F
anditisassumedthatthedensitydV /dF
isknown.Considertheestimatorof
ζ = θ −1
givenbyζ b T = 1
T
T X −1 t=0
u(X t )
, whereu(x) = 1 θ
dV dF h
(x)
. (2.1)Notethat
u
doesnotdependonθ
be auseV ≪ F h
andthereforeu(x) = 1 θ
dV dF h
(x) = 1 h(x)
dV dF (x),
for
x
su hthath(x) > 0
. Theestimator(2.1)isageneralisationoftheestimator (1.3) where one an think ofv
as the density ofV
with respe t to Lebesguemeasure. Anestimatorof
θ
an then onstru tedasθ b T = b ζ T −1
.The varian e analysis of
ζ b T
follows pre isely the steps outlined in Se tion1.2. Thenormalisedvarian eis
θ 2 Var F h (b ζ T ) = θ 2
T Var F h u(X 0 ) + 2θ 2
T 2
T X −1 t=0
T−1 X
s=t+1
Cov F h u(X s ), u(X t )
, (2.2)
wheretherstterm anberewritten,similarlyto thedisplay(1.4),as
θ 2
T Var F h u(X 0 )
= 1 T
E V h dV
dF h
i
− 1
.
The analysis above indi ates that an appropriate hoi e of
V
is su h thatE V [ dF dV
h ]
is loseto1
. Again,theideal hoi ewouldbetakingV = F h
leadingtozerovarian e. This hoi eisnotfeasiblebutneverthelesssuggestssele ting
V
asanapproximationof
F h
. Asalreadynotedthisissimilartotheideologybehindhoosingane ientimportan esamplingestimator. Thedieren ebeingthat
here
V ≪ F
isrequiredwhereasinimportan esamplingF
needsbeabsolutelyontinuouswith respe tto thesampling distribution. The rudeupperbound
forthe ovarian etermin (2.2)is valid,justasinSe tion1.2.
Asymptoti e ien y anbe onvenientlyformulatedintermsofalimit riteria
as alarge deviation parametertends to innity. As is ustomaryin problems
relatedtorare-eventsimulationtheproblemathandisembeddedinasequen e
ofproblems,indexedby
n = 1, 2, . . .
. Thegeneralsetupisformalisedasfollows.Let
(X (n) ) n≥1
bea sequen eof random variables withX (n)
having distri-bution
F (n)
. Leth
beanon-negativefun tion,integrablewithrespe ttoF (n)
,forea h
n
. Supposeθ (n) = E
h(X (n) )
= Z
h(x)dF (n) (x) → 0,
as
n → ∞
. Theproblemisto omputeθ (n)
forsomelargen
.Denote by
F h (n)
the distribution withdF h (n) /dF (n) = h/θ (n)
. For then
thproblem, aMarkov hain
(X t (n) ) T t=0 −1
withinvariantdistributionF h (n)
is gener-ated byan MCMCalgorithm. The estimatorof
ζ (n) = (θ (n) ) −1
isbased onaprobabilitydistribution
V (n)
, su hthatV (n) ≪ F h (n)
,withknowndensitywithrespe tto
F (n)
. Anestimatorζ b T (n)
ofζ
isgivenbyζ b T (n) = 1
T
T X −1 t=0
u (n) (X t (n) ),
where
u (n) (x) = 1 h(x)
dV (n) dF (n) (x).
Theheuristi e ien y riteriainSe tions1.2 annowberigorouslyformu-
latedasfollows:
1. Rare-evente ien y: Sele t theprobabilitydistributions
V (n)
su hthat(θ (n) ) 2 Var F (n)
h (u (n) (X)) → 0,
asn → ∞.
2. Largesamplesizee ien y: DesigntheMCMCsampler,byndinganap-
propriateGibbssampleroraproposaldensityfortheMetropolis-Hastings
algorithm, su hthat,forea h
n ≥ 1
, theMarkov hain(X t (n) ) t≥0
isgeo-metri ally ergodi .
Remark 2.1. The rare-event e ien y riteria is formulated in terms of the
e ien y of estimating
(θ (n) ) −1
byζ b T (n)
. If oneinsists on studying the meanandvarian eof
θ b (n) T = (b ζ T (n) ) −1
,thentheee tsofthetransformationx 7→ x −1
must be taken into a ount. Forinstan e, the estimator
b θ (n) T
is biasedand itsvarian e ould beinnite. Thebias anberedu ed forinstan e viathedelta
methodillustratedin [3,p. 76℄. Wealsoremark thatevenin theestimation of
(θ (n) ) −1
byζ b T (n)
thereisabias omingfromthefa tthattheMarkov hainnotbeingperfe tlystationary.
The MCMC methodology presented in Se tion 2 is here applied to ompute
theprobability that arandom walk
S n = Y 1 + · · · + Y n
, whereY 1 , . . . , Y n
arenon-negative,independentandheavy-tailed,ex eedsahigh threshold
a n
. Thisproblemhasre eivedsomeattentionin the ontextof onditionalMonte Carlo
algorithms[2, 4℄andimportan e samplingalgorithms[35,16,11,8℄.
In this se tion a Gibbs sampler is presented for sampling from the on-
ditional distribution
P((Y 1 , . . . , Y n ) ∈ · | S n > a n )
. The resulting Markovhain isprovedtobeuniformlyergodi . An estimatorfor
(p (n) ) −1
oftheform(2.1)issuggestedwith
V (n)
asthe onditionaldistributionof(Y 1 , . . . , Y n )
givenmax{Y 1 , . . . , Y n } > a n
. Theestimatoris proved to havevanishing normalisedvarian ewhenthedistributionof
Y 1
belongstothe lassofsubexponentialdis- tributions. Theproofiselementaryandis ompleted inafewlines. This isinsharp ontrast to e ien y proofs forimportan e sampling algorithms for the
sameproblem,whi hrequiremorerestri tiveassumptionsonthetailof
Y 1
andtend to be long and te hni al [16, 11, 9℄. The se tion is on luded with nu-
meri alexperimentstoillustratethe omparativenesswithexistingimportan e
samplingalgorithmandstandardMonteCarlo.
3.1 A Gibbs sampler for omputing
P(S n > a n )
Let
Y 1 , . . . , Y n
benon-negativeindependentandidenti allydistributed random variables with ommon distributionF Y
and densityf Y
with respe t to somereferen e measure
µ
. Consider therandom walkS n = Y 1 + · · · + Y n
and theproblemof omputingtheprobability
p (n) = P(S n > a n )
,where
a n → ∞
su ientlyfastthatp (n) → 0
asn → ∞
.Itis onvenientto denoteby
Y (n)
then
-dimensionalrandomve torY (n) = (Y 1 , . . . , Y n ) ⊤
,andtheset
A n = {y ∈ R n : 1 ⊤ y > a n }
,where
1 = (1, . . . , 1) ⊤ ∈ R n
andy = (y 1 , . . . , y n ) ⊤
. With thisnotationp (n) = P(S n > a n ) = P(1 ⊤ Y (n) > a n ) = P(Y (n) ∈ A n )
.The onditionaldistribution
F A (n) n (·) = P(Y (n) ∈ · | Y (n) ∈ A n )
,hasdensity
dF A (n) n
dµ (y 1 , . . . , y n ) = Q n
j=1 f Y (y j )I{y 1 + · · · + y n > a n }
p (n)
. (3.1)The rst step towards dening the estimator of
p (n)
is to onstru t theMarkov hain
(Y (n) t ) t≥0
whoseinvariantdensityisgivenby(3.1)usingaGibbssampler. In short,the Gibbs samplerupdates oneelementof
Y t (n)
at atimekeepingtheotherelements onstant. Formallythealgorithmpro eedsasfollows.
Algorithm3.1. Startataninitialstate
Y 0 (n) = (Y 0,1 , . . . , Y 0,n ) ⊤
whereY 0,1 +
· · · + Y 0,n > a n
. GivenY (n) t = (Y t,1 , . . . , Y t,n ) ⊤
, forsomet = 0, 1, . . .
,thenextstate
Y (n) t+1
issampledasfollows:1. Draw
j 1 , . . . , j n
from{1, . . . , n}
withoutrepla ementandpro eedbyup- datingthe omponentsofY (n) t
intheorderthusobtained.2. Forea h
k = 1, . . . , n
,repeatthefollowing.(a) Let
j = j k
betheindextobeupdatedandwriteY t,−j = (Y t,1 , . . . , Y t,j−1 , Y t,j+1 , . . . , Y t,n ) ⊤
.Sample
Y t,j ′
fromthe onditionaldistributionofY
giventhatthesumex eedsthethreshold. That is,
P(Y t,j ′ ∈ · | Y t,−j ) = P
Y ∈ · | Y + X
k6=j
Y t,k > a n
.
(b) Put
Y ′ t = (Y t,1 , . . . , Y t,j−1 , Y t,j ′ , Y t,j+1 , . . . , Y t,n ) ⊤
.3. Drawarandompermutation
π
ofthenumbers{1, . . . , n}
fromtheuniformdistributionandput
Y t+1 (n) = (Y t,π(1) ′ , . . . , Y t,π(n) ′ ) ⊤
.Iteratesteps(1)-(3)untiltheentireMarkov hain
(Y (n) t ) T−1 t=0
is onstru ted.Remark3.2. (i)Intheheavy-tailedsettingthetraje toriesoftherandomwalk
leading to the rare event are likelyto onsist of one largein rement (the big
jump)whiletheotherin rementsareaverage.Thepurposeofthepermutation
step is to for e the Markov hain to mix faster by moving the big jump to
dierentlo ations. However,thepermutationstepinAlgorithm3.1isnotreally
needed when onsidering the probability
P(S n > a n )
. This isdue to thefa tthatthesummationisinvariantoftheorderingofthesteps.
(ii)Thealgorithmrequiressamplingfromthe onditionaldistribution
P(Y ∈
· | Y > c)
for arbitraryc
. This is easy wheneverinversionis feasible, see [3,p. 39℄,ora eptan e/reje tionsampling anbeemployed. Thereare,however,
situations where sampling from the onditional distribution
P(Y ∈ · | Y > c)
maybedi ult,see[33,Se tion2.2℄.
Thefollowingproposition onrmsthattheMarkov hain
(Y (n) t ) t≥0
,gener-atedbyAlgorithm3.1,has
F A (n) n
asitsinvariantdistribution.Proposition 3.3. The Markov hain
(Y t (n) ) t≥0
, generated by Algorithm 3.1,has the onditionaldistribution
F A (n) n
asitsinvariant distribution.Proof. The goalis to show that ea h updating step(Step 2 and 3) of the al-
gorithmpreserves stationarity. Sin e the onditional distribution
F A (n) n
is per-mutation invariantitis learthat Step3preservesstationarity. Thereforeitis
su ientto onsiderStep2ofthealgorithm.
Let
P j (y, ·)
denotethetransitionprobabilityoftheMarkov hain(Y t (n) ) t≥0
orrespondingtothe
j
th omponentbeingupdated. Itissu ienttoshowthat,forall
j = 1, . . . , m
andallBorelsetsofprodu tformB 1 × · · · × B n ⊂ A n
,thefollowingequalityholds:
F A (n) n (B 1 × · · · × B n ) = E F (n)
An [P j (Y, B 1 × · · · × B n )].
Observethat,be ause
B 1 × · · · × B n ⊂ A n
,F A (n) n (B 1 × · · · × B n ) = E h Y n
k=1
I{Y k ∈ B k } | S n > a n
i
= E [I{Y j ∈ B j }I{S n > a n } Q
k6=j I{Y k ∈ B k }]
P(S n > a n )
=
E h E[I{Y j ∈B j }|Y j >a n −S n,−j ,Y (n)
−j ] Q
k6=j I{Y k ∈B k } P(Y j >a n −S n,−j |Y (n) −j )
i
P(S n > a n )
= E[P j (Y (n) , B 1 × · · · × B n ) Q
k6=j I{Y k ∈ B k }]
P (S n > a n )
= E[P j (Y (n) , B 1 × · · · × B n ) | S n > a n ]
= E
F An (n) [P j (Y, B 1 × · · · × B n )]
,withthe onventionalnotationofwriting
Y (n) = (Y 1 , . . . , Y n ) ⊤
,S n = Y 1 + · · · + Y n
,Y −j (n) = (Y 1 , . . . , Y j−1 , Y j+1 , Y n ) ⊤
andS n,−j = Y 1 + · · · + Y j−1 + Y j+1 + · · · + Y n
.Asfortheergodi properties,Algorithm3.1produ esaMarkov hainwhi h
isuniformlyergodi .
Proposition 3.4. For ea h
n ≥ 1
, the Markov hain(Y (n) t ) t≥0
is uniformlyergodi . In parti ular, it satises the following minorisation ondition: there
exists
δ > 0
su hthatP(Y 1 (n) ∈ B | Y (n) 0 = y) ≥ δF A (n) n (B),
for all
y ∈ A n
andall BorelsetsB ⊂ A n
.Proof. Takean arbitrary
n ≥ 1
. Uniform ergodi ity anbe dedu ed from thefollowing minorisation ondition (see [44℄): there exists a probability measure
ν
,δ > 0
,andanintegert 0
su hthatP(Y (n) t 0 ∈ B | Y (n) 0 = y) ≥ δν(B)
,forevery
y ∈ A n
andBorelsetB ⊂ A n
. Takey ∈ A n
andwriteg( · | y)
forthedensity of
P(Y 1 (n) ∈ · | Y (n) 0 = y)
. Thegoalis to show thatthe minorisation onditionholds witht 0 = 1
,δ = p (n) /n!
,andν = F A (n) n
.Forany
x ∈ A n
thereexistsanorderingj 1 , . . . , j n
ofthenumbers{1, . . . , n}
su h that