SlideShare a Scribd company logo
MAXIMUM ENTROPY APPROACH TO THE LOSS DATA
AGGREGATION PROBLEM
Erika Gomes
epgomes@emp.uc3m.es
joint work with
Silvia Mayoral Henryk Gzyl
Department of Business Administration
Universidad Carlos III de Madrid
June, 2016
1 / 38
Outline
1 Introduction
Motivation
Methodology: Loss distribution Approach
Univariate case
Multivariate case
2 Maximum Entropy Approach
Examples and Applications
Theory
3 Numerical Results
Joint marginal density distributions
Joint vector of losses
4 Conclusions
2 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
The differences between these two sectors lies in the availability of the data. In
Operational risk the size of the historical data is small. So, the results may vary
widely.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
The differences between these two sectors lies in the availability of the data. In
Operational risk the size of the historical data is small. So, the results may vary
widely.
More precisely, we are interested in the calculation of regulatory/economic
capital using advanced models (LDA: loss distribution approach) allowed by
Basel II.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
The differences between these two sectors lies in the availability of the data. In
Operational risk the size of the historical data is small. So, the results may vary
widely.
More precisely, we are interested in the calculation of regulatory/economic
capital using advanced models (LDA: loss distribution approach) allowed by
Basel II.
The problem is the calculation of the amount of money you may need in order
to be hedged at a high level of confidence (VaR: 99,9%).
The regulation states that the allocated capital charge should correspond to a
1-in-1000 (quantile .999) years worst possible loss event.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
The differences between these two sectors lies in the availability of the data. In
Operational risk the size of the historical data is small. So, the results may vary
widely.
More precisely, we are interested in the calculation of regulatory/economic
capital using advanced models (LDA: loss distribution approach) allowed by
Basel II.
The problem is the calculation of the amount of money you may need in order
to be hedged at a high level of confidence (VaR: 99,9%).
The regulation states that the allocated capital charge should correspond to a
1-in-1000 (quantile .999) years worst possible loss event.
It is necessary to calculate the distribution of the losses, and the methodology
used has to take in consideration challenges related with size of the data sets,
bimodality, heavy tails, dependence, between others.
3 / 38
Introduction
Motivation
Banks developed a conceptual framework to characterize and quantify risk, to
put money aside to cover large-scale losses and to ensure the stability of the
financial system.
In this sense, a similar problem appears also in Insurance, to set premiums and
optimal reinsurance levels.
The differences between these two sectors lies in the availability of the data. In
Operational risk the size of the historical data is small. So, the results may vary
widely.
More precisely, we are interested in the calculation of regulatory/economic
capital using advanced models (LDA: loss distribution approach) allowed by
Basel II.
The problem is the calculation of the amount of money you may need in order
to be hedged at a high level of confidence (VaR: 99,9%).
The regulation states that the allocated capital charge should correspond to a
1-in-1000 (quantile .999) years worst possible loss event.
It is necessary to calculate the distribution of the losses, and the methodology
used has to take in consideration challenges related with size of the data sets,
bimodality, heavy tails, dependence, between others.
We propose to model the total losses by maximizing an entropy measure.
3 / 38
Introduction
Loss distribution Approach (LDA): Univariate Case
Operational risk has to do with
losses due failures in processes,
technology, people, etc.
Two variables play a role in operational risk:
Severity (X): Lognormal, gamma,
weibull distributions, Subexponential
distributions...
Frequency (N): Poisson, negative
binomial , binomial distributions.
S = X1 + X2 + ... + XN =
N
n=1
Xn
where S represents the aggregate claim
amount in a fixed time period (typically one
year) per risk event.
Approach used: Fit parametric distributions
to N and X and obtain fS through recursive
models or convolutions
No single distribution fits well over the en-
tire data set.
4 / 38
Introduction
Loss distribution Approach (LDA): Multivariate Case
Then, each of these loss distributions are further summed among all types of risk, to
arrive at the total aggregate loss.
(S1, ..., Sm) = (
N1
i=1
X1i, ...,
Nm
i=1
Xmi)
ST =
m
i=1
Si = S1 + S2 + ... + Sm
where b = 1, . . . , 8(business lines), l = 1, . . . , 7(event types),
m = 8 × 7, types of risks in Operational Risk
Dependence structure between risks Si , i.e. choice of a copula model.
5 / 38
Introduction
Loss distribution Approach (LDA): Illustrative Example
Loss distribution Approach (LDA): Illustrative Example
Estimate parametric distributions for the Frequency N and Severity X for each
individual risk (Maximum-Likelihood Estimation - MLE).
Compound the distributions (Panjer, Convolutions, Fourier, ...). Then, we have
fSi
(Univariate Case)
Then, the density fST
of the sum
ST = S1 + ... + SB
(Multivariate case) can be obtained by a sequential convolution procedure:
1 Derive the distribution of the sum of a pair of values S1 + S2 from the
joint density fS1,S2
(s1, s2) = fS1
(s1)fS2
(s2)c(s1, s2), where C is the copula
model .
2 Apply the convolution integral
fS1+S2
= s1
fS1,S2
(s1, l12 − s1) = s2
fS1,S2
(s2, l12 − s2)
steps (1) and (2) are repeated for the rest of the sum.
6 / 38
Introduction
Loss distribution Approach (LDA): Copulas
It is a method used to introduce dependence among random
variables.
C(u1, u2, ..., un) = P[U1 ≤ u1, U2 ≤ u2, ..., Un ≤ un]
7 / 38
Introduction
Illustrative Example
200 samples of size 10
Aggregation: 7 independent types of risks
What happen when the data is scarce, as
in common in banking?
Problems
Parameter uncertainty.
Bad fit in the tails.
Scarcity of data, impossibility to fit
tails and body separately.
Underestimation of the regulatory
capital charge.
This methodology gives a bad fit
even when re-sampling is an
alternative.
8 / 38
Maximum entropy approach
Illustrative example - size and tail concern
Parametric Approach
3.5 4.0 4.5 5.0 5.5 6.0 6.5 7.001234
Losses
density
True
AVERAGE
Reconstructions
Maxent Approach
Maxent provides a density reconstruction over the entire range of values.
9 / 38
Maximum entropy approach
Illustrative example - bimodal concern
(1)
(2)
Error (1) (2)
MAE 0.02652 0.01291
RMSE 0.03286 0.01647
Table : Errors.
Maxent is able to model asymmetries.
10 / 38
Maximum entropy approach
Question
How can we use maxent methodologies to model dependencies between
different types of risks in the framework of Operational risk?
11 / 38
Maximum entropy approach
Find a probability distribution P on some measure
space (Ω, F), which is absolutely continuous respect
to some (usually σ-finite) measure on Q and on
(Ω, F)
max
P
HQ (P) = −
Ω
ρ(ξ)lnρ(ξ)dQ(ξ)
satisfying
{P << Q such that EP [AX] = Y }
Ω ρ(ξ)dQ(ξ) = 1
This method consist in to find the probability measure which
best represent the current state of knowledge which is the one
with the largest information theoretical entropy.
12 / 38
Maximum entropy approach
Jaynes, 1957
This concept was used by Jaynes (1957) for first time as a method
of statistical inference in the case of a under-determined problem.
For example:
We rolled 1000 times, a six-sided die comes up with an average of
4.7 dots. We want to estimate, as best we can, the probability
distribution of the faces.
There are infinitely many 6-tuples (p1, ..., p6) with pi ≥ 0,
i
pi = 1 and
i
ipi = 4.7
13 / 38
Maximum entropy approach
Entropy in Finance
In the last two decades, this concept which has its original role in
statistical physics, has found important applications in several
fields, especially in finance
It has been an important tool for portfolio selection, as a
measure of the degree of diversification.
In asset pricing, to tackle the problem of extracting asset
probability distributions from limited data.
14 / 38
Maximum Entropy Approach
General overview
The essence of the Maxent method consist in transform a problem of the type
AX = Y X : Ω → C
into a problem of convex optimization, by maximizing the entropy measure.
Where C is a constraint set of possible reconstructions.
Then, we have a
Unique and robust solution
15 / 38
Laplace Transform
In probability theory and statistics, the Laplace transform is defined as expectation of
a random variable.
ψ(α) = E[e−αS
] =
∞
0
e−αs
dFS (s) S ∈ IR+
If any two continuous functions have the same Laplace transform, then those
functions must be identical.
The Laplace transforms of some pdf’s are not easy to invert and there is not
a completely general method which works equally well for all possible
transforms.
16 / 38
Laplace Transform
In probability theory and statistics, the Laplace transform is defined as expectation of
a random variable.
ψ(α) = E[e−αS
] =
∞
0
e−αs
dFS (s) S ∈ IR+
If any two continuous functions have the same Laplace transform, then those
functions must be identical.
All the information about the problem can be compressed in a set of moments
obtained from the Laplace transform, through a change of variables.
ψ(α) = E[e−αS
] = E[Y α
] =
1
0
Y α
dFY (y) with Y = e−St and Y ∈ (0, 1)
The selection of those moments should be in a way that we use only those that
are the more relevant or informative (Lin-1992 and Entropy Convergence
Theorem).
17 / 38
Laplace Transform
We want to model fS with S > 0.
When N = 0 ⇒ S = 0 and we rewrite the Laplace transform as
ψ(α) = E[e−αS
] = P(S = 0) · E[e−αS
|S = 0] + P(S > 0) · E[e−αS
|S > 0]
ψ(α) = E[e−αS
] = P(N = 0) · E[e−αS
|N = 0] + P(N > 0) · E[e−αS
|N > 0]
where P(S = 0) = P(N = 0) = po, then
ψ(α) = E[e−αS
] = po · 1 + (1 − po) · E[e−αS
|N > 0]
µ(α) = E[e−αS
|N > 0] =
ψ(α) − po
1 − po
ψ(α) and po has to be estimated from the data, or any other procedure.
18 / 38
Input of the Methodology
Univariate Case
Thus, the problem becomes to determine fS from the integral constraint after a
change of variables,
E[e−αS
|S > 0] =
1
0
yαj fY (y)dy = µ(αj ), j = 0, ..., K.
Analytical form
ψ(αk ) = E(e−αSt ) = ∞
n=0(ϕX (t))k pn = G(φX (αk )) with αk = α0/k
Numerical form
ψ(αk ) =
1
T
T
i=1
eαk si with αk = α0/k
where
α0 = 1.5 : fractional value, k = 1, ..., K optimal number of moments.
φX (αk ): Laplace transform of X, αk ∈ R+
G(·): probability generation function of the frequencies
ψ(αk ): Laplace transform of the total losses
T: sample size. 19 / 38
Input of the Methodology
Univariate Case
Thus, the problem becomes to determine fS from the integral constraint after a
change of variables,
E[e−αS
|S > 0] =
1
0
yαj fY (y)dy = µ(αj ), j = 0, ..., K.
Analytical form
Fit parametrically the frequency and severity distributions and calculate the Laplace
transform through the probability generation function.
Poisson-Gamma
ψ(αk ) = exp(− (1 − ba
(αk + b)−a
))with αk = α0/k
The quality of the results is linked to how well the data fit to the defined
distributions.
It is not possible to find a closed form of ψ(αk ) for some pdf’s. This is
particular true for long tail pdf’s, as for example the lognormal distribution.
20 / 38
Input of the Methodology
Univariate Case
Thus, the problem becomes to determine fS from the integral constraint after a
change of variables,
E[e−αS
|S > 0] =
1
0
yαj fY (y)dy = µ(αj ), j = 0, ..., K.
Analytical form
ψ(αk ) = E(e−αSt ) = ∞
n=0(ϕX (t))k pn = G(φX (αk )) with αk = α0/k
Numerical form
ψ(αk ) =
1
T
T
i=1
eαk si with αk = α0/k
where
α0 = 1.5 : fractional value, k = 1, ..., K optimal number of moments.
φX (αk ): Laplace transform of X, αk ∈ R+
G(·): probability generation function of the frequencies
ψ(αk ): Laplace transform of the total losses
T: sample size. 21 / 38
Input of the Methodology
Multivariate Case
Here we explore how the maxent methodologies can handle with dependencies
between different types of risks, when
1 The marginal densities for each risk fS1
, ..., fSm are independently modeled, and
the input of the maxent methodology should be adapted to model
dependencies between risks, through a copula model. In order to see the
quality of the results we compare with a sequential convolution procedure.
2 Choose the incorrect copula can harm the fit of the distribution. In the case
where the losses are collected as a joint vector (where the dependencies are
included), the maxent techniques can find directly the density distribution.
We compare the results with a sequential convolution procedure with a correct
and incorrect copula to see the differences.
22 / 38
Input of the Methodology
Multivariate Case
(1) We can aggregate dependencies to our input, knowing each fSi
ψ(α) = E[e−αk (S1+S2+...+SB )] =
=
N−1
i=1
e−(s1i +s2i +...+sBi )αk f (s1i , s2i , ..., sBi )∆s1∆s2...∆sB
where N is the number of partitions used in the discretization and
f (s1, s2, ..., sB ) = c[F1(s1), ..., F1(sB )]
B
i=1
fSi
(xi )
the joint distribution, c is the density of the copula model C, and fS1
, ..., fSB
are
marginal densities.
(2) Simply ψ(α) = 1
T
T
i=1
e−αk (s1i +s2i +...+sBi )
where T is the sample size.
23 / 38
Maximum Entropy Methods
max H(f ) = −
1
0
fY (y)lnfY (y)dy
SME approach. Find the probability density on [0,1]
ψ(α) =
1
0
yαk f (y)dy = µ(αk ) with Y = e−S
where µ =
ψ(αk )−P(N=0)
1−P(N=0)
MEM approach: Extension of the SME approach, allows to include a reference
measure Q, which is a parametric distribution.
SMEE approach: Extension of the SME approach when we assume that the
data has noise.
1
0
yαk f (y)dy ∈ Ck = [ak , bk ] with Y = e−S
These methods consist in to find the probability measure which best represent
the current state of knowledge which is the one with the largest information
theoretical entropy.
24 / 38
Standard Maximum Entropy Method (SME)
In general, the maximum entropy density is obtained by maximizing the entropy
measure
Max H(f ) = −
1
0
fY (y)lnfY (y)dy
satisfying
E(yαk ) =
1
0
yαk fY (y)dy = µαk , k = 1, 2, ..., K with K = 8
1
0 fY (y)dy = 1
where
µk : k-th moment, which it is positive and it is a known value
K = 8: number of moments
Fractional value: αk = α0/k, α0 = 1.5
25 / 38
Standard Maximum Entropy Method (SME)
When the problem has a solution can be expressed in terms of the Lagrange
multipliers as
f ∗
Y (y) =
1
Z(λ)
exp −
K
k=1
λk yαk = exp −
K
k=0
λk yαk
the normalization constant is determined by
Z(λ) =
Ω
exp −
K
k=1
λk yαk dy
Then it is necessary to find λ∗, that is the minimizer of the dual entropy that is
in function of the Lagrange multipliers λ and is given by
H(λ) = lnZ(λ)+ < λ, µ >= (λ, µ)
Basically it is a problem of minimizing a convex function and there have to re-
duce the step size as it progresses (Barzilai and Borwein non-monotone gradient
method)
f ∗
Y (y) =
1
Z(λ∗)
exp −
K
k=1
λ∗
k yαk Y ∈ (0, 1)
26 / 38
Standard Maximum Entropy Method (SME)
1 The starting point is ∞
0 e−αs dFS (s) = µk , S ∈ (0, ∞)
2 Make a change of variables, setting Y = e−S(N), Y ∈ (0, 1),
3 Find a minimum of the dual entropy that is in function of λ,
min
λ
(λ, µ) = lnZ(λ)+ < λ, µ >
where
Z(λ) =
1
0
e− K
k=1 λk yαk
dy.
4 The solution
f ∗
Y (y) =
1
Z(λ∗
)
e− K
k=1 λ∗
k yαk
= e− K
k=0 λ∗
k yαk
Y ∈ (0, 1)
5 Return the change of variables
f ∗
S (s) = e−s
f ∗
Y (e−s
), S ∈ (0, ∞)
27 / 38
Numerical Results
Multivariate Case
Here we explore how the maxent methodologies can handle with dependencies
between different types of risks, when
1 The marginal densities for each risk fS1
, ..., fSm are independently modeled, and
the input of the maxent methodology should be adapted to model
dependencies between risks, through a copula model. In order to see the
quality of the results we compare with a sequential convolution procedure.
2 Choose the incorrect copula can harm the fit of the distribution. In the case
where the losses are collected as a joint vector (where the dependencies are
included), the maxent techniques can find directly the density distribution.
We compare the results with a sequential convolution procedure with a correct
and incorrect copula to see the differences.
28 / 38
Numerical Results
To test the methodology we consider different combinations of frequencies and
severity losses.
We use a sample large enough for not to worry about the effect of the size in
the results.
We use several methods to verify the quality of the results: L1 & L2 distances,
MAE & RMSE distances, visual comparisons, and goodness of fit tests.
MAE =
1
T
T
n=1
|ˆF(xn) − Fe (xn)|
RMSE =
1
T
T
n=1
ˆF(xn) − Fe (xn)
2
RMSE is more sensitive to outliers, because this measure gives a relatively high
weight to large errors. So, the greater the difference between MAE and RMSE,
the greater the variance of the individual errors in the sample.
29 / 38
Numerical Results
S
Density
1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
01234
(a) S1
S
Density
1.5 2.0 2.5 3.0
0.00.51.01.52.02.53.0
(b) S2
S
Density
4.0 4.5 5.0 5.5 6.0 6.5 7.0
0.00.51.01.5
(c) S3
Figure : Losses for each line of activity, reconstructed by SME
Error
S MAE RMSE
S1 0.0054 0.0072
S2 0.0241 0.0282
S3 0.0061 0.0071
30 / 38
31 / 38
Error
Copula MAE RMSE
Independent 0.027 0.039
Gaussian, ρ = 0.5 0.004 0.005
Gaussian, ρ = 0.8 0.004 0.005
t − student, ρ = 0.7 , ν = 10 0.004 0.005
32 / 38
Numerical Results
Multivariate Case
Here we explore how the maxent methodologies can handle with dependencies
between different types of risks, when
1 The marginal densities for each risk fS1
, ..., fSm are independently modeled, and
the input of the maxent methodology should be adapted to model
dependencies between risks, through a copula model. In order to see the
quality of the results we compare with a sequential convolution procedure.
2 Choose the incorrect copula can harm the fit of the distribution. In the case
where the losses are collected as a joint vector (where the dependencies are
included), the maxent techniques can find directly the density distribution.
We compare the results with a sequential convolution procedure with a correct
and incorrect copula to see the differences.
33 / 38
Gaussian copula with weak Negative correlation
S
Density
6.5 7.0 7.5 8.0 8.5 9.0
0.00.51.01.5
SME (Gaussian)
Convolution (t−student)
Convolution (Gaussian)
Error SME Convolution(Gaussian) Convolution (t-Student)
MAE 0.007989 0.01071 0.01430
RMSE 0.009605 0.01264 0.01652
Table : Errors SME approach.
34 / 38
Computation of the regulatory capital
VaR
Approaches Errors Conf. Interval 95%
γ SME Conv. Empirical SME err. Conv.err. VaRinf VaRsup
0.900 7.237 7.293 7.236 0.001 0.057 7.212 7.263
0.950 7.399 7.293 7.365 0.034 0.072 7.309 7.389
0.990 7.682 7.569 7.658 0.024 0.089 7.516 7.689
0.995 7.803 7.707 7.719 0.084 0.012 7.595 7.856
0.999 8.175 8.534 8.601 0.426 0.067 7.689 8.926
Table : Comparison of VaR for the SME & Convolution approach
35 / 38
Computation of the regulatory capital
TVaRγ(S) = E[S|S > VaRγ]
TVaR
Approaches Errors Conf. Interval 95%
γ SME Conv. Emp. SME err. Conv.err. TVaRinf TVaRsup
0.900 7.439 7.404 7.419 0.020 0.015 7.336 7.536
0.950 7.571 7.514 7.549 0.022 0.035 7.415 7.735
0.990 7.892 7.837 7.920 0.028 0.083 7.551 8.443
0.995 8.052 8.089 8.047 0.005 0.042 7.578 8.926
0.999 8.529 8.758 8.334 0.195 0.424 7.658 8.926
Table : Comparison of TVaR for the SME & Convolution approaches.
36 / 38
Conclusions
In this work we present an application of the maxent methodologies to
Operational Risk. We showed that this methodology can provide a good density
reconstruction over the entire range, in the case of scarcity, heavy tails and
asymmetries, using only eight moments as input of the methodology.
This methodology gives the possibility to obtain the density distributions from
different levels of aggregation and allow us to include dependencies between
different types of risks.
We can joint marginal densities obtained from any methodology and give
them any relation or
we can obtain the joint distribution directly from the data and avoid bad
estimations
The estimation of the underlying loss process provides a starting point to design
policies, set premiums and reserves, calculate optimal reinsurance levels and
calculate risk pressures for solvency purposes in insurance and risk management.
Also, this is useful in structural engineering to describe the accumulated damage
of a structure, just to mention one more possible application.
37 / 38
QUESTIONS, COMMENTS
38 / 38
Ad

More Related Content

What's hot (20)

NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
Christian Robert
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forests
Christian Robert
 
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Christian Robert
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like sampler
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
Christian Robert
 
4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar
Christian Robert
 
06 Machine Learning - Naive Bayes
06 Machine Learning - Naive Bayes06 Machine Learning - Naive Bayes
06 Machine Learning - Naive Bayes
Andres Mendez-Vazquez
 
CS229 Machine Learning Lecture Notes
CS229 Machine Learning Lecture NotesCS229 Machine Learning Lecture Notes
CS229 Machine Learning Lecture Notes
Eric Conner
 
Yes III: Computational methods for model choice
Yes III: Computational methods for model choiceYes III: Computational methods for model choice
Yes III: Computational methods for model choice
Christian Robert
 
NIPS2007: learning using many examples
NIPS2007: learning using many examplesNIPS2007: learning using many examples
NIPS2007: learning using many examples
zukun
 
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game TheoryUncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Rikiya Takahashi
 
Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Learning for exploration-exploitation in reinforcement learning. The dusk of ...Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Université de Liège (ULg)
 
18.1 combining models
18.1 combining models18.1 combining models
18.1 combining models
Andres Mendez-Vazquez
 
Meta-learning of exploration-exploitation strategies in reinforcement learning
Meta-learning of exploration-exploitation strategies in reinforcement learningMeta-learning of exploration-exploitation strategies in reinforcement learning
Meta-learning of exploration-exploitation strategies in reinforcement learning
Université de Liège (ULg)
 
ABC workshop: 17w5025
ABC workshop: 17w5025ABC workshop: 17w5025
ABC workshop: 17w5025
Christian Robert
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chapters
Christian Robert
 
Lecture10 - Naïve Bayes
Lecture10 - Naïve BayesLecture10 - Naïve Bayes
Lecture10 - Naïve Bayes
Albert Orriols-Puig
 
MaxEnt 2009 talk
MaxEnt 2009 talkMaxEnt 2009 talk
MaxEnt 2009 talk
Christian Robert
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systems
recsysfr
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
Christian Robert
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forests
Christian Robert
 
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Discussion of ABC talk by Stefano Cabras, Padova, March 21, 2013
Christian Robert
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like sampler
Christian Robert
 
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
Christian Robert
 
4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar
Christian Robert
 
CS229 Machine Learning Lecture Notes
CS229 Machine Learning Lecture NotesCS229 Machine Learning Lecture Notes
CS229 Machine Learning Lecture Notes
Eric Conner
 
Yes III: Computational methods for model choice
Yes III: Computational methods for model choiceYes III: Computational methods for model choice
Yes III: Computational methods for model choice
Christian Robert
 
NIPS2007: learning using many examples
NIPS2007: learning using many examplesNIPS2007: learning using many examples
NIPS2007: learning using many examples
zukun
 
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game TheoryUncertainty Awareness in Integrating Machine Learning and Game Theory
Uncertainty Awareness in Integrating Machine Learning and Game Theory
Rikiya Takahashi
 
Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Learning for exploration-exploitation in reinforcement learning. The dusk of ...Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Learning for exploration-exploitation in reinforcement learning. The dusk of ...
Université de Liège (ULg)
 
Meta-learning of exploration-exploitation strategies in reinforcement learning
Meta-learning of exploration-exploitation strategies in reinforcement learningMeta-learning of exploration-exploitation strategies in reinforcement learning
Meta-learning of exploration-exploitation strategies in reinforcement learning
Université de Liège (ULg)
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chapters
Christian Robert
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systems
recsysfr
 

Viewers also liked (14)

SSL12 Entropy
SSL12 EntropySSL12 Entropy
SSL12 Entropy
Keith Vaugh
 
The importance of strong entropy for iot
The importance of strong entropy for iotThe importance of strong entropy for iot
The importance of strong entropy for iot
Arm
 
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
RMS
 
Maxentropic and quantitative methods in operational risk modeling
Maxentropic and quantitative methods in operational risk modelingMaxentropic and quantitative methods in operational risk modeling
Maxentropic and quantitative methods in operational risk modeling
Erika G. G.
 
Marnus Koorts - Masters Project
Marnus Koorts - Masters ProjectMarnus Koorts - Masters Project
Marnus Koorts - Masters Project
Marnus Koorts
 
NLP Secrets Seminar 2012 - Quantum Energy - 21062012
NLP Secrets Seminar 2012 - Quantum Energy - 21062012NLP Secrets Seminar 2012 - Quantum Energy - 21062012
NLP Secrets Seminar 2012 - Quantum Energy - 21062012
Grant Hamel
 
Heat transfer seminar
Heat transfer seminarHeat transfer seminar
Heat transfer seminar
amaljo joju e
 
Information entropy
Information entropyInformation entropy
Information entropy
Oliver Reichenstein
 
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Spark Summit
 
New shell & tube heat exchanger
New shell & tube heat exchangerNew shell & tube heat exchanger
New shell & tube heat exchanger
vatsalpateln
 
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS  ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
Vineeth Sundar
 
Entropy
EntropyEntropy
Entropy
Krushal Kakadia
 
10.3 - Second law of thermodynamics
10.3 - Second law of thermodynamics10.3 - Second law of thermodynamics
10.3 - Second law of thermodynamics
simonandisa
 
Al2O3 Nanofluid
Al2O3 NanofluidAl2O3 Nanofluid
Al2O3 Nanofluid
Suman Krishna Kanth
 
The importance of strong entropy for iot
The importance of strong entropy for iotThe importance of strong entropy for iot
The importance of strong entropy for iot
Arm
 
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
#StumpStumpo: Challenge RMS to model any re/insurance contract in RMS's Contr...
RMS
 
Maxentropic and quantitative methods in operational risk modeling
Maxentropic and quantitative methods in operational risk modelingMaxentropic and quantitative methods in operational risk modeling
Maxentropic and quantitative methods in operational risk modeling
Erika G. G.
 
Marnus Koorts - Masters Project
Marnus Koorts - Masters ProjectMarnus Koorts - Masters Project
Marnus Koorts - Masters Project
Marnus Koorts
 
NLP Secrets Seminar 2012 - Quantum Energy - 21062012
NLP Secrets Seminar 2012 - Quantum Energy - 21062012NLP Secrets Seminar 2012 - Quantum Energy - 21062012
NLP Secrets Seminar 2012 - Quantum Energy - 21062012
Grant Hamel
 
Heat transfer seminar
Heat transfer seminarHeat transfer seminar
Heat transfer seminar
amaljo joju e
 
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Modeling Catastrophic Events in Spark: Spark Summit East Talk by Georg Hofman...
Spark Summit
 
New shell & tube heat exchanger
New shell & tube heat exchangerNew shell & tube heat exchanger
New shell & tube heat exchanger
vatsalpateln
 
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS  ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
ENHANCEMENT OF HEAT TRANSFER IN SHELL AND TUBE EXCHANGER USING NANO FLUIDS
Vineeth Sundar
 
10.3 - Second law of thermodynamics
10.3 - Second law of thermodynamics10.3 - Second law of thermodynamics
10.3 - Second law of thermodynamics
simonandisa
 
Ad

Similar to A Maximum Entropy Approach to the Loss Data Aggregation Problem (20)

V. pacáková, d. brebera
V. pacáková, d. breberaV. pacáková, d. brebera
V. pacáková, d. brebera
logyalaa
 
Risk Aggregation Inanoglu Jacobs 6 09 V1
Risk Aggregation Inanoglu Jacobs 6 09 V1Risk Aggregation Inanoglu Jacobs 6 09 V1
Risk Aggregation Inanoglu Jacobs 6 09 V1
Michael Jacobs, Jr.
 
Team 1 post-challenge final report
Team 1 post-challenge final reportTeam 1 post-challenge final report
Team 1 post-challenge final report
Gaukhar Shaimerdenova
 
Portfolio Optimization Using MAD
Portfolio Optimization Using MADPortfolio Optimization Using MAD
Portfolio Optimization Using MAD
Filippas Beteniotis
 
chain leader
chain leaderchain leader
chain leader
marhadi .
 
Va R Is Wrong !
Va R Is Wrong !Va R Is Wrong !
Va R Is Wrong !
Abdelilah BOUALAM
 
Historical Simulation with Component Weight and Ghosted Scenarios
Historical Simulation with Component Weight and Ghosted ScenariosHistorical Simulation with Component Weight and Ghosted Scenarios
Historical Simulation with Component Weight and Ghosted Scenarios
simonliuxinyi
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data science
pujashri1975
 
Xue paper-01-13-12
Xue paper-01-13-12Xue paper-01-13-12
Xue paper-01-13-12
Yuhong Xue
 
Error-proneness and social security design
Error-proneness and social security designError-proneness and social security design
Error-proneness and social security design
GRAPE
 
A value at risk framework for longevity risk printversion 0
A value at risk framework for longevity risk printversion 0A value at risk framework for longevity risk printversion 0
A value at risk framework for longevity risk printversion 0
Okuda Boniface
 
Study of the Barrier Probability in a Classical Risk Model for two different ...
Study of the Barrier Probability in a Classical Risk Model for two different ...Study of the Barrier Probability in a Classical Risk Model for two different ...
Study of the Barrier Probability in a Classical Risk Model for two different ...
ijsrd.com
 
Optimal investment and reinsurance for mean-variance insurers under variance ...
Optimal investment and reinsurance for mean-variance insurers under variance ...Optimal investment and reinsurance for mean-variance insurers under variance ...
Optimal investment and reinsurance for mean-variance insurers under variance ...
International Journal of Business Marketing and Management (IJBMM)
 
Predictive Modeling in Insurance in the context of (possibly) big data
Predictive Modeling in Insurance in the context of (possibly) big dataPredictive Modeling in Insurance in the context of (possibly) big data
Predictive Modeling in Insurance in the context of (possibly) big data
Arthur Charpentier
 
Skepticism
SkepticismSkepticism
Skepticism
Elsa von Licy
 
Ruin Theory
Ruin TheoryRuin Theory
Ruin Theory
David Rechavel
 
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
NABIH IBRAHIM BAWAZIR
 
Chronology
ChronologyChronology
Chronology
NABIH IBRAHIM BAWAZIR
 
Measuring credit risk in a large banking system: econometric modeling and emp...
Measuring credit risk in a large banking system: econometric modeling and emp...Measuring credit risk in a large banking system: econometric modeling and emp...
Measuring credit risk in a large banking system: econometric modeling and emp...
SYRTO Project
 
Mean-Distortion Risk Measure Portfolio Optimization under the Distribution Un...
Mean-Distortion Risk Measure Portfolio Optimization under the Distribution Un...Mean-Distortion Risk Measure Portfolio Optimization under the Distribution Un...
Mean-Distortion Risk Measure Portfolio Optimization under the Distribution Un...
International Journal of Business Marketing and Management (IJBMM)
 
V. pacáková, d. brebera
V. pacáková, d. breberaV. pacáková, d. brebera
V. pacáková, d. brebera
logyalaa
 
Risk Aggregation Inanoglu Jacobs 6 09 V1
Risk Aggregation Inanoglu Jacobs 6 09 V1Risk Aggregation Inanoglu Jacobs 6 09 V1
Risk Aggregation Inanoglu Jacobs 6 09 V1
Michael Jacobs, Jr.
 
Portfolio Optimization Using MAD
Portfolio Optimization Using MADPortfolio Optimization Using MAD
Portfolio Optimization Using MAD
Filippas Beteniotis
 
chain leader
chain leaderchain leader
chain leader
marhadi .
 
Historical Simulation with Component Weight and Ghosted Scenarios
Historical Simulation with Component Weight and Ghosted ScenariosHistorical Simulation with Component Weight and Ghosted Scenarios
Historical Simulation with Component Weight and Ghosted Scenarios
simonliuxinyi
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data science
pujashri1975
 
Xue paper-01-13-12
Xue paper-01-13-12Xue paper-01-13-12
Xue paper-01-13-12
Yuhong Xue
 
Error-proneness and social security design
Error-proneness and social security designError-proneness and social security design
Error-proneness and social security design
GRAPE
 
A value at risk framework for longevity risk printversion 0
A value at risk framework for longevity risk printversion 0A value at risk framework for longevity risk printversion 0
A value at risk framework for longevity risk printversion 0
Okuda Boniface
 
Study of the Barrier Probability in a Classical Risk Model for two different ...
Study of the Barrier Probability in a Classical Risk Model for two different ...Study of the Barrier Probability in a Classical Risk Model for two different ...
Study of the Barrier Probability in a Classical Risk Model for two different ...
ijsrd.com
 
Predictive Modeling in Insurance in the context of (possibly) big data
Predictive Modeling in Insurance in the context of (possibly) big dataPredictive Modeling in Insurance in the context of (possibly) big data
Predictive Modeling in Insurance in the context of (possibly) big data
Arthur Charpentier
 
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
A Short Glimpse Intrododuction to Multi-Period Fuzzy Bond Imunization for Con...
NABIH IBRAHIM BAWAZIR
 
Measuring credit risk in a large banking system: econometric modeling and emp...
Measuring credit risk in a large banking system: econometric modeling and emp...Measuring credit risk in a large banking system: econometric modeling and emp...
Measuring credit risk in a large banking system: econometric modeling and emp...
SYRTO Project
 
Ad

Recently uploaded (20)

DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
jqrbt
 
GCF - Master Presentation - UK GCF - 0525 .pdf
GCF - Master Presentation - UK GCF - 0525 .pdfGCF - Master Presentation - UK GCF - 0525 .pdf
GCF - Master Presentation - UK GCF - 0525 .pdf
hkmd5mqzjb
 
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdfThe Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
nickysharmasucks
 
GCF - Master Presentation - UK - 0525.pdf
GCF - Master Presentation - UK - 0525.pdfGCF - Master Presentation - UK - 0525.pdf
GCF - Master Presentation - UK - 0525.pdf
hkmd5mqzjb
 
Моніторинг ІТ сектору України GET_UKR_PB_03_2025.pdf
Моніторинг ІТ сектору України GET_UKR_PB_03_2025.pdfМоніторинг ІТ сектору України GET_UKR_PB_03_2025.pdf
Моніторинг ІТ сектору України GET_UKR_PB_03_2025.pdf
Інститут економічних досліджень та політичних консультацій
 
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docxTelegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Henry Tapper
 
stereochemistryuyguhihihihihuiihiuuhiuhiuhi
stereochemistryuyguhihihihihuiihiuuhiuhiuhistereochemistryuyguhihihihihuiihiuuhiuhiuhi
stereochemistryuyguhihihihihuiihiuuhiuhiuhi
4mg22ec401
 
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
Adnet Communications
 
Lundin Gold Corporate Presentation - May 2025
Lundin Gold Corporate Presentation -  May 2025Lundin Gold Corporate Presentation -  May 2025
Lundin Gold Corporate Presentation - May 2025
Adnet Communications
 
20250514-assessing-the-uk-retirement-income-market-vfm.pdf
20250514-assessing-the-uk-retirement-income-market-vfm.pdf20250514-assessing-the-uk-retirement-income-market-vfm.pdf
20250514-assessing-the-uk-retirement-income-market-vfm.pdf
Henry Tapper
 
LCP-Pensions-Powerbrokers-04-2025.pdf excellent
LCP-Pensions-Powerbrokers-04-2025.pdf excellentLCP-Pensions-Powerbrokers-04-2025.pdf excellent
LCP-Pensions-Powerbrokers-04-2025.pdf excellent
Henry Tapper
 
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docxTelegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Henry Tapper
 
Weekly Regional & Global Insights (1 - 8 May 2025)
Weekly Regional & Global Insights (1 - 8 May 2025)Weekly Regional & Global Insights (1 - 8 May 2025)
Weekly Regional & Global Insights (1 - 8 May 2025)
Maher Itani
 
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdfGCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
hkmd5mqzjb
 
2025 0507 Macro Trends and their impact on Enterprise AI.pptx
2025 0507 Macro Trends and their impact on Enterprise AI.pptx2025 0507 Macro Trends and their impact on Enterprise AI.pptx
2025 0507 Macro Trends and their impact on Enterprise AI.pptx
Sunil Grover
 
GCF - Master Presentation - UK GCF - 0525.pdf
GCF - Master Presentation - UK GCF - 0525.pdfGCF - Master Presentation - UK GCF - 0525.pdf
GCF - Master Presentation - UK GCF - 0525.pdf
hkmd5mqzjb
 
GCF - Master Presentation Fundraising - UK - 0525.pdf
GCF - Master Presentation Fundraising - UK - 0525.pdfGCF - Master Presentation Fundraising - UK - 0525.pdf
GCF - Master Presentation Fundraising - UK - 0525.pdf
hkmd5mqzjb
 
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Philip Caputo Toms River
 
report in economics abcdefghijklmnop.pptx
report in economics abcdefghijklmnop.pptxreport in economics abcdefghijklmnop.pptx
report in economics abcdefghijklmnop.pptx
msalvador21
 
Format Meeting Bulanan Minimalist Aesthetic
Format Meeting Bulanan Minimalist AestheticFormat Meeting Bulanan Minimalist Aesthetic
Format Meeting Bulanan Minimalist Aesthetic
frenkywhijaya
 
DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
DeFi Revolution: How JQRBT Players Can Benefit from Aave's Record-Breaking $4...
jqrbt
 
GCF - Master Presentation - UK GCF - 0525 .pdf
GCF - Master Presentation - UK GCF - 0525 .pdfGCF - Master Presentation - UK GCF - 0525 .pdf
GCF - Master Presentation - UK GCF - 0525 .pdf
hkmd5mqzjb
 
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdfThe Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
The Future of Debt Recovery_ Poonawalla Fincorp’s AI Revolution.pdf
nickysharmasucks
 
GCF - Master Presentation - UK - 0525.pdf
GCF - Master Presentation - UK - 0525.pdfGCF - Master Presentation - UK - 0525.pdf
GCF - Master Presentation - UK - 0525.pdf
hkmd5mqzjb
 
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docxTelegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025 2.docx
Henry Tapper
 
stereochemistryuyguhihihihihuiihiuuhiuhiuhi
stereochemistryuyguhihihihihuiihiuuhiuhiuhistereochemistryuyguhihihihihuiihiuuhiuhiuhi
stereochemistryuyguhihihihihuiihiuuhiuhiuhi
4mg22ec401
 
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
2025-05-08 - Torex Gold - Corporate Presentation - May 2025.pdf
Adnet Communications
 
Lundin Gold Corporate Presentation - May 2025
Lundin Gold Corporate Presentation -  May 2025Lundin Gold Corporate Presentation -  May 2025
Lundin Gold Corporate Presentation - May 2025
Adnet Communications
 
20250514-assessing-the-uk-retirement-income-market-vfm.pdf
20250514-assessing-the-uk-retirement-income-market-vfm.pdf20250514-assessing-the-uk-retirement-income-market-vfm.pdf
20250514-assessing-the-uk-retirement-income-market-vfm.pdf
Henry Tapper
 
LCP-Pensions-Powerbrokers-04-2025.pdf excellent
LCP-Pensions-Powerbrokers-04-2025.pdf excellentLCP-Pensions-Powerbrokers-04-2025.pdf excellent
LCP-Pensions-Powerbrokers-04-2025.pdf excellent
Henry Tapper
 
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docxTelegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Telegraph - 'Rachel Reeves paves way for Dutch-style pensions' May 2025.docx
Henry Tapper
 
Weekly Regional & Global Insights (1 - 8 May 2025)
Weekly Regional & Global Insights (1 - 8 May 2025)Weekly Regional & Global Insights (1 - 8 May 2025)
Weekly Regional & Global Insights (1 - 8 May 2025)
Maher Itani
 
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdfGCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
GCF - Our added Value in Cosmetics & Fragrances sector 0425.pdf
hkmd5mqzjb
 
2025 0507 Macro Trends and their impact on Enterprise AI.pptx
2025 0507 Macro Trends and their impact on Enterprise AI.pptx2025 0507 Macro Trends and their impact on Enterprise AI.pptx
2025 0507 Macro Trends and their impact on Enterprise AI.pptx
Sunil Grover
 
GCF - Master Presentation - UK GCF - 0525.pdf
GCF - Master Presentation - UK GCF - 0525.pdfGCF - Master Presentation - UK GCF - 0525.pdf
GCF - Master Presentation - UK GCF - 0525.pdf
hkmd5mqzjb
 
GCF - Master Presentation Fundraising - UK - 0525.pdf
GCF - Master Presentation Fundraising - UK - 0525.pdfGCF - Master Presentation Fundraising - UK - 0525.pdf
GCF - Master Presentation Fundraising - UK - 0525.pdf
hkmd5mqzjb
 
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Commerce on the Rise_ Toms River’s Economic Revival by Philip Caputo Toms Riv...
Philip Caputo Toms River
 
report in economics abcdefghijklmnop.pptx
report in economics abcdefghijklmnop.pptxreport in economics abcdefghijklmnop.pptx
report in economics abcdefghijklmnop.pptx
msalvador21
 
Format Meeting Bulanan Minimalist Aesthetic
Format Meeting Bulanan Minimalist AestheticFormat Meeting Bulanan Minimalist Aesthetic
Format Meeting Bulanan Minimalist Aesthetic
frenkywhijaya
 

A Maximum Entropy Approach to the Loss Data Aggregation Problem

  • 1. MAXIMUM ENTROPY APPROACH TO THE LOSS DATA AGGREGATION PROBLEM Erika Gomes epgomes@emp.uc3m.es joint work with Silvia Mayoral Henryk Gzyl Department of Business Administration Universidad Carlos III de Madrid June, 2016 1 / 38
  • 2. Outline 1 Introduction Motivation Methodology: Loss distribution Approach Univariate case Multivariate case 2 Maximum Entropy Approach Examples and Applications Theory 3 Numerical Results Joint marginal density distributions Joint vector of losses 4 Conclusions 2 / 38
  • 3. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. 3 / 38
  • 4. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. 3 / 38
  • 5. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. The differences between these two sectors lies in the availability of the data. In Operational risk the size of the historical data is small. So, the results may vary widely. 3 / 38
  • 6. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. The differences between these two sectors lies in the availability of the data. In Operational risk the size of the historical data is small. So, the results may vary widely. More precisely, we are interested in the calculation of regulatory/economic capital using advanced models (LDA: loss distribution approach) allowed by Basel II. 3 / 38
  • 7. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. The differences between these two sectors lies in the availability of the data. In Operational risk the size of the historical data is small. So, the results may vary widely. More precisely, we are interested in the calculation of regulatory/economic capital using advanced models (LDA: loss distribution approach) allowed by Basel II. The problem is the calculation of the amount of money you may need in order to be hedged at a high level of confidence (VaR: 99,9%). The regulation states that the allocated capital charge should correspond to a 1-in-1000 (quantile .999) years worst possible loss event. 3 / 38
  • 8. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. The differences between these two sectors lies in the availability of the data. In Operational risk the size of the historical data is small. So, the results may vary widely. More precisely, we are interested in the calculation of regulatory/economic capital using advanced models (LDA: loss distribution approach) allowed by Basel II. The problem is the calculation of the amount of money you may need in order to be hedged at a high level of confidence (VaR: 99,9%). The regulation states that the allocated capital charge should correspond to a 1-in-1000 (quantile .999) years worst possible loss event. It is necessary to calculate the distribution of the losses, and the methodology used has to take in consideration challenges related with size of the data sets, bimodality, heavy tails, dependence, between others. 3 / 38
  • 9. Introduction Motivation Banks developed a conceptual framework to characterize and quantify risk, to put money aside to cover large-scale losses and to ensure the stability of the financial system. In this sense, a similar problem appears also in Insurance, to set premiums and optimal reinsurance levels. The differences between these two sectors lies in the availability of the data. In Operational risk the size of the historical data is small. So, the results may vary widely. More precisely, we are interested in the calculation of regulatory/economic capital using advanced models (LDA: loss distribution approach) allowed by Basel II. The problem is the calculation of the amount of money you may need in order to be hedged at a high level of confidence (VaR: 99,9%). The regulation states that the allocated capital charge should correspond to a 1-in-1000 (quantile .999) years worst possible loss event. It is necessary to calculate the distribution of the losses, and the methodology used has to take in consideration challenges related with size of the data sets, bimodality, heavy tails, dependence, between others. We propose to model the total losses by maximizing an entropy measure. 3 / 38
  • 10. Introduction Loss distribution Approach (LDA): Univariate Case Operational risk has to do with losses due failures in processes, technology, people, etc. Two variables play a role in operational risk: Severity (X): Lognormal, gamma, weibull distributions, Subexponential distributions... Frequency (N): Poisson, negative binomial , binomial distributions. S = X1 + X2 + ... + XN = N n=1 Xn where S represents the aggregate claim amount in a fixed time period (typically one year) per risk event. Approach used: Fit parametric distributions to N and X and obtain fS through recursive models or convolutions No single distribution fits well over the en- tire data set. 4 / 38
  • 11. Introduction Loss distribution Approach (LDA): Multivariate Case Then, each of these loss distributions are further summed among all types of risk, to arrive at the total aggregate loss. (S1, ..., Sm) = ( N1 i=1 X1i, ..., Nm i=1 Xmi) ST = m i=1 Si = S1 + S2 + ... + Sm where b = 1, . . . , 8(business lines), l = 1, . . . , 7(event types), m = 8 × 7, types of risks in Operational Risk Dependence structure between risks Si , i.e. choice of a copula model. 5 / 38
  • 12. Introduction Loss distribution Approach (LDA): Illustrative Example Loss distribution Approach (LDA): Illustrative Example Estimate parametric distributions for the Frequency N and Severity X for each individual risk (Maximum-Likelihood Estimation - MLE). Compound the distributions (Panjer, Convolutions, Fourier, ...). Then, we have fSi (Univariate Case) Then, the density fST of the sum ST = S1 + ... + SB (Multivariate case) can be obtained by a sequential convolution procedure: 1 Derive the distribution of the sum of a pair of values S1 + S2 from the joint density fS1,S2 (s1, s2) = fS1 (s1)fS2 (s2)c(s1, s2), where C is the copula model . 2 Apply the convolution integral fS1+S2 = s1 fS1,S2 (s1, l12 − s1) = s2 fS1,S2 (s2, l12 − s2) steps (1) and (2) are repeated for the rest of the sum. 6 / 38
  • 13. Introduction Loss distribution Approach (LDA): Copulas It is a method used to introduce dependence among random variables. C(u1, u2, ..., un) = P[U1 ≤ u1, U2 ≤ u2, ..., Un ≤ un] 7 / 38
  • 14. Introduction Illustrative Example 200 samples of size 10 Aggregation: 7 independent types of risks What happen when the data is scarce, as in common in banking? Problems Parameter uncertainty. Bad fit in the tails. Scarcity of data, impossibility to fit tails and body separately. Underestimation of the regulatory capital charge. This methodology gives a bad fit even when re-sampling is an alternative. 8 / 38
  • 15. Maximum entropy approach Illustrative example - size and tail concern Parametric Approach 3.5 4.0 4.5 5.0 5.5 6.0 6.5 7.001234 Losses density True AVERAGE Reconstructions Maxent Approach Maxent provides a density reconstruction over the entire range of values. 9 / 38
  • 16. Maximum entropy approach Illustrative example - bimodal concern (1) (2) Error (1) (2) MAE 0.02652 0.01291 RMSE 0.03286 0.01647 Table : Errors. Maxent is able to model asymmetries. 10 / 38
  • 17. Maximum entropy approach Question How can we use maxent methodologies to model dependencies between different types of risks in the framework of Operational risk? 11 / 38
  • 18. Maximum entropy approach Find a probability distribution P on some measure space (Ω, F), which is absolutely continuous respect to some (usually σ-finite) measure on Q and on (Ω, F) max P HQ (P) = − Ω ρ(ξ)lnρ(ξ)dQ(ξ) satisfying {P << Q such that EP [AX] = Y } Ω ρ(ξ)dQ(ξ) = 1 This method consist in to find the probability measure which best represent the current state of knowledge which is the one with the largest information theoretical entropy. 12 / 38
  • 19. Maximum entropy approach Jaynes, 1957 This concept was used by Jaynes (1957) for first time as a method of statistical inference in the case of a under-determined problem. For example: We rolled 1000 times, a six-sided die comes up with an average of 4.7 dots. We want to estimate, as best we can, the probability distribution of the faces. There are infinitely many 6-tuples (p1, ..., p6) with pi ≥ 0, i pi = 1 and i ipi = 4.7 13 / 38
  • 20. Maximum entropy approach Entropy in Finance In the last two decades, this concept which has its original role in statistical physics, has found important applications in several fields, especially in finance It has been an important tool for portfolio selection, as a measure of the degree of diversification. In asset pricing, to tackle the problem of extracting asset probability distributions from limited data. 14 / 38
  • 21. Maximum Entropy Approach General overview The essence of the Maxent method consist in transform a problem of the type AX = Y X : Ω → C into a problem of convex optimization, by maximizing the entropy measure. Where C is a constraint set of possible reconstructions. Then, we have a Unique and robust solution 15 / 38
  • 22. Laplace Transform In probability theory and statistics, the Laplace transform is defined as expectation of a random variable. ψ(α) = E[e−αS ] = ∞ 0 e−αs dFS (s) S ∈ IR+ If any two continuous functions have the same Laplace transform, then those functions must be identical. The Laplace transforms of some pdf’s are not easy to invert and there is not a completely general method which works equally well for all possible transforms. 16 / 38
  • 23. Laplace Transform In probability theory and statistics, the Laplace transform is defined as expectation of a random variable. ψ(α) = E[e−αS ] = ∞ 0 e−αs dFS (s) S ∈ IR+ If any two continuous functions have the same Laplace transform, then those functions must be identical. All the information about the problem can be compressed in a set of moments obtained from the Laplace transform, through a change of variables. ψ(α) = E[e−αS ] = E[Y α ] = 1 0 Y α dFY (y) with Y = e−St and Y ∈ (0, 1) The selection of those moments should be in a way that we use only those that are the more relevant or informative (Lin-1992 and Entropy Convergence Theorem). 17 / 38
  • 24. Laplace Transform We want to model fS with S > 0. When N = 0 ⇒ S = 0 and we rewrite the Laplace transform as ψ(α) = E[e−αS ] = P(S = 0) · E[e−αS |S = 0] + P(S > 0) · E[e−αS |S > 0] ψ(α) = E[e−αS ] = P(N = 0) · E[e−αS |N = 0] + P(N > 0) · E[e−αS |N > 0] where P(S = 0) = P(N = 0) = po, then ψ(α) = E[e−αS ] = po · 1 + (1 − po) · E[e−αS |N > 0] µ(α) = E[e−αS |N > 0] = ψ(α) − po 1 − po ψ(α) and po has to be estimated from the data, or any other procedure. 18 / 38
  • 25. Input of the Methodology Univariate Case Thus, the problem becomes to determine fS from the integral constraint after a change of variables, E[e−αS |S > 0] = 1 0 yαj fY (y)dy = µ(αj ), j = 0, ..., K. Analytical form ψ(αk ) = E(e−αSt ) = ∞ n=0(ϕX (t))k pn = G(φX (αk )) with αk = α0/k Numerical form ψ(αk ) = 1 T T i=1 eαk si with αk = α0/k where α0 = 1.5 : fractional value, k = 1, ..., K optimal number of moments. φX (αk ): Laplace transform of X, αk ∈ R+ G(·): probability generation function of the frequencies ψ(αk ): Laplace transform of the total losses T: sample size. 19 / 38
  • 26. Input of the Methodology Univariate Case Thus, the problem becomes to determine fS from the integral constraint after a change of variables, E[e−αS |S > 0] = 1 0 yαj fY (y)dy = µ(αj ), j = 0, ..., K. Analytical form Fit parametrically the frequency and severity distributions and calculate the Laplace transform through the probability generation function. Poisson-Gamma ψ(αk ) = exp(− (1 − ba (αk + b)−a ))with αk = α0/k The quality of the results is linked to how well the data fit to the defined distributions. It is not possible to find a closed form of ψ(αk ) for some pdf’s. This is particular true for long tail pdf’s, as for example the lognormal distribution. 20 / 38
  • 27. Input of the Methodology Univariate Case Thus, the problem becomes to determine fS from the integral constraint after a change of variables, E[e−αS |S > 0] = 1 0 yαj fY (y)dy = µ(αj ), j = 0, ..., K. Analytical form ψ(αk ) = E(e−αSt ) = ∞ n=0(ϕX (t))k pn = G(φX (αk )) with αk = α0/k Numerical form ψ(αk ) = 1 T T i=1 eαk si with αk = α0/k where α0 = 1.5 : fractional value, k = 1, ..., K optimal number of moments. φX (αk ): Laplace transform of X, αk ∈ R+ G(·): probability generation function of the frequencies ψ(αk ): Laplace transform of the total losses T: sample size. 21 / 38
  • 28. Input of the Methodology Multivariate Case Here we explore how the maxent methodologies can handle with dependencies between different types of risks, when 1 The marginal densities for each risk fS1 , ..., fSm are independently modeled, and the input of the maxent methodology should be adapted to model dependencies between risks, through a copula model. In order to see the quality of the results we compare with a sequential convolution procedure. 2 Choose the incorrect copula can harm the fit of the distribution. In the case where the losses are collected as a joint vector (where the dependencies are included), the maxent techniques can find directly the density distribution. We compare the results with a sequential convolution procedure with a correct and incorrect copula to see the differences. 22 / 38
  • 29. Input of the Methodology Multivariate Case (1) We can aggregate dependencies to our input, knowing each fSi ψ(α) = E[e−αk (S1+S2+...+SB )] = = N−1 i=1 e−(s1i +s2i +...+sBi )αk f (s1i , s2i , ..., sBi )∆s1∆s2...∆sB where N is the number of partitions used in the discretization and f (s1, s2, ..., sB ) = c[F1(s1), ..., F1(sB )] B i=1 fSi (xi ) the joint distribution, c is the density of the copula model C, and fS1 , ..., fSB are marginal densities. (2) Simply ψ(α) = 1 T T i=1 e−αk (s1i +s2i +...+sBi ) where T is the sample size. 23 / 38
  • 30. Maximum Entropy Methods max H(f ) = − 1 0 fY (y)lnfY (y)dy SME approach. Find the probability density on [0,1] ψ(α) = 1 0 yαk f (y)dy = µ(αk ) with Y = e−S where µ = ψ(αk )−P(N=0) 1−P(N=0) MEM approach: Extension of the SME approach, allows to include a reference measure Q, which is a parametric distribution. SMEE approach: Extension of the SME approach when we assume that the data has noise. 1 0 yαk f (y)dy ∈ Ck = [ak , bk ] with Y = e−S These methods consist in to find the probability measure which best represent the current state of knowledge which is the one with the largest information theoretical entropy. 24 / 38
  • 31. Standard Maximum Entropy Method (SME) In general, the maximum entropy density is obtained by maximizing the entropy measure Max H(f ) = − 1 0 fY (y)lnfY (y)dy satisfying E(yαk ) = 1 0 yαk fY (y)dy = µαk , k = 1, 2, ..., K with K = 8 1 0 fY (y)dy = 1 where µk : k-th moment, which it is positive and it is a known value K = 8: number of moments Fractional value: αk = α0/k, α0 = 1.5 25 / 38
  • 32. Standard Maximum Entropy Method (SME) When the problem has a solution can be expressed in terms of the Lagrange multipliers as f ∗ Y (y) = 1 Z(λ) exp − K k=1 λk yαk = exp − K k=0 λk yαk the normalization constant is determined by Z(λ) = Ω exp − K k=1 λk yαk dy Then it is necessary to find λ∗, that is the minimizer of the dual entropy that is in function of the Lagrange multipliers λ and is given by H(λ) = lnZ(λ)+ < λ, µ >= (λ, µ) Basically it is a problem of minimizing a convex function and there have to re- duce the step size as it progresses (Barzilai and Borwein non-monotone gradient method) f ∗ Y (y) = 1 Z(λ∗) exp − K k=1 λ∗ k yαk Y ∈ (0, 1) 26 / 38
  • 33. Standard Maximum Entropy Method (SME) 1 The starting point is ∞ 0 e−αs dFS (s) = µk , S ∈ (0, ∞) 2 Make a change of variables, setting Y = e−S(N), Y ∈ (0, 1), 3 Find a minimum of the dual entropy that is in function of λ, min λ (λ, µ) = lnZ(λ)+ < λ, µ > where Z(λ) = 1 0 e− K k=1 λk yαk dy. 4 The solution f ∗ Y (y) = 1 Z(λ∗ ) e− K k=1 λ∗ k yαk = e− K k=0 λ∗ k yαk Y ∈ (0, 1) 5 Return the change of variables f ∗ S (s) = e−s f ∗ Y (e−s ), S ∈ (0, ∞) 27 / 38
  • 34. Numerical Results Multivariate Case Here we explore how the maxent methodologies can handle with dependencies between different types of risks, when 1 The marginal densities for each risk fS1 , ..., fSm are independently modeled, and the input of the maxent methodology should be adapted to model dependencies between risks, through a copula model. In order to see the quality of the results we compare with a sequential convolution procedure. 2 Choose the incorrect copula can harm the fit of the distribution. In the case where the losses are collected as a joint vector (where the dependencies are included), the maxent techniques can find directly the density distribution. We compare the results with a sequential convolution procedure with a correct and incorrect copula to see the differences. 28 / 38
  • 35. Numerical Results To test the methodology we consider different combinations of frequencies and severity losses. We use a sample large enough for not to worry about the effect of the size in the results. We use several methods to verify the quality of the results: L1 & L2 distances, MAE & RMSE distances, visual comparisons, and goodness of fit tests. MAE = 1 T T n=1 |ˆF(xn) − Fe (xn)| RMSE = 1 T T n=1 ˆF(xn) − Fe (xn) 2 RMSE is more sensitive to outliers, because this measure gives a relatively high weight to large errors. So, the greater the difference between MAE and RMSE, the greater the variance of the individual errors in the sample. 29 / 38
  • 36. Numerical Results S Density 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 01234 (a) S1 S Density 1.5 2.0 2.5 3.0 0.00.51.01.52.02.53.0 (b) S2 S Density 4.0 4.5 5.0 5.5 6.0 6.5 7.0 0.00.51.01.5 (c) S3 Figure : Losses for each line of activity, reconstructed by SME Error S MAE RMSE S1 0.0054 0.0072 S2 0.0241 0.0282 S3 0.0061 0.0071 30 / 38
  • 38. Error Copula MAE RMSE Independent 0.027 0.039 Gaussian, ρ = 0.5 0.004 0.005 Gaussian, ρ = 0.8 0.004 0.005 t − student, ρ = 0.7 , ν = 10 0.004 0.005 32 / 38
  • 39. Numerical Results Multivariate Case Here we explore how the maxent methodologies can handle with dependencies between different types of risks, when 1 The marginal densities for each risk fS1 , ..., fSm are independently modeled, and the input of the maxent methodology should be adapted to model dependencies between risks, through a copula model. In order to see the quality of the results we compare with a sequential convolution procedure. 2 Choose the incorrect copula can harm the fit of the distribution. In the case where the losses are collected as a joint vector (where the dependencies are included), the maxent techniques can find directly the density distribution. We compare the results with a sequential convolution procedure with a correct and incorrect copula to see the differences. 33 / 38
  • 40. Gaussian copula with weak Negative correlation S Density 6.5 7.0 7.5 8.0 8.5 9.0 0.00.51.01.5 SME (Gaussian) Convolution (t−student) Convolution (Gaussian) Error SME Convolution(Gaussian) Convolution (t-Student) MAE 0.007989 0.01071 0.01430 RMSE 0.009605 0.01264 0.01652 Table : Errors SME approach. 34 / 38
  • 41. Computation of the regulatory capital VaR Approaches Errors Conf. Interval 95% γ SME Conv. Empirical SME err. Conv.err. VaRinf VaRsup 0.900 7.237 7.293 7.236 0.001 0.057 7.212 7.263 0.950 7.399 7.293 7.365 0.034 0.072 7.309 7.389 0.990 7.682 7.569 7.658 0.024 0.089 7.516 7.689 0.995 7.803 7.707 7.719 0.084 0.012 7.595 7.856 0.999 8.175 8.534 8.601 0.426 0.067 7.689 8.926 Table : Comparison of VaR for the SME & Convolution approach 35 / 38
  • 42. Computation of the regulatory capital TVaRγ(S) = E[S|S > VaRγ] TVaR Approaches Errors Conf. Interval 95% γ SME Conv. Emp. SME err. Conv.err. TVaRinf TVaRsup 0.900 7.439 7.404 7.419 0.020 0.015 7.336 7.536 0.950 7.571 7.514 7.549 0.022 0.035 7.415 7.735 0.990 7.892 7.837 7.920 0.028 0.083 7.551 8.443 0.995 8.052 8.089 8.047 0.005 0.042 7.578 8.926 0.999 8.529 8.758 8.334 0.195 0.424 7.658 8.926 Table : Comparison of TVaR for the SME & Convolution approaches. 36 / 38
  • 43. Conclusions In this work we present an application of the maxent methodologies to Operational Risk. We showed that this methodology can provide a good density reconstruction over the entire range, in the case of scarcity, heavy tails and asymmetries, using only eight moments as input of the methodology. This methodology gives the possibility to obtain the density distributions from different levels of aggregation and allow us to include dependencies between different types of risks. We can joint marginal densities obtained from any methodology and give them any relation or we can obtain the joint distribution directly from the data and avoid bad estimations The estimation of the underlying loss process provides a starting point to design policies, set premiums and reserves, calculate optimal reinsurance levels and calculate risk pressures for solvency purposes in insurance and risk management. Also, this is useful in structural engineering to describe the accumulated damage of a structure, just to mention one more possible application. 37 / 38
  翻译: