**EduPristine (****http://www.edupristine.com/ca****) and Bionic Turtle have entered into a partnership to promote practical applications of the concepts related to risk analytics and modeling. Practical and hands-on understanding of building excel based models related to operational and credit risk is necessary for any job related to risk management. For this purpose, we would be illustrating step by step model building techniques for risk management. Registrations for Operational Risk are OPEN. Please ****Sign-up**** ****now.**

We will now discuss one of the most pertinent areas of Operational Risk Modeling.

Operational risk modeling uses Loss Data Analysis (internal or external), Scenario Analysis and data on Business Environment and Internal Controls. Loss data analysis and Scenario Analysis require fitting probability distribution to loss data or scenario data i.e. identifying which distribution best describes the empirical or expert judgement data.

There are various methods of fitting distributions to data such as:

Moments Matching (estimating distribution parameters such that moments such as mean, variance of the data are matched)

Quantile/Percentile Matching (estimating distribution parameters such that quantiles like 50^{th} Quantile i.e. Median, 99% percentile etc. are matched)

Probability Weighted Moments

Maximum Likelihood Estimation (MLE)

MLE method estimates the distribution parameter such that the joint likelihood of observing all empirical data points together gets maximized. Quite unlike moment matching and quantile matching, MLE makes use of all data points instead of only specific moments/quantiles. MLE also allows fitting distribution to truncated and censored data (common feature of operational risk data).

Start with seed parameter values for the distribution to be fitted to the data

Find the probability of observing each data point (using appropriate PDF function)

Assuming all data points are independent, joint probability of observing all data points together is the product of their probability density functions

*If two events A and B are independent Joint probability of A and B happening together is P(A) x P(B)*

Overall probability of observing all loss data amounts together would be all PDF multiplied together.

We would like to maximize the joint likelihood to observe all data points. This is achieved either

**Maximizing P(A)*P(B) or**

**Maximizing log(P(A)) + log(P(B)) or**

**Minimizing [log(P(A)) + log(P(B))]**

Use an optimization algorithm (like **Excel Solver**) to maximize joint log-likelihood by changing parameter values

Let's take a small operational loss severity data set comprising of only 20 loss data points. Illustration shows estimation of parameters of Gamma distribution using MLE:

Gamma distribution has two parameters:

Shape/Alpha which controls the shape of the distribution and impacts skewness and kurtosis of the distribution and

Scale/Beta which controls the dispersion (variance) of the distribution

For Gamma distribution, shape and scale parameters have to be positive, so we start with positive seed values.

Usually, CDF and PDF functions in Excel have Dist suffixed to the distribution name. For instance, NORMDIST function gives CDF of normal distribution if cumulative argument is TRUE and gives PDF if cumulative argument is entered as FALSE.

Similarly, GAMMADIST gives CDF (cumulative = TRUE) and PDF (cumulative = FALSE) of Gamma distributed random variable.

Joint probability of observing all data points together is calculated as product of PDF of individual data point. We can either

Maximise product of all PDF or

Maximize the sum of logs of individual PDF (maximising log likelihood) or

Minimise the sum of logs of individual PDF (minimizing log likelihood).

We decide to minimize the negative logarithm. Column C we calculate the logs of PDF and in C7 we take negative sum of all log PDF. We then use Excel solver to minimize cell C7 by changing parameter values.

Excel Solver can be used to minimize the log likelihood function on (Cell C7), by changing parameter values (B3 and B4), subject to the constraints that parameters are positive (B3, B4 > = small positive value)

In practice and as allowed in regulatory guidelines, Banks may not collect data on small operational losses below a threshold (called truncation of loss data). MLE parameter estimates need to be corrected for this to avoid bias in estimates.

Say if the loss collection threshold is USD 10000, then:

Find conditional PDF, conditional on the fact that losses above 10000 are only being collected.

To find conditional PDF, we divide PDF by the cumulative probability that losses are above the threshold (as we are capturing only data points above the threshold)

Equivalently we divide by (1 cumulative probability that losses are less than the threshold).

Notice that we are now minimizing cell E7, sum of conditional PDF. Conditional probability of each data point is calculated as

**PDF(Loss Amount)/(1-CDF(Loss threshold))**

**GAMMADIST(Loss amount, Param-1, Param-2, Cumulative = False)/ (1- GAMMADIST(Loss Threshold, Param-1, Param-2, Cumulative = True)).**

Another noteworthy point is that after considering loss threshold, shape parameter has changed. Therefore, not adjusting for loss threshold would lead to biased parameter estimates.

A template has been created for you, where the subheadings are given and you have to link the model to get the cash numbers! You can download the same from here**.** You can go through the case and fill in the yellow boxes. You are also recommended to try to create this structure on your own (so that you get a hang of what information is to be recorded).

Also you can download this filled template and check, if the information you recorded matches what has already been mentioned or not! (Check sheet 2 also)

Excel functions can be used to model operational loss data. Similar techniques can be used for other common distributions fitted to operational loss data, use:

Weibull function for Weibull distribution

Expondist function for exponential distribution

Normdist function for lognormal distribution after taking Natural Log of all data points

Also in the next few tutorials we would find the methodology of fitting distributions to data and the practical problems that we face in the same.

By the way, we have launched a course on Operational Risk Modeling, which covers all these practical concepts through video lectures. If you want to join the same, you can join by clicking here.

If you have queries, post them on our forum!

The case has been drafted for discussion purpose. It has been written by EduPristine (http://www.edupristine.com/ca) and would be discussed by experts from EduPristine & Bionic Turtle. There would be a step by step analysis and financial model building to come to a conclusion on the decision.

To guarantee your success at passing the FRM exam at one go, follow the link to FRM Concept Checkers for many more concepts!

## Comments