Model Management Guidance PDF
Model Management Guidance PDF
Model Management
Guidance
CONTENTS
INTRODUCTION
Section 1 Context and objective
APPENDIX
Numerical thresholds
TABLE OF CONTENTS
Definitions and interpretations .................................................................................................. 5
1 Context and Objective ...................................................................................................... 11
1.1 Regulatory context .................................................................................................... 11
1.2 Objectives ................................................................................................................. 11
1.3 Document structure................................................................................................... 11
1.4 Scope of Application ................................................................................................ 11
1.5 Requirements and timeframe .................................................................................... 12
1.6 Scope of Models ....................................................................................................... 12
2 Rating Models.................................................................................................................... 13
2.1 Scope ........................................................................................................................ 13
2.2 Governance and strategy........................................................................................... 13
2.3 Data collection and analysis ..................................................................................... 15
2.4 Segmentation ............................................................................................................ 15
2.5 Default definition ...................................................................................................... 16
2.6 Rating scale............................................................................................................... 17
2.7 Model construction ................................................................................................... 17
2.8 Model documentation ............................................................................................... 19
2.9 Usage of rating models ............................................................................................. 20
2.10 Rating override ......................................................................................................... 20
2.11 Monitoring and Validation ....................................................................................... 21
3 PD Models .......................................................................................................................... 23
3.1 Scope ........................................................................................................................ 23
3.2 Key definitions and interpretations ........................................................................... 23
3.3 Default rate estimation .............................................................................................. 23
3.4 Rating-to-PD ............................................................................................................. 25
3.5 PIT PD and PD terms structure ................................................................................ 27
3.6 PIT PD with transition matrices................................................................................ 27
3.7 Portfolio scaling approach ........................................................................................ 29
3.8 The Vasicek credit framework.................................................................................. 30
3.9 Validation of PD models .......................................................................................... 31
4 LGD Models ...................................................................................................................... 33
4.1 Scope ........................................................................................................................ 33
4.2 Data collection .......................................................................................................... 34
4.3 Historical realised LGD ............................................................................................ 35
4.4 Analysis of realised LGD ......................................................................................... 38
4.5 TTC LGD modelling ................................................................................................ 39
4.6 PIT LGD modelling .................................................................................................. 42
4.7 Validation of LGD models ....................................................................................... 43
The following terms shall have the meaning assigned to them for the purpose of interpreting these
Standards and the related Guidance:
is known. In the case of off-balance sheet exposures such as credit cards, guarantees, working
capital facilities or derivatives, the EAD is not certain on the date of measurement and should
be estimated with statistical models.
13. EAR (Earning At Risk): Refer to NII.
14. ECL (Expected Credit Loss): Metric supporting the estimation of provisions under IFRS9 to
cover credit risk arising from facilities and bonds in the banking book. It is designed as a
probability-weighted expected loss.
15. Economic Intuition (written in lower case as “economic intuition”): Also referred to as
economic intuition and business sense. Property of a model and its output to be interpreted in
terms and metrics that are commonly employed for business and risk decisions. It also refers
to the property of the model variables and the model outputs to meet the intuition of experts
and practitioners, in such a way that the model can be explained and used to support decision-
making.
16. Effective Challenge: Characteristic of a validation process. An effective model validation
ensures that model defects are suitably identified, discussed and addressed in a timely fashion.
Effectiveness is achieved via certain key features of the validation process such as
independence, expertise, clear reporting and prompt action from the development team.
17. EVE (Economic Value of Equity): It is defined as the difference between the present value
of the institution’s assets minus the present value of liabilities. The EVE is sensitive to
changes in interest rates. It is used in the measurement of interest rate risk in the banking
book.
18. Expert-Based Models (written in lower case as “expert-based models”): Also referred to as
judgemental models, these models rely on the subjective judgement of expert individuals
rather than on quantitative data. In particular, this type of model is used to issue subjective
scores in order to rank corporate clients.
19. Institutions (written in lower case as “institution(s)”): All banks licenced by the CBUAE.
Entities that take deposits from individuals and/or corporations, while simultaneously issuing
loans or capital market securities.
20. LGD (Loss Given Default): Estimation of the potential loss incurred by a lending institution
upon the default of an obligor (or a facility), measured as a percentage of the EAD. It also
refers to the actual loss incurred upon past defaults also expressed as a percentage of EAD.
The observed LGD levels tend to be related to PD levels with various strength of correlation.
21. Limits and limitations (written in lower case as “limits” and “limitations”): Model limits are
thresholds applied to a model’s outputs and/or its parameters in order to control its
performance. Model limitations are boundary conditions beyond which the model ceases to
be accurate.
22. LSI (Large and/or Sophisticated Institutions): This group comprises DSIBs and any other
institutions that are deemed large and/or with mature processes and skills. This categorisation
is defined dynamically based on the outcome of regular banking supervision.
23. Macroeconomic Model (written in lower case as “macroeconomic model” or “macro
model”): Refers to two types of models. (i) A model that links a set of independent macro
variables to another single dependent macro variable or to several other dependent macro
variables or (ii) a model that links a set of independent macro variables to a risk metric (or a
set of risk metrics) such as probabilities of default or any other business metric such as
revenues.
24. Market Data: Refers to the various data attributes of a traded financial instrument reported
by a trading exchange. It includes the quoted value of the instrument and/or the quoted
parameters of that instrument that allow the derivation of its value. It also includes transaction
information including the volume exchanged and the bid-ask spread.
25. Materiality: The materiality of a model represents the financial scope covered by the model
in the context of a given institution. It can be used to estimate the potential loss arising from
model uncertainty (see Model Risk). Model materiality can be captured by various metrics
depending on model types. Typically, total exposure can be used as a metric for credit models.
26. MMG: CBUAE’s Model Management Guidance.
27. MMS: CBUAE’s Model Management Standards.
28. Model (written in lower case as “model”): A quantitative method, system, or approach that
applies statistical, economic, financial, or mathematical theories, techniques, and assumptions
to process input data into quantitative estimates. For the purpose of the MMS and MSG,
models are categorised in to three distinct groups: statistical models, deterministic models
and expert-based models.
29. Model Calibration (written in lower case as “model calibration”): Key step of the model
development process. Model calibration means changing the values of the parameters and/or
the weights of a model, without changing the structure of the model, i.e. without changing the
nature of the variables and their transformations.
30. Model Complexity (written in lower case as “model complexity”): Overall characteristic of
a model reflecting the degree of ease (versus difficulty) with which one can understand the
model conceptual framework, its practical design, calibration and usage. Amongst other
things, such complexity is driven by, the number of inputs, the interactions between variables,
the dependency with other models, the model mathematical concepts and their
implementation.
31. Model Construction (written in lower case as “model construction”): Key step of the model
development process. The construction of a model depends on its nature, i.e. statistical or
deterministic. For the purpose of the MMS and the MMG, model construction means the
following: for statistical models, for a given methodology and a set of data and transformed
variables, it means estimating and choosing, with a degree of confidence, the number and
nature of the dependent variables along with their associated weights or coefficients. For
deterministic models, for a given methodology, it means establishing the relationship between
a set of input variables and an output variable, without statistical confidence intervals.
32. Model Development (written in lower case as “model development”): Means creating a
model by making a set of sequential and recursive decisions according to the steps outlined
in the dedicated sections of the MMS. Model re-development means conducting the model
development steps again with the intention to replace an existing model. The replacement
may, or may not, occur upon re-development.
33. Modelling Decision (written in lower case as “modelling decision”): A modelling decision
is a deliberate choice that determines the core functionality and output of a model. Modelling
decisions relate to each of the steps of the data acquisition, the development and the
implementation phase. In particular, modelling decisions relate to (i) the choice of data,
(ii) the analysis of data and sampling techniques, (iii) the methodology, (iv) the calibration
and (v) the implementation of models. Some modelling decisions are more material than
others. Key modelling decisions refer to decisions with strategic implications and/or with
material consequences on the model outputs.
34. Model Risk: Potential loss faced by institutions from making decisions based on inaccurate
or erroneous outputs of models due to errors in the development, the implementation or the
inappropriate usage of such models. Losses incurred from Model Risk should be understood
in the broad sense as Model Risk has multiple sources. This definition includes direct
quantifiable financial loss but also any adverse consequences on the ability of the institution
to conduct its activities as originally intended, such as reputational damage, opportunity costs
or underestimation of capital. In the context of the MMS and the MMG, Model Risk for a
given model should be regarded as the combination of its materiality and the uncertainty
surrounding its results.
35. Model Selection (written in lower case as “model selection”): This step is part of the
development process. This means choosing a specific model amongst a pool of available
models, each with a different set of variables and parameters.
36. Model Uncertainty (written in lower case as “model uncertainty”): This refers to the
uncertainty surrounding the results generated by a model. Such uncertainty can be quantified
as a confidence interval around the model output values. It is used as a component to estimate
Model Risk.
37. Multivariate Analysis (written in lower case as “multivariate analysis”): For a given set of
data and variables, this is a process of observing the joint distribution of the dependent and
independent variables together and drawing conclusions regarding their degree of correlation
and causality.
38. NII (Net Interest Income): To simplify notations, both Net Interest Income (for conventional
products) and/or Net Profit Income (for Islamic Products) are referred to as “NII”. In this
context, ‘profit’ is assimilated as interest. It is defined as the difference between total interest
income and total interest expense, over a specific time horizon and taking into account
hedging. The change in NII (“∆NII”) is defined as the difference between the NII estimated
with stressed interest rates under various scenarios, minus the NII estimated with the interest
rates as of the portfolio reporting date. ∆NII is also referred to as earnings at risk (“EAR”).
39. NPV (Net Present Value): Present value of future cash flows minus the initial investment, i.e.
the amount that a rational investor is willing to pay today in exchange for receiving these cash
flows in the future. NPV is estimated through a discounting method. It is commonly used to
estimate various metrics for the purpose of financial accounting, risk management and
business decisions.
40. PD (Probability of Default): Probability that an obligor fails to meet its contractual obligation
under the terms of an agreed financing contract. Such probability is computed over a given
horizon, typically 12 months, in which case it is referred to as a 1-year PD. It can also be
computed over longer horizons. This probability can also be defined at several levels of
granularity, including, but not limited to, single facility, pool of facilities, obligor, or
consolidated group level.
41. PD Model (written as “PD model”): This terminology refers to a wide variety of models with
several objectives. Amongst other things, these models include mapping methods from scores
generated by rating models onto probability of defaults. They also include models employed
to estimate the PD or the PD term structure of facilities, clients or pool of clients.
42. PD Term Structure (written as “PD term structure”): Refers to the probability of default
over several time horizons, for instance 2 years, 5 years or 10 years. A distinction is made
between the cumulative PD and the marginal PD. The cumulative PD is the total probability
of default of the obligor over a given horizon. The marginal PD is the probability of default
between two dates in the future, provided that the obligor has survived until the first date.
43. PIT (Point-In-Time) and TTC (Through-The-Cycle): A point-in-time assessment refers to
the value of a metric (typically PD or LGD) that incorporates the current economic conditions.
This contrasts with a through-the-cycle assessment that refers to the value of the same metric
across a period covering one or several economic cycles.
44. Qualitative validation: A review of model conceptual soundness, design, documentation,
and development and implementation process.
45. Quantitative validation: A review of model numerical output, covering at least its accuracy,
degree of conservatism, stability, robustness and sensitivity.
46. Rating / Scoring (written in lower case “rating or scoring”): For the purpose of the MMS and
the MMG, a rating and a score are considered as the same concept, i.e. an ordinal quantity
representing the relative creditworthiness of an obligor (or a facility) on a predefined scale.
‘Ratings’ are commonly used in the context of corporate assessments whilst ‘scores’ are used
for retail client assessments.
47. Restructuring (written in lower case “restructuring”): The definition of restructuring /
rescheduling used for modelling in the context of the MMS and MMG should be understood
as the definition provided in the dedicated CBUAE regulation and, in particular, in the
Circular 28/2010 on the classification of loans, with subsequent amendments to this Circular
and any new CBUAE regulation on this topic.
48. Rating Model (written in lower case “rating model”): The objective of such model is to
discriminate ex-ante between performing clients and potentially non-performing clients. Such
models generally produce a score along an arbitrary scale reflecting client creditworthiness.
This score can subsequently mapped to a probability of default. However, rating models
should not be confused with PD models.
49. Retail Clients (written in lower case as “retail clients”): Retail clients refer to individuals to
whom credit facilities are granted for the following purpose: personal consumer credit
facilities, auto credit facilities, overdraft and credit cards, refinanced government housing
credit facilities, other housing credit facilities, credit facilities against shares to individuals. It
also includes small business credit facilities for which the credit risk is managed using similar
methods as applied for personal credit facilities.
50. Segment (written in lower case as “segment”): Subsets of an institution’s portfolio obtained
by splitting the portfolio by the most relevant dimensions which explain its risk profile.
Typical dimensions include obligor size, industries, geographies, ratings, product types, tenor
and currency of exposure. Segmentation choices are key drivers of modelling accuracy and
robustness.
51. Senior Management: As defined in the CBUAE’s Corporate Governance Regulation for
Banks.
52. Statistical Model (written in lower case as “statistical model”): A statistical model is a
mathematical construction achieved by the application of statistical techniques to samples of
data. The model links one or several dependent variables to one or several independent
variables. The objective of such a model is to predict, with a confidence interval, the values
of the dependent variables given certain values of the independent variables. Examples of
statistical models include rating models or value-at-risk (VaR) models. Statistical models are
different from deterministic models. By construction, statistical models always include a
degree of Model Risk.
53. Tiers: Models are allocated to different groups, or Tiers, depending on their associated Model
Risk.
54. Time series analysis (written in lower case as “time series analysis”): For a given set of data
and variables, this is a process of observing the behaviour of these variables through time.
This can be done by considering each variable individually or by considering the joint pattern
of the variables together.
55. UAT (User Acceptance Testing): Phase of the implementation process during which users
rigorously test the functionalities, robustness, accuracy and reliability of a system containing
a new model before releasing it into production.
56. Variable Transformation (written in lower case as “variable transformation”): Step of the
modelling process involving a transformation of the model inputs before developing a model.
Amongst others, common transformations include (i) relative or absolute differencing
between variables, (ii) logarithmic scaling, (iii) relative or absolute time change, (iv) ranking,
(v) lagging, and (vi) logistic or probit transformation.
57. Wholesale Clients (written in lower case as “wholesale clients”): Wholesale clients refer to
any client that is not considered as a retail client as per the definition of these Standards.
1.1.1 The Risk Management Regulation (Circular No. 153/2018) issued by the Central Bank
of the UAE (“CBUAE”) on 27th May 2018 states that banks must have robust systems
and tools to assess and measure risks.
1.1.2 To set out modelling requirements for licenced banks, the CBUAE has issued Model
Management Standards (“MMS”) and Model Management Guidance (“MMG”). Both
MMS and MMG should be read jointly as they constitute a consistent set of requirements
and guidance, as follows:
(i) The MMS outline general standards applicable to all models and constitute
minimum requirements that must be met by UAE banks.
(ii) The MMG expands on technical aspects that are expected to be implemented by
UAE banks for certain types of models. Given the wide range of models and the
complexity of some, the CBUAE recognises that alternative approaches can be
envisaged on specific technical points. Whilst this MMG neither constitutes
additional legislation or regulation nor replaces or supersedes any legal or
regulatory requirements or statutory obligations, deviations from the MMG should
be clearly justified and will be subject to CBUAE supervisory review.
1.2 Objectives
1.2.1 Both the MMS and MMG share three key objectives. The first objective is to ensure that
models employed by UAE banks meet quality standards to adequately support decision-
making and reduce Model Risk. The second objective is to improve the homogeneity of
model management across UAE banks. The third objective is to mitigate the risk of
potential underestimation of provisions and capital across UAE banks.
1.2.2 The MMG outlines techniques based on commonly accepted practices by practitioners
and academics, internationally. The majority of its content has been subject to
consultation with numerous subject matter experts in the UAE and therefore it also
reflects expected practices amongst UAE institutions.
1.3.1 Each section of the MMG addresses a different type of model. The MMG is constructed
in such a way that the numbering of each article is sequentially and each article is a unique
reference across the entire MMG.
1.3.2 Both the MMS and the MMG contain an appendix summarising the main numerical limits
included throughout each document respectively. Such summary is expected to ease the
implementation and monitoring of these limits by both institution and the CBUAE.
1.4.1 The MMG applies to all licensed banks in the UAE, which are referred to herein as
“institutions”.
1.4.2 The scope of institutions is consistent across the MMS and the MMG. Details about the
scope of institutions are available in the MMS.
1.4.3 Branches or subsidiaries of foreign institutions should apply the most conservative
practices between the MMG and the expectations from their parent company’s regulator.
1.4.4 Institutions with a parent company incorporated in the UAE should ensure that all their
branches and subsidiaries comply with the MMG.
1.5.1 The implementation timeframe is consistent across the MMS and the MMG. Further
details are available in the MMS.
1.6.1 The MMG focuses on the main credit risk models entering the computation of the
Expected Credit Loss in the context of the current accounting requirements, due to their
materiality and their relevance across the vast majority of institutions. The MMG also
provides guidance on other models used for the assessment of interest rate risk in the
banking book and net present values.
1.6.2 The MMG does not impose the use of these models. The MMG outlines minimum
expected practices if institutions decide to use such models, in order to manage Model
Risk appropriately.
1.6.3 As model management matures across UAE institutions, additional model types may be
included in the scope of the MMG in subsequent publications.
2 RATING MODELS
2.1 Scope
2.1.1 The vast majority of institutions employ rating models to assess the credit worthiness of
their obligors. Rating models provide essential metrics used as foundations to multiple
core processes within institutions. Ratings have implications for key decisions, including
but not limited to, risk management, provisioning, pricing, capital allocation and Pillar II
capital assessment. Institutions should pay particular attention to the quality of their rating
models and subsequent PD models, presented in the next section.
2.1.2 Inadequate rating models can result in material financial impacts due to a potentially
incorrect estimation of credit risk. The CBUAE will pay particular attention to suitability
of the design and calibration of rating and PD models. Rating models that are
continuously underperforming even after several recalibrations should be replaced. These
models should no longer be used for decision making and reporting.
2.1.3 For the purpose of the MMG, a rating and a score should be considered as identical
concepts, that is a numerical quantity without units representing the relative
creditworthiness of an obligor or a facility on predefined scale. The main objective of
rating models is to segregate obligors (or facilities) that are likely to perform under their
current contractual obligations from the ones that are unlikely to perform, given a set of
information available at the rating assessment date.
2.1.4 The construction of rating models is well documented by practitioners and in academic
literature. Therefore, it is not the objective of this section to elaborate on the details of
modelling techniques. Rather, this section focuses on minimum expected practices and
the challenging points that should attract institutions’ attention.
2.2.1 The management of rating models should follow all the steps of the model life-cycle
articulated in the MMS. The concept of model ownership and independent validation is
particularly relevant to rating models due to their direct business implications.
2.2.2 It is highly recommended that institutions develop rating models internally based on their
own data. However, in certain circumstances such as for low default portfolios,
institutions may rely on the support from third party providers. This support can take
several forms that are presented below through simplified categorisation. In all cases, the
management and calibration of models should remain the responsibility of institutions.
Consequently, institutions should define, articulate and justify their preferred type of
modelling strategy surrounding rating models. This strategy will have material
implications on the quality, accuracy and reliability of the outputs.
2.2.3 The choice of strategy has a material impact on the methodology employed. Under all
circumstances, institutions remain accountable for the modelling choices embedded in
their rating models and their respective calibrations.
2.2.4 Various combinations of third party contributions exist. These can be articulated around
the supplier’s contribution to the model development, the IT system solution and/or the
data for the purpose of calibration. Simplified categories are presented hereby, for the
purpose of establishing minimum expected practices:
(i) Type 1 – Support for modelling: A third party consultant is employed to build a
rating model based on the institution’s own data. The IT infrastructure is fully
developed internally. In this case, institutions should work in conjunction with
consultants to ensure that sufficient modelling knowledge is retained internally.
Institutions should ensure that the modelling process and the documentation are
compliant with the principles articulated in the MMS.
(ii) Type 2 – Support for modelling and infrastructure: A third party consultant
provides a model embedded in a software that is calibrated based on the
institution’s data. In this case, the institution has less control over the design of the
rating model. The constraints of such approach are as follows:
a. Institutions should ensure that they understand the modelling approach being
provided to them.
b. Institutions should fully assess the risks related to using a system solution
provided by external parties. At a minimum, this assessment should be made
in terms of performance, system security and stability.
c. Institutions should ensure that a comprehensive set of data is archived in
order to perform validations once the model is implemented. This data should
cover both the financial and non-financial characteristics of obligors and the
performance data generated by the model. The data should be stored at a
granular level, i.e. at a factor level, in order to fully assess the performance
of the model.
(iii) Type 3 – Support of modelling, infrastructure and data: In addition to Type 2
support, a third party consultant provides data and/or a ready-made calibration.
This is the weakest form of control by institutions. For such models, the institution
should demonstrate that additional control and validation are implemented in order
to reduce Model Risk. Immediately after the model implementation, the institution
should start collecting internal data (where possible) to support the validation
process. Such validation could result in a material shift in obligors’ rating and lead
to financial implications.
(iv) Type 4 – Various supports: In the case of various supports, the minimum expected
practices are as follows:
a. If a third party provides modelling services, institutions should ensure that
sufficient knowledge is retained internally.
b. If a third party provides software solutions, institutions should ensure that
they have sufficient controls over parameters and that they archive data
appropriately.
c. If a third party provides data for calibration, institutions should take the
necessary steps to collect internal data in accordance with the data
management framework articulated in the MMS.
2.2.5 In conjunction with the choice of modelling strategy, institutions should also articulate
their modelling method of rating models. A range of possible approaches can be
envisaged between two distinct categories: (i) data-driven statistical models that can rely
on both quantitative and qualitative (subjective) factors, or (ii) expert-based models that
rely only on views from experienced individuals without the use of statistical data.
Between these two categories, a range of options exist. Institutions should consciously
articulate the rationale for their modelling approach.
2.2.6 Institutions should aim to avoid purely expert based models, i.e. models with no data
inputs. Purely expert-based models should be regarded as the weakest form of models
and therefore should be seen as the least preferable option. If the portfolio rated by such
a model represents more than 10% of the institution’s loan book (other than facilities
granted to governments and financial institutions), then the institution should demonstrate
that additional control and validation are implemented in order to reduce Model Risk. It
should also ensure that Senior Management and the Board are aware of the uncertainty
arising from such model. Immediately after the model implementation, the institution
should start collecting internal data to support the validation process.
2.3.1 Institutions should manage and collect data for rating models, in compliance with the
MMS. The data collection, cleaning and filtering should be fully documented in such way
that it can be traced by any third party.
2.3.2 A rigorous process for data collection is expected. The type of support strategy presented
in earlier sections has no implications on the need to collect data for modelling and
validation.
2.3.3 For the development of rating models, the data set should include, at a minimum, (i) the
characteristics of the obligors and (ii) their performance, i.e. whether they were flagged
as default. For each rating model, the number of default events included in the data sample
should be sufficiently large to permit the development of a robust model. This minimum
number of defaults will depend on business segments and institutions should demonstrate
that this minimum number is adequate. If the number of defaults is too small, alternative
approaches should be considered.
2.3.4 At a minimum, institutions should ensure that the following components of the data
management process are documented. These components should be included in the scope
of validation of rating models.
(i) Analysis of data sources,
(ii) Time period covered,
(iii) Descriptive statistics about the extracted data,
(iv) Performing and non-performing exposures,
(v) Quality of the financial statements collected,
(vi) Lag of financial statements,
(vii) Exclusions and filters, and
(viii) Final number of performing and defaulted obligors by period.
2.4 Segmentation
2.4.1 Segmentation means splitting a statistical sample into several groups in order to improve
the accuracy of modelling. This concept applies to any population of products or
customers. The choice of portfolio, customer and/or product segmentation has a material
impact on the quality of rating models. Generally, the behavioural characteristics of
obligors and associated default rates depend on their industry and size (for wholesale
portfolios) and on product types (for retail portfolios). Consequently, institutions should
thoroughly justify the segmentation of their rating models as part of the development
process.
2.4.2 The characteristics of obligors and/or products should be homogeneous within each
segment in order to build appropriate models. First, institutions should analyse the
representativeness of the data and pay particular attention to the consistency of obligor
characteristics, industry, size and lending standards. The existence of material industry
bias in data samples should result in the creation of a rating model specific to that
industry. Second, the obligor sample size should be sufficient to meet minimum statistical
performance. Third, definition of default employed to identify default events should also
be homogeneous across the data sample.
2.5.1 Institutions should define and document two definitions of default, employed in two
different contexts: (i) for the purpose of rating model development and (ii) for the purpose
of estimating and calibrating probabilities of defaults. These two definitions of default
can be identical or different, if necessary. The scope of these definitions should cover all
credit facilities and all business segments of the institution. In this process, institutions
should apply the following principles.
2.5.2 For rating models: The definition of default in the context of a rating model is a choice
made to achieve a meaningful discrimination between performing and non-performing
obligors (or facilities). The terminology ‘good’ and ‘bad’ obligors is sometimes
employed by practitioners in the context of this discrimination. Institutions should define
explicitly the definition of default used as the dependent variable when building their
rating models.
(i) This choice should be guided by modelling considerations, not by accounting
considerations. The level of conservatism embedded in the definition of default
used to develop rating models has no direct impact on the institution’s P&L. It
simply supports a better identification of customers unlikely to perform.
(ii) Consequently, institutions can choose amongst several criteria to identify default
events in order to maximise the discriminatory power of their rating models. This
choice should be made within boundaries. At a minimum, they should rely on the
concept of days-past-due (“DPD”). An obligor should be considered in default if
its DPD since the last payment due is greater or equal to 90 or if it is identified as
defaulted by the risk management function of the institution.
(iii) If deemed necessary, institutions can use more conservative DPD thresholds in
order to increase the predictive power of rating models. For low default portfolios,
institutions are encouraged to use lower thresholds, such as 60 days in order to
capture more default events. In certain circumstances, restructuring events can also
be included to test the power of the model to identify early credit events.
2.5.3 For PD estimation: The definition of default in the context of PD estimation has direct
financial implications through provisions, capital assessment and pricing.
(i) This choice should be guided by accounting and regulatory principles. The
objective is to define this event in such a way that it reflects the cost borne by
institutions upon the default of an obligor.
(ii) For that purpose, institutions should employ the definition of default articulated in
the CBUAE credit risk regulation, separately from the MMS and MMG. As the
regulation evolves, institutions should update the definition employed for
modelling and recalibrate their models.
2.6.1 Rating models generally produce an ordinal indicator on a predefined scale representing
creditworthiness. The scores produced by each models should be mapped to a fixed
internal rating scale employed across all aspects of credit risk management, in particular
for portfolio management, provision estimation and capital assessment. The rating scale
should be the result of explicit choices that should be made as part of the model
governance framework outlined in the MMS. At a minimum, the institution’s master
rating scale should comply with the below principles:
(i) The granularity of the scale should be carefully defined in order to support credit
risk management appropriately. An appropriate balance should be found regarding
the number of grades. A number of buckets that is too small will reduce the
accuracy of decision making. A number of buckets that is too large will provide a
false sense of accuracy and could be difficult to use for modelling.
(ii) Institutions should ensure that the distribution of obligors (or exposures) spans
across most rating buckets. High concentration in specific grades should be
avoided, or conversely the usage of too many grades with no obligors should also
be avoided. Consequently, institution may need to redefine their rating grades
differently from rating agencies’ grades, by expanding or grouping certain grades.
(iii) The number of buckets should be chosen in such a way that the obligors’
probability of default in each grade can be robustly estimated (as per the next
section on PD models).
(iv) The rating scale from external rating agencies may be used as a benchmark,
however their granularity may not be the most appropriate for a given institution.
Institutions with a large proportion of their portfolio in non-investment grade rating
buckets should pay particular attention to bucketing choices. They are likely to
require more granular buckets in this portion of the scale to assess their risk more
precisely than with standard scales from rating agencies.
(v) The choice of an institution’s rating scale should be substantiated and documented.
The suitability of rating scale should be assessed on a regular basis as part of model
validation.
2.7.1 The objective of this section is to draw attention to the key challenges and minimum
expected practices to ensure that institutions develop effective rating models. The
development of retail scorecards is a standardised process that all institutions are expected
to understand and implement appropriately on large amounts of data. Wholesale rating
models tend to be more challenging due to smaller population sizes and the complexity
of the factors driving defaults. Consequently, this section related to model construction
focuses on wholesale rating models.
2.7.2 Wholesale rating models should incorporate, at a minimum, financial information and
qualitative inputs. The development process should include a univariate analysis and a
multivariate analysis, both fully documented. All models should be constructed based on
a development sample and tested on a separate validation sample. If this is not possible
in the case of data scarcity, the approach should be justified and approved by the validator.
2.7.3 Quantitative factors: These are characteristics of the obligors that can be assessed
quantitatively, most of which are financial variables. For wholesale rating models,
institutions should ensure that the creation of financial ratios and subsequent variable
2.7.4 Qualitative subjective factors: These are characteristics of the obligor that are not easily
assessed quantitatively, for instance the experience of management or the dependency of
the obligors on its suppliers. The following categories of subjective factors are commonly
used to assess the risk of corporate lending: industry performance, business
characteristics and performance, management character and experience, and quality of
financial reporting and reliability of auditors. The assessment of these factors is generally
achieved via bucketing that relies on experts’ judgement. When using such qualitative
factors, the following principles should apply:
(i) Institutions should ensure that this assessment is based upon a rigorous governance
process. The collection of opinions and views from experienced credit officers
should be treated as a formal data collection process. The data should be subject to
quality control. Erroneous data points should also be removed.
(ii) If the qualitative subjective factors are employed to adjust the outcome of the
quantitative factors, institutions should control and limit this adjustment. Institutions
should demonstrate that the weights given to the expert-judgement section of the
model is appropriate. Institutions should not perform undue rating overrides with
expert judgement.
2.7.5 Univariate analysis: In the context of rating model development, this step involves
assessing the discriminatory power of each quantitative factor independently and
assessing the degree of correlation between these quantitative factors.
(i) The assessment of the discriminatory power should rely on clearly defined metrics,
such as the accuracy ratio (or Gini coefficient). Variables that display no
relationship or counterintuitive relationships with default rates should preferably
be excluded. They can be included in the model only after a rigorous
documentation of the rationale supporting their inclusion.
(ii) Univariate analysis should also involve an estimation of the correlations between
the quantitative factors with the aim to avoid multicolinearity in the next step of
the development.
(iii) The factors should be ranked according to their discriminatory power. The
development team should comment on whether the observed relationship is
meeting economic and business expectations.
2.7.6 Multivariate analysis: This step involves establishing a link between observed defaults
and the most powerful factors identified during the univariate analysis.
(i) Common modelling techniques include, amongst others, logistic regressions and
neural networks. Institutions can chose amongst several methodologies, provided
that the approach is fully understood and documented internally. This is
particularly relevant if third party consultants are involved.
(ii) Institutions should articulate clearly the modelling technique employed and the
process of model selection. When constructing and choosing the most appropriate
model, institutions should pay attention to the following:
(a) The number of variables in the model should be chosen to ensure a right
balance. An insufficient number of variables can lead to a sub-optimal model
2.8.1 Rigorous documentation should be produced for each rating model as explained in the
MMS. The documentation should be sufficiently detailed to ensure that the model can
be fully understood and validated by any independent party.
2.8.2 In addition to the elements articulated in the MMS, the following components should be
included:
(i) Dates: The model development date and implementation date should be explicitly
mentioned in all rating model documentation.
(ii) Materiality: The model materiality should be quantified, for instance as the number
of rated customers and their total corresponding gross exposure.
(iii) Overview: An executive summary with the model rating strategy, the expected
usage, an overview of the model structure and the data set employed to develop
and test the model.
(iv) Key modelling choices: The default definition, the rating scale and a justification
of the chosen segmentation as explained in earlier sections.
(v) Data: A description of the data employed for development and testing, covering
the data sources and the time span covered. The cleaning process should be
explained including the filter waterfall and/or any other adjustments used.
(vi) Methodology: The development approach covering the modelling choices, the
assumptions, limits, the parameter estimation technique. Univariate and
multivariate analyses discussing in detail the construction of factors, their
transformation and their selection.
(vii) Expert judgement inputs: All choices supporting the qualitative factors. Any
adjustments made to the variables or the model based on expert opinions. Any
contributions from consulted parties.
(viii) Validation: Details of testing and validation performed during the development
phase or immediately after.
2.9.1 Upon the roll-out of a new rating model and/or a newly recalibrated (optimised) rating
model, institutions should update client ratings as soon as possible. Institutions should
assign new ratings with the new model to 70% of the existing obligors (entering the
model scope) within six (6) months and to 95% of the existing obligors within nine (9)
months. The assignment of new ratings should be based on financials that have been
updated since the issuance of the previous rating, if they exist. Otherwise prior financials
should be used. This expectation applies to wholesale and retail models.
2.9.2 In order to achieve this client re-rating exercise in a short timeframe, institutions are
expected to rate clients in batches, performed by a team of rating experts, independently
from actual, potential or perceived business line interests.
2.9.3 Institutions should put in place a process to monitor the usage of rating models. At a
minimum, they should demonstrate that the following principles are met:
(i) All ratings should be archived with a date that reflects the last rating update. This
data should be stored in a secure database destined to be employed on a regular
basis to manage the usage of rating models.
(ii) The frequency of rating assignment should be tracked and reported to ensure that
all obligors are rated appropriately in a timely fashion.
(iii) Each rating model should be employed on the appropriate type of obligor defined
in the model documentation. For instance, a model designed to assess large
corporates should not be used to assess small enterprises.
(iv) Institutions should ensure that the individuals assigning and reviewing ratings are
suitably trained and can demonstrate a robust understanding of the rating models.
(v) If the ratings are assigned by the business lines, these should be reviewed and
independently signed-off by the credit department to ensure that the estimation of
ratings is unbiased from short term potential or perceived business line interests.
2.10.1 In the context of the MMG, rating override means rating upgrade or rating downgrade.
Overrides are permitted; however, they should follow a rigorously documented process.
This process should include a clear governance mirroring the credit approval process
based on obligor type and exposure materiality. The decision to override should be
controlled by limits expressed in terms of number of notches and number of times a rating
can be overridden. Eligibility criteria and the causes for override should be clearly
documented. Causes may include, amongst others: (i) events specific to an obligor, (ii)
systemic events in a given industry or region, and/or (iii) changes of a variable that is not
included in the model.
2.10.2 Rating overrides should be monitored and be included in the model validation process.
The validator should estimate the frequency of overrides and the number of notches
between the modelled rating and the new rating. A convenient approach to monitor
overrides is to produce an override matrix.
2.10.3 In some circumstances, the rating of a foreign obligor should not be better than the rating
of its country of incorporation. Such override decision should be justified and
documented.
2.10.4 A contractual guarantee of a parent company can potentially result in the rating
enhancement of an obligor, but conditions apply:
(i) The treatment of parental support for a rating enhancement should be recognised
only based on the production of an independent legal opinion confirming the
enforceability of the guarantee upon default. The rating enhancement should only
apply to the specific facility benefiting from the guarantee. The process for rating
enhancement should be clearly documented. For the avoidance of doubt, a sole
letter of intent from the parent company should not be considered as a guarantee
for enforceability purpose. A formal legal guarantee is the only acceptable
documentation.
(ii) For modelling purpose, an eligible parent guarantee can be reflected in the PD or
the LGD of the facility benefiting from this guarantee. If the rating of the facility
is enhanced, then the guarantee will logically be reflected in the PD parameter. If
the rating of the obligor is not enhanced but the guarantee is deed eligible, then the
guarantee can be reflected in the LGD parameter. The rationale behind such choice
should be fully documented.
2.11.1 Institutions should demonstrate that their rating models are performing over time. All
rating models should be monitored on a regular basis and independently validated
according to all the principles articulated in the MMS. For that purpose, institutions
should establish a list of metrics to estimate the performance and stability of models and
compare these metrics against pre-defined limits.
2.11.2 The choice of metrics to validate rating models should be made carefully. These metrics
should be sufficiently granular and capture performance through time. It is highly
recommended to estimate the change in the model discriminatory power through time,
for instance by considering a maximum acceptable drop in accuracy ratio.
2.11.3 In addition to the requirement articulated in the MMS related to the validation step, for
rating models in particular, institutions should ensure that validation exercises include the
following components:
(i) Development data: A review of the data collection and filtering process performed
during the development phase and/or the last recalibration. In particular, this
should cover the definition of default and data quality.
(ii) Model usage: A review of the governance surrounding model usage. In particular,
the validator should comment on (a) the frequency of rating issuance, (b) the
governance of rating production, and (c) the variability of ratings produced by the
model. The validator should also liaise with the credit department to form a view
on (d) the quality of financial inputs and (e) the consistency of the subjective inputs
and the presence of potential bias.
(iii) Rating override: A review of rating overrides. This point does not apply to newly
developed models.
(iv) Model design: A description of the model design and its mathematical formulation.
A view on the appropriateness of the design, the choice of factors and their
transformations.
(v) Key assumptions: A review of the appropriateness of the key assumptions,
including the default definition, the segmentation and the rating scale employed
when developing the model.
(vi) Validation data: The description of the data set employed for validation.
(vii) Quantitative review: An analysis of the key quantitative indicators covering, at a
minimum, the model stability, discriminatory power, sensitivity and calibration.
This analysis should cover the predictive power of each quantitative and subjective
factor driving the rating.
(viii) Documentation: A review on the quality of the documentation surrounding the
development phase and the modelling decisions.
(ix) Suggestions: When deemed appropriate, the validator can make suggestions for
defect remediation to be considered by the development team.
3 PD MODELS
3.1 Scope
3.1.1 The majority of institutions employ models to estimate the probability of default of their
obligors (or facilities), for risk management purpose and to comply with accounting and
regulatory requirements. These models are generally referred to as ‘PD models’, although
this broad definition covers several types of models. For the purpose of the MMG, and to
ensure appropriate model management, the following components should be considered
as separate models:
(i) Rating-to-PD mapping models, and
(ii) Point-in-Time PD Term Structure models.
3.1.2 These models have implications for key decisions including, but not limited to, risk
management, provisioning, pricing, capital allocation and Pillar II capital assessment.
Institutions should manage these models through a complete life-cycle process in line
with the requirements articulated in the MMS. In particular, the development, ownership
and validation process should be clearly organised and documented.
3.2.1 The following definitions are complementing the definitions provided at the beginning of
the MMG. The probability of default of a borrower or of a facility is noted “PD”. The
loss proportion of exposure arising after default, or “loss given default” is noted “LGD”.
3.2.2 A point-in-time assessment (“PIT”) refers to the value of a metric (typically PD or LGD)
that incorporates the current economic conditions. This contrasts with a through-the-cycle
assessment (“TTC”) that refers to the value of the same metric across a period covering
one or several economic cycles.
3.2.3 A PD is associated with a specific time horizon, which means that the probability of
default is computed over a given period. A 1-year PD refers to the PD over a one year
period, starting today or at any point in the future. A PD Term Structure refers to a
cumulative PD over several years (generally starting at the portfolio estimation date). This
contrasts with a marginal forward 1-year PD, which refers to a PD starting at some point
in the future and covering a one year period, provided that the obligor has survived until
that point.
3.2.4 A rating transition matrix is a square matrix that gives the probabilities to migrate from a
rating to another rating. This probability is expressed over a specific time horizon,
typically one year, in which case we refer to a ‘one-year transition matrix’. Transitions
can also be expressed over several years.
3.3.1 Prior to engaging in modelling, institutions should implement a robust process to compute
time series of historical default rates, for all portfolios where data is available. The results
should be transparent and documented. This process should be governed and approved
by the Model Oversight Committee. Once estimated, historical default rates time series
should only be subject to minimal changes. Any retroactive updates should be approved
by the Model Oversight Committee and by the bank’s risk management committee.
3.3.2 Institutions should estimate default rates at several levels of granularity: (i) for each
portfolio, defined by obligor type or product, and (ii) for each rating grade within each
portfolio, where possible. In certain circumstances, default rate estimation at rating
grade level may not be possible and institutions may only rely on pool level estimation.
In this case, institutions should justify their approach by demonstrating clear evidence
based on data, that grade level estimation is not deemed sufficiently robust.
3.3.3 Institutions should compute the following default ratio, based on the default definition
described in the previous section. This ratio should be computed with an observation
window of 12 months to ensure comparability across portfolios and institutions. In
addition, institutions are free estimate this ratio for other windows (e.g. quarterly) for
specific modelling purposes.
(i) The denominator is composed of performing obligors with any credit obligation,
including off and on balance sheet facilities, at the start of the observation window.
(ii) The numerator is composed of obligors that defaulted at least once during the
observation window, on the same scope of facilities.
Formally the default rate can be expressed as shown by the formula below, where t
represented the date of estimation. Notice that if the ratio is reported at time t, then the
ratio is expressed as a backward looking metrics. This concept is particularly relevant for
the construction of macro models as presented in subsequent sections. The frequency of
computation should be at least quarterly and possibly monthly for some portfolios.
3.3.4 When the default rate is computed by rating grade, the denominator should refer to all
performing obligors assigned to a rating grade at the beginning of the observation
window. When the default rate is computed at segment level, the denominator should
refer to all performing obligors assigned to that segment at the beginning of the
observation window.
3.3.5 For wholesale portfolios, this ratio should be computed in order to obtain quarterly
observations over long time periods covering one or several economic cycles. For
wholesale portfolios, institutions should aim to gather at least 5 years of data, and
preferably longer. For retail portfolios or for portfolios with frequent changes in product
offerings, the period covered may be shorter, but justification should be provided.
3.3.6 Institutions should ensure that default time series are homogeneous and consistent
through time, i.e. relate to a portfolio with similar characteristics, consistent lending
standards and consistent definition of default. Adjustments may be necessary to build
time series representative of the institution current portfolios. Particular attention should
be given to changes in the institution’s business model through time. This is relevant is
the case of rapidly growing portfolios or, conversely, in the case of portfolio run-off
strategies. This point is further explained in the MMG section focusing on macro models.
3.3.7 If an obligor migrates between ratings or between segments during the observation
period, the obligor should be included in the original rating bucket and/or original
segment for the purpose of estimating a default rate. Institutions should document any
changes in portfolio segmentation that occurred during the period of observation.
3.3.8 When the default rate is computed by rating grade, the ratings at the beginning of the
observation window should not reflect risk transfers or any form of parent guaranties, in
order to capture the default rates pertaining to the original creditworthiness of the
obligors. The ratings at the start of the observation window can reflect rating overrides if
these overrides relate to the obligors themselves, independently of guarantees.
3.3.9 When default rate series are computed over long time periods, it could happen that
obligors come out of their default status after a recovery and a cure period. In subsequent
observation windows, such obligors could be performing again and default again, in
which case another default event should be recorded. For that purpose, institutions should
define minimum cure periods per product and/or portfolio type. If a second default occurs
after the end of the cure period, it should be recorded as an addition default event. These
cure periods should be based on patterns observed in data sets.
3.3.10 Provided that institutions follow the above practices, the following aspects remain subject
to the discretion of each institution. First, they may choose to exclude obligors with only
undrawn facilities from the numerator and denominator to avoid lowering unduly the
default rate of obligors with drawn credit lines. Second, institutions may also choose to
estimate default rates based on exposures rather than on counts of obligors; such
estimation provides additional statistical information on expected exposures at default.
3.4 Rating-to-PD
3.4.1 For the purpose of risk management, the majority of institutions employ a dedicated
methodology to estimate a TTC PD associated with each portfolio and, where possible,
associated with each rating grade (or score) generated by their rating models. This
estimation is based on the historical default rates previously computed, such that the TTC
PD reflects the institution’s experience.
3.4.2 This process results in the construction of a PD scale, whereby the rating grades (or
scores) are mapped to a single PD master scale, often common across several portfolios.
This mapping exercise is referred to as ‘PD calibration’. It relies on assumptions and
methodological choices separate from the rating model, therefore it is recommended to
considered such mapping as a standalone model. This choice is left to each institution and
should be justified. The approach should be tested, documented and validated.
3.4.3 Institutions should demonstrate that the granularity of segmentation employed for PD
modelling is an appropriate reflection of the risk profile of their current loan book. The
segmentation granularity of PD models should be based on the segmentation of rating
models. In other words, the segmentation of rating models should be used as a starting
point, from which segments can be grouped or split further depending on portfolio
features, provided it is fully justified. This modelling choice has material consequences
on the quality of PD models; therefore, it should be documented and approved by the
Model Oversight Committee. Finally, the choice of PD model granularity should be
formally reviewed as part of the validation process.
3.4.4 The rating-to-PD mapping should be understood as a relationship in either direction since
no causal relationship is involved. The volatility of the grade PD through time depends
on the sensitivity of the rating model and on the rating methodology employed. Such
volatility will arise from a combination of migrations across rating grades and changes in
the DR observed for each grade. Two situations can arise:
(i) If rating models are sensitive to economic conditions, ratings will change and the
exposures will migrate across grades, while the DR will remain stable within each
grade. In this case, client ratings will change and the TTC PD assigned to each
rating bucket will remain stable.
(ii) If rating models are not sensitive to economic conditions, then the exposures will
not migrate much through grades but the DR of each grade will change. In this
case, client ratings will remain stable but the observed DR will deviate from the
TTC PD assigned to each rating bucket.
3.4.5 Institutions should estimate the degree to which they are exposed to each of the situations
above. Institutions are encouraged to favour the first situation, i.e. implement rating
models that are sensitive to economic conditions, favour rating migrations and keep the
1-year TTC PD assigned to each rating relatively stable. For the purpose of provisioning,
economic conditions should be reflected in the PIT PD in subsequent modelling.
Deviation from this practice is possible but should be justified and documented.
3.4.6 The estimation of TTC PD relies on a set of principles that have been long established in
the financial industry. At a minimum, institutions should ensure that they cover the
following aspects:
(i) The TTC PD associated with each portfolio or grade should be the long-run
average estimation of the 1-year default rates for each corresponding portfolio or
grade.
(ii) The DR time series should be homogeneous and consistent through time, i.e. relate
to a portfolio with similar characteristics and grading method.
(iii) TTC PDs should incorporate an appropriate margin of conservatism depending on
the time span covered and the population size.
(iv) TTC PDs should be estimated over a minimum of five (5) years and preferably
longer for wholesale portfolios. For retail portfolios, changes in product offerings
should be taken into account when computing TTC PD.
(v) The period employed for this estimation should cover at least one of the recent
economic cycles in the UAE: (i) the aftermath of the 2008 financial crisis, (ii) the
2015-2016 economic slowdown after a sharp drop in oil price, and/or (iii) the
Covid-19 crisis.
(vi) If the estimation period includes too many years of economic expansion or
economic downturn, the TTD PD should be adjusted accordingly.
3.4.7 For low default portfolios, institutions should employ a separate approach to estimate
PDs. They should identify an appropriate methodology suitable to the risk profile of their
portfolio. It is recommended to refer to common methods proposed by practitioners and
academics to address this question. Amongst others, the Pluto & Tasche method or the
Benjamin, Cathcart and Ryan method (BCR) are suitable candidates.
3.4.8 For portfolios that are externally rated by rating agencies, institutions can use the
associated TTC PDs provided by rating agencies. However, institutions should
demonstrate that (i) they do not have sufficient observed DR internally to estimate TTC
PDs, (ii) each TTC PD is based on a blended estimation across the data provided by
several rating agencies, (iii) the external data is regularly updated to include new
publications from rating agencies, and (iv) the decision to use external ratings and PDs is
reconsidered by the Model Oversight Committee on a regular basis.
3.5.1 Modelling choices surrounding PIT PD and PD term structure have material
consequences on the estimation of provisions and subsequent management decisions.
Several methodologies exist with benefits and drawbacks. The choice of methodology is
often the result of a compromise between several dimensions, including but not limited
to: (i) rating granularity, (ii) time step granularity and (iii) obligor segmentation
granularity. It is generally challenging to produce PD term structures with full granularity
in all dimensions. Often, one or two dimensions have to be reduced, i.e. simplified.
3.5.2 Institutions should be aware of this trade-off and should choose the most appropriate
method according to the size and risk profile of their books. The suitability of a
methodology should be reviewed as part of the validation process. The methodology
employed can change with evolving portfolios, risk drivers and modelling techniques.
This modelling choice should be substantiated, documented and approved by the Model
Oversight Committee. Modelling suggestions made by third party consultants should also
be reviewed through a robust governance process.
3.5.3 For the purpose of the MMG, minimum expected practices are articulated for the
following common methods. Other methodologies exist and are employed by
practitioners. Institutions are encouraged to make research and consider several
approaches.
(i) The transition matrix approach,
(ii) The portfolio average approach, and
(iii) The Vasicek credit framework.
3.5.4 Irrespective of the modelling approach, institutions should ensure that the results
produced by models meet business sense and economic intuition. This is particularly true
when using sophisticated modelling techniques. Ultimately, the transformation and the
adjustment of data should lead to forecasted PDs that are coherent with the historical
default rates experienced by the institution. Deviations should be clearly explained.
3.6.1 This section applies to institutions choosing to use transition matrices as a methodology
to model PD term structures.
3.6.2 Transition matrices are convenient tools; however, institutions should be aware of their
theoretical limitations and practical challenges. Their design and estimation should follow
the decision process outlined in the MMS. Institutions should assess the suitability of this
methodology vs. other possible options as part of the model development process. If a
third party consultant suggests using transition matrices as a modelling option,
institutions should ensure that sufficient analysis is performed, documented and
communicated to the Model Oversight Committee prior to choosing such modelling path.
3.6.3 One of the downsides of using transition matrices is the excessive generalization and the
lack of industry granularity. To obtain robust matrices, pools of data are often created
with obligors from various background (industry, geography and size). This reduces the
accuracy of the PD prediction across these dimensions. Consequently, institutions should
analyse and document the implications of this dimensionality reduction.
3.6.4 The construction of the TTC matrix should meet a set of properties, that should be clearly
defined in advance by the institution. The matrix should be based on the institution’s
internal data as it is not recommended to use external data for this purpose. If an
institution does not have sufficient internal data to construct a transition matrix, or if the
matrix does not meet the following properties, then other methodologies should be
considered to develop PD term structures.
3.6.5 At a minimum, the following properties should be analysed, understood and documented:
(i) Matrix robustness: Enough data should be available to ensure a robust estimation
of each rating transition point. Large confidence intervals around each transition
probabilities should be avoided. Consequently, institutions should estimate and
document these confidence intervals as part of the model development phase.
These should be reviewed as part of the model validation phase.
(ii) Matrix size: The size of the transition matrix should be chosen carefully as for the
size of a rating scale. A number of buckets that is too small will reduce the accuracy
of decision making. A number of buckets that is too large will lead to an unstable
matrix and provide a false sense of accuracy. Generally, it is recommended to
reduce the size of the transition matrix compared to the full rating scale of the
institution. In this case, a suitable interpolation method should be created as a
bridge from the reduced matrix size, back to the full rating scale.
(iii) Matrix estimation method: Amongst others, two estimation methods are commonly
employed; the cohort approach and the generator approach. The method of choice
should be tested, documented and reviewed as part of the model validation process.
(iv) Matrix smoothing: Several inconsistencies often occur in transition matrices, for
instance (a) transition probabilities can be zero in some rating buckets, and/or (b)
the transition distributions for a given origination rating can be bi-modal.
Institutions should ensure that the transition matrix respect Markovian properties.
3.6.6 If the institution decides to proceed with the transition matrix appraoch, the modelling
approach should be clearly articulated as a clear sequence of steps to ensure transparency
in the decision process. At a minimum, the following sequence should be present in the
modelling documentation. The MMG does not intend to elaborate on the exact
methodology of each step. Rather, the MMG intends to draw attention to modelling
challenges and set minimum expected practices as follows:
(i) TTC transition matrix: The first step is the estimation of a TTC matrix that meets
the properties detailed in the previous article.
(ii) Credit Index: The second step is the construction a Credit Index (“CI”) reflecting
appropriately the difference between the observed PIT DR and TTC DR (after
probit or logit transformation). The CI should be coherent with the TTC transition
matrix. This means that the volatility of the CI should reflect the volatility of the
transition matrix through time. For that purpose the CI and the TTC transition
matrix should be based on the same data. If not, justification should be provided.
(iii) Forecasted CI: The third step involves forecasting the CI with a macroeconomic
model. However, a segmentation issue often arises. If the matrix was created by
pooling obligors from several segments, then only one blended CI will be
estimated. This may be insufficient to capture the relationship between
macroeconomic variables and the creditworthiness of obligors at segment level for
the purpose of PIT modelling. Institutions should be cognisant of such limitation
and provide solutions to overcome it. An option is to adjust the blended forecasted
CI to create several granular CIs that would reflect the behaviour of each segment.
(iv) Adjusted transition matrix: The fourth step is the development of a mechanism to
adjust the TTC transition matrix with the forecasted CI or the adjusted CIs. Several
forward PIT transition matrices should be obtained at several points in the future.
(v) PD term structure: Finally, a PD term structure should be created based on the
forward PIT transition matrices. Methodologies based on matrix multiplication
techniques should be robust and consistently applied.
3.6.7 As part of the development process, several pre-implementation validation checks should
be performed on the TTC transition matrix in order to verify that the above properties are
met. In addition, for each segment being modelled, the matrix should be constructed such
that two logical properties are met by the PD outputs:
(i) The weighted average TTC PD based on the default column of the TTC transition
matrix should be reasonably close to the long term default rate of the obligors from
the same segment(s) employed to build the matrix.
(ii) The weighted average PIT PD based on the default column of the PIT transition
matrix for the next few time steps, should be coherent with the current default rate
of the obligors from the same segment(s) employed to build the matrix or the
segment(s) employed to derived the adjusted CIs.
3.7.1 This section applies to institutions using a portfolio-level scaling approach to model the
term structure of PIT PD. This approach is simpler to implement than transition matrices
and is suitable for portfolios with smaller data sets. In this method, average PD across
ratings are being modelled, instead of all transition points between ratings. This approach
tends to be preferred for smaller segments. The obligor segmentation granularity is
preserved at the expense of a simplification of the rating granularity.
3.7.2 In order to ensure transparency in the decision process, the modelling approach should
be clearly articulated as a clear sequence of steps. It is not the object of the MMG to
elaborate on the exact methodology of each step. Rather, the MMG intends to draw
attention to modelling challenges and set minimum expected practices. At a minimum,
the following sequence should be present in the modelling documentation:
(i) Forecast portfolio average PIT PD per segment based on macro-PD models.
(ii) Estimate the deviation of the portfolio PIT PD from its long term average PD.
(iii) Apply this deviation to PDs at lower granularity levels, for instance pools or rating
grades. This can be achieved by scaling in logit or probit space.
(iv) Construct a PIT PD term structure. It is highly recommended to compare several
methods and test their impacts on PD outcomes and risk metrics.
3.7.3 The drawback of this method is the generalisation of the PD volatility across grades (or
pools) with the use of scalers. Certain rating grades are more prone to volatility than
others, which is not reflected in this type of model. Therefore this method could result in
an underestimation of the PIT PD of the lowest rating grades. Consequently, institutions
should demonstrate that they assess and understand this effect.
3.7.4 Institutions should ensure that scalers lead to explainable shifts of the PD curve across
rating grades and across time steps. The scalers will not be static. They will change
through the forecasted time steps, since they follow the path of the PD forecasts.
3.7.5 Institutions should be aware of the theoretical and practical limitations of this approach.
Its design and estimation should follow the decision process outlined in the MMS. As for
any other models, institutions should assess the suitability of this methodology vs. other
possible options as part of the model development process.
3.8.1 The Vasicek credit framework is sometimes used to model PIT PD term structures.
Institutions should be cognisant of the material challenges arising from using the Vasicek
one-factor credit model (or similar derivations) for the purpose of ECL estimation, for
the following reasons:
(i) This model has been originally designed to model economic capital and extreme
losses at portfolio level. It is designed to replicate the behaviour of credit risk for
a granular and homogeneous portfolio. Whilst it is an elegant modelling
construction, it might not be the most suitable approach to model expected loss
behaviours at the level of individual obligors.
(ii) It relies on parameters that are challenging to calibrate, in particular the asset
correlation representing the correlation between (a) obligors’ asset value and (b) a
non-observable systemic factor - generally assimilated to an industry factor for
practical reasons. The model results are highly sensitive to the choice of this
parameter. When modelling PIT PD, the introduction of this correlation parameter
tends to reduce the impact of macroeconomic factors.
(iii) When it is used for ECL, the Vasicek model is often combined with a
macroeconomic regression model. In this case, the non-observable systemic factor
is not a given input. Rather, it is partially driven by macro variables. Consequently,
the commonly used one- factor model should be adjusted to account for the
variance of the residuals, i.e. the part of the systemic factor that is not explained by
the macro variables.
3.8.2 If an institution decides to use this methodology, this choice should be approved by the
Model Oversight Committee, with a clearly documented rationale. The asset correlation
parameters should be chosen carefully, bearing in mind the following principle: the lower
the PD, the higher the asset correlation because the obligor’s performance is expected to
be mostly driven by systemic factors rather than by idiosyncratic factors.
3.8.3 It is common practice to calibrate the asset correlation based on the values suggested by
the Basel Framework employed for capital calculation. However, institutions should
consider the possibility that the interconnectedness of corporates in the UAE could lead
to higher systemic correlations for some industries. Consequently, institutions should, at
a minimum, perform sensitivity analysis to assess the implications of this calibration
uncertainty on PDs.
3.9.1 Irrespective of their choice of methods, institutions should validate PD models according
to the principles articulated in the MMS. In particular for PD models, both qualitative and
quantitative assessments are required.
3.9.2 Institutions should ensure that the following metrics represent accurately the risk profile
of their books at segment-level: TTC 1y PD, PIT 1y PD and PD term structure. For that
purpose, these metrics should be validated at several granularity levels (e.g. rating grades,
segments, industries). Statistical tests alone are generally not sufficient to conduct
appropriate validation of PD at segment level. Consequently, institutions should combine
statistical tests, judgement and checks across several metrics to ensure that the calibration
of these metrics are meaningful and accurate.
3.9.4 If insufficient data is available to estimate some of the metrics in the below table,
institutions should demonstrate that actions are taken to collect data to produce these
metrics. Given the lack of data, they should also explain their approach to assess the
suitability of the PIT PD calibration currently used in production.
3.9.5 Upon the estimation of the above metrics, institutions should perform the following
comparisons at segment level. Institutions should implement acceptable limits to monitor
each of the following comparison, i.e. the difference between each two quantities. These
limits should be included in the model validation framework and agreed by the Model
Oversight Committee. Frequent and material breaches should trigger actions as
articulated in the governance section of the MMS.
(i) TTC 1y DR vs. TTC 1y PD per segment: The objective is to verify that the central
tendency observed historically is in line with the PD estimated based on the
institution’s master-scale.
(ii) PIT 1y DR vs. PIT 1y PD estimated over the same historical period, per segment:
This is a back testing exercise. The objective is to verify that the default rates
observed historically are falling within a reasonable confidence interval around the
PD forecasted over the same period.
(iii) PIT 1y DR recently computed vs. PIT 1y PD estimated over the next 12 months:
The objective is to verify that the default rate recently observed is coherent with
the PD forecasted from the reporting date over the next 12 months. These two
quantities can diverge due to the effect of economic forecasts. There should be a
logical and intuitive link between the two and material differences should be
explained.
(iv) TTC CDR vs. PIT CPD per segment: The objective is to verify that the shape of
the cumulative default rates observed historically is similar with the shape of the
cumulative default rate forecasted by the model from the portfolio reporting date.
The shape can be expressed as a multiplier of the 1-year PD.
(v) TTC CDR vs. PIT PD derived analytically: A PD term structures can be
estimated simply by using survival probabilities derived from the institution’s PD
mater scale. This is referred as the analytical PD term structure, that serves as a
theoretical benchmark. The objective is to compare this analytical benchmark vs.
(a) observed CDR and (b) the PD term structure generated by the model. Material
deviations should be understood and documented. If CDR and CPD are materially
lower than the analytical approach, adjustments should be considered.
3.9.6 In addition to segment level validation, institutions should ensure that the PIT PD profile
across rating grades is logical and consistent. This is particularly relevant in the case of
transition matrices. PIT adjustments should be coherent across different ratings.
Technically, for a given segment and a given horizon forecast, the log-odd ratio of the
PIT PD for a given rating over the TTC PD for the same rating, should be of similar
magnitude between ratings.
3.9.7 Finally, economic consistency between segments is also part of the validation process of
PD models. Institutions should ensure that such considerations are included in the scope
of model validation. PIT PDs generated by models should be coherent between industries
and between segments. For instance, if a given portfolio displayed high historical PD
volatility, then such volatility is expected to be reflected in the forecasted PIT PD.
Material deviations from coherent expectations should be explained and documented.
4 LGD MODELS
4.1 Scope
4.1.1 For risk management purpose and to comply with accounting and regulatory
requirements, the majority of institutions develop models to estimate the potential loss
arising in the event of the default of a facility or obligor. These are referred to as Loss
Given Default models (“LGD”). These models serve several purposes including, but not
limited to provision estimation, credit portfolio management, the economic capital
estimation and capital allocation. For the purpose of the MMG, and to ensure appropriate
model management, the following components are considered as separate models:
(i) TTC LGD models, and
(ii) PIT LGD models.
The definitions of through-the-cycle (“TTC”) and point-in-time (“PIT”) are similar to
those used under the section on PD models.
4.1.2 Institutions should develop and manage these models through a complete life-cycle
process in line with the requirements articulated in the MMS. In particular, the
development, ownership and validation process should be clearly managed and
documented.
4.1.3 Institutions are expected to meet minimum expected practices for the development of
LGD models. For that purpose, the construction of LGD models should include, at a
minimum, the following steps:
(i) Regular and comprehensive collection of data,
(ii) Accurate computation of realised historical LGD,
(iii) Analysis of the LGD drivers and identification of the most relevant segmentation,
(iv) Development of TTC LGD model(s), and
(v) Development of PIT LGD model(s).
4.1.4 This section elaborates on the concepts and methods that institutions should incorporate
in their modelling practice. In particular, institutions should pay attention to the
appropriate estimation of recovery and losses arising from restructured facilities.
Restructuring should not always be considered as a zero-sum game leading to no financial
impact. In some cases the present value (“PV”) mechanics can lead to limited impact;
however, restructuring events generate execution costs, delays and uncertainty that
should be fully identified and incorporated into LGD modelling.
4.1.5 Institutions are strongly recommended to apply floors on TTC LGD and PIT LGD across
all portfolios for several reasons. There is limited evidence that default events lead to zero
realised losses. An LGD of zero leads to zero expected loss and zero capital consumption,
thereby creating a biased perception of risk and misguided portfolio allocation. LGD
floors contribute to sound decision making for risk management and lead to more realistic
provisioning. The value of the floor should be five percent (5%) for all facilities, except
in the following circumstances:
(i) The facility is fully secured by cash collateral, bank guarantees or government
guarantees, and/or
(ii) The institution provides robust evidence that historical LGDs are lower than 5%.
In all circumstances, LGD should not be lower than one percent (1%).
4.2.1 A robust data collection process should be put in place in order to support the estimation
of LGD parameters. This process should follow the requirements pertaining to the Data
Management Framework articulated in the MMS. If the data listed below is not currently
available in the institution’s data bases, a formal project should be put in place in order
to collect and store this data from the date of the MMG issuance.
4.2.2 Governance: The data types to be collected for the estimation of LGD are often spread
across several teams and departments within institution. Consequently, close
collaboration is expected between departments to (i) fully retrieve historical default cases,
(ii) understand the reasons and the context for default and (iii) reconstruct long recovery
processes that can last several years. In particular, close collaboration is expected between
the risk analytics department in charge of model development and the department in
charge of managing non-performing credit facilities.
4.2.3 Default definition: Institutions should ensure consistency between (i) the default
definition used to collect data for LGD estimation and (ii) the default definition used to
estimate PD for the same segment. PD and LGD are necessarily linked to each other and
their estimation should be coherent. This is particularly relevant in the context of facility
restructuring. If restructured facilities are included in the estimation of LGD, they should
also be included in the estimation of PD.
4.2.4 Data types: The collection of data should cover all the elements necessary to estimate
recoveries and historical LGDs, following each default event. At a minimum, the data
collection process should include the following elements:
(i) Default events: An exhaustive list of default events should be collected to support
a robust estimation of LGD. They should be consistent with the default events
employed for PD modelling. Institutions are expected to collect as many default
events as possible covering at least one economic cycle.
(ii) Exposure At Default: As per the definition section. For non-contingent facilities,
the EAD should be the outstanding amount at the date of default. For contingent
facilities, the EAD should be the outstanding amount at the date of default plus any
other amounts that become due during the recovery process. This should include
any additional drawings that occurred after default and before foreclosure or cure.
It should also include any part of the original exposure that had been written-off
before the default event was recorded.
(iii) Outcome categorisation: Each default event should be linked to the categories
presented in the next article, depending on the outcome of the recovery process,
namely (i) cured & not restructured, (ii) cured & restructured, (iii) not cured &
secured, and (iv) not cured & not secured.
(iv) Obligor information: For each default event, client related information should be
collected including, at a minimum, client segment, industry and geography.
(v) Facility details: This should include the type of facility and the key elements of the
facility terms such as the tenor, the seniority ranking and the effective interest rate.
(vi) Restructuring: Each restructuring and rescheduling event should be identified and
used in the categorisation of outcomes presented in the next articles.
(vii) Collateral: For each default event related to collateralised facilities, institutions
should collect all relevant collateral information, including, but not limited to, the
type of collateral, the last valuation prior to the default event and the associated
valuation date, the liquidation value after default and the associated liquidation
date, the currencies of collateral values and unfunded credit protections. If several
valuations are available, institutions have a method to estimate a single value.
(viii) Historical asset prices: In order to estimate collateral haircut, historical time series
should be collected, including amongst others, real estate prices, traded securities
and commodity prices.
(ix) Collected cash flows: For each default event, the data set should include the cash
flow profile received through time, related to this specific default event. Provided
that collected cash inflows are related to the specific default event, they can arise
from any party, including the obligor itself, any guarantor or government entities.
(x) Direct costs: These costs are directly linked to the collection of the recovery cash
flows. They should include outsourced collection costs, legal fees and any other
fees charged by third parties. If the facility is secured, the data set should include
costs associated with the sale of collateral, including auction proceedings and any
other fees charges by third party during the collateral recovery process.
(xi) Indirect costs: Institutions are encourage to collect internal and external costs that
cannot be directly attributed to the recovery process of a specific facility. Internal
costs relate to the institution’s recovery process, i.e. mostly the team that manages
non-performing credit facilities and obligors. External costs relate mostly to
overall costs of outsourced collection services.
4.2.5 Categorisation: The outcome of default events should be clearly categorised. Institutions
are free to define these categories, provided that these include, at a minimum, the below
concepts.
Table 3: Typical outcomes of default events
Outcome Description Detailed outcome Category
Cured The obligor has returned to a No restructuring / Cured & not
performing situation after a cure rescheduling restructured
period – as defined in the
Restructuring / Cured &
CBUAE Credit Risk regulation.
rescheduling restructured
Not cured The obligor has not returned to The facility is Not cured &
a performing status after secured by collateral secured
defaulting.
The facility is not Not cured &
secured unsecured
Unresolved The outcome remains uncertain Unresolved
until a maximum recovery
period beyond which all cases
should be considered closed for
LGD estimation.
4.3.1 The next step in the modelling process involves the computation of historical realised
LGD based on the data previously collected. The objective is to estimate the recovery and
loss through a ‘workout’ approach for each of the identified default event.
4.3.2 The computation of LGD relies on the appropriate identification and quantification of the
total recoveries and expenses linked to each default event. Institutions should implement
a robust process to perform the necessary computation to estimate LGD at the lowest
possible granularity level.
4.3.3 Institutions can develop their own methodologies for the estimation of historical realised
LGD. However, their approach should incorporate, at a minimum, the components listed
in this section and the corresponding categories of workout outcomes.
4.3.4 Institutions are expected to compute LGD at the time default (𝑡) as the ratio of the total
loss incurred divided by the Exposure At Default. When modelling LGD time series, the
time of reference should be the date of default. We note LGD as a function of time t, as
𝐿𝐺𝐷𝑡 , then t is the date of default, which is different from the time at which the recovery
cash flows where collected. The total recovery is noted 𝑅𝑒𝑐𝑜𝑣𝑒𝑟𝑦𝑡 and the total loss is
noted 𝐿𝑜𝑠𝑠𝑡 . Institutions should therefore estimate realised LGD for each default event
with the following formula:
𝐸𝐴𝐷𝑡 − 𝑅𝑒𝑐𝑜𝑣𝑒𝑟𝑦𝑡 𝐿𝑜𝑠𝑠𝑡
𝑅𝑒𝑎𝑙𝑖𝑠𝑒𝑑 𝐿𝐺𝐷 𝑎𝑡 𝑡𝑖𝑚𝑒 𝑡 = =
𝐸𝐴𝐷𝑡 𝐸𝐴𝐷𝑡
4.3.5 The recovery is derived from all cash inflows and expenses occurring at future times after
the default event and discounted back to the default date. The recovery cash flows should
not be higher than the amount of recoveries that can legally be obtained by the institution.
The discount rates should reflect the time-value of money plus the uncertainty
surrounding the cash flows. Additional considerations for restructured facilities are
presented at the end of this section. If several facilities are secured by one or several
collaterals, institutions should implement a clear collateral allocation mechanism from
the obligor level to each facility. The computations of 𝑅𝑒𝑐𝑜𝑣𝑒𝑟𝑦𝑡 and 𝐿𝑜𝑠𝑠𝑡 depend on
the workout outcome. The estimation method should incorporate the following
components.
Table 4: Recovery and costs per type of outcome
Outcome Components of recovery and costs
(1) Cured & not (a) Indirect costs, as defined in the data collection section.
restructured
(2) Cured & (a) Potential reduction in present value upon restructuring the
restructured facility.
(b) Direct costs, as defined in the data collection section.
(c) Indirect costs, as defined in the data collection section.
(3) Not cured & (a) When applicable, present value of discounted collateral
secured proceedings back to the default date.
(b) Direct costs, as defined in the data collection section.
(c) Indirect costs, as defined in the data collection section.
(d) Cash flows received but not associated with collateral
liquidation.
(4) Not cured & (a) Recovered cash flows. Effect of discounting these cash
unsecured flows back to the default date, function of the time to
recovery.
(b) Indirect costs, as defined in the data collection section.
(5) Unresolved These should be treated as per the following article.
4.3.6 The treatment of unresolved default cases (incomplete workouts) creates a material bias
in the estimation of LGD. Consequently, institutions should establish a clear process for
the treatment of these cases and understand their impact on the estimation of historical
realised LGD.
(i) Institutions should define criteria to decide on whether the recovery process of a
default case should be considered closed. A degree of conservativeness should be
included in this estimation to reflect the uncertainty of the recovery process. This
means that if doubts persist regarding the possibility of future cash inflows, the
recovery process should be considered closed.
(ii) Institutions should put in place clear criteria to include or exclude unresolved cases
in their estimation samples. For that purpose, a maximum length of resolution
period (from the date of default) should be established by obligor segment. The
objective is to choose a duration that is sufficiently short to maximise the number
of recovery cases flagged as ‘closed’ and sufficiently long to capture a fair recovery
period.
(iii) It is recommended that open default cases with a recovery period longer than four
(4) years should be included in the LGD estimation process, irrespective of
whether they are considered closed. For the avoidance of doubt, all closed cases
with a shorter recovery period should, of course, be included. Banks are free to use
a shorter maximum duration. Longer maximum duration, however, should be
avoided and can only be used upon robust evidence provided by the institution.
(iv) Default cases that are still unresolved within the maximum length of the recovery
process (i.e. shorter than 4 years) should preferably be excluded for the purpose of
estimating historical LGDs. Institutions have the option to consider adjustments by
extrapolating the remaining completion of the workout process up to the maximum
resolution period. Such extrapolation should be based on documented analysis of
the recovery pattern by obligor segment and/or product type observed for closed
cases. This extrapolation should be conservative and incorporate the possibility of
lower recovered cash-flows.
4.3.7 Institutions should not assume that restructuring and rescheduling events necessarily lead
to zero economic loss. For restructuring associated with material exposures, an estimation
of their associated present value impact should be performed. If no PV impact is readily
available, then the terms of the new and old facilities should be collected in order to
estimate a PV impact, according to the methodology outlined in the dedicated section of
the MMG. In particular, if the PV impact of the cash flow adjustment is compensated for
by a capitalisation of interest, institutions should include an incremental credit spread in
discounting to reflect the uncertainty arising from postponing principal repayments at
future dates. Such credit spread should then lead to a negative PV impact.
4.3.8 For low default portfolios, institutions may not have enough data to estimate robust
historical recovery patterns. In this case, institutions should be in a position to
demonstrate that data collection efforts have taken place. They should also justify
subsequent modelling choices based on alternative data sources and/or comparables.
Furthermore, portfolios with high frequency of cure via restructuring should not be
considered as portfolios with low default frequency. Restructured facilities could be
recognised as defaults depending on circumstances and in compliance with the CBUAE
credit risk regulation.
4.4.1 Once institutions have estimated and categorised realised LGD, they should analyse and
understand the drivers of realised LGD in order to inform modelling choices in the
subsequent step.
4.4.2 At a minimum, institutions should analyse and understand the impact of the following
drivers on LGD:
(i) The time at which LGD was observed in the economic cycle. The profile of the
recovery pattern and the effect of the economic cycle on this pattern.
(ii) The effect of collateral on the final LGD including the time to realise the collateral,
the impact of the type of collateral, the difference between the last valuation and
the liquidation value.
(iii) The link between LGD and the obligor’s credit worthiness at the time of default
captured by its rating or its PD.
(iv) The type of facility and its seniority ranking, where applicable.
(v) The obligor segments expressed by size, industry, and/or geography.
(vi) Any change in the bankruptcy legal framework of the jurisdiction of exposure.
4.4.3 Institutions should identify the most appropriate segmentation of historical realised LGD,
because this choice will subsequently inform model segmentation. Portfolio segmentation
should be based upon the characteristics of the obligors, its facilities and its collateral
types, if any.
4.4.4 Institutions should be cautious when using ‘Secured LGD’ and ‘Unsecured LGDs’ as
portfolio segments. A secured LGD is a loss obtained from a facility secured by a
collateral. It is based upon the estimation of a collateral coverage (defined as the ratio of
the collateral value to the exposure). The loss resulting from such coverage can spread
across a large range: from low values in the case of over-collateralization, up to high
values in the case of small collateral amounts. An average (referred as Secured LGD)
across such large range of values is likely to suffer from a lack of accuracy. Thus, it is
preferable to employ collateral as a direct continuous driver of LGD, rather than use it to
split a population of obligors.
4.4.5 Once segments have been identified, institutions should produce three types of LGD per
segment to support the estimation of ECL as per accounting principles. These estimates
should be used to calibrate the TTC LGD and PIT LGD models in subsequent modelling
steps. The estimation of averages can be exposure-weighted or count-weighted. This
choice depends on each portfolio and thus each institution.
(i) The long run average by segment, through time across business cycles, estimated
as the average of realised LGDs over the observation period.
4.4.6 When analysing the effect of collateral on LGD outcomes, institutions should consider,
at a minimum, the following collateral types. Note that personal guarantees should not be
considered as eligible collateral for the purpose of modelling LGD. This list may evolve
with the CBUAE regulation.
4.5.1 The objective of TTC LGD models is to estimate LGD, independently of the
macroeconomic circumstances at the time of default. Therefore, these models should not
depend on macroeconomic variables. These models can take several forms depending on
the data available and the type of portfolio. Institutions are free to choose the most suitable
approach, provided that it meets the minimum expectations articulated in this section.
4.5.2 Defaulted vs. non-defaulted cases: LGD should be modelled and estimated separately
between defaulted obligors (or facilities) and non-defaulted obligors. Whilst the
methodology should be similar between these two cases, several differences exist:
(i) Upon a default event, the estimation of recovery relies on assumptions and on a
live process with regular information updates. Therefore, for defaulted obligors (or
facilities), as the recovery process unfolds, institutions should collect additional
information to estimate recovery rates with further accuracy and thus obtain more
specific LGD estimation.
(ii) For defaulted obligors (or facilities), particular attention should be given to PV
modelling as per the dedicated section of the MMG. Discount factors should reflect
the circumstances of default and the uncertainty surrounding the recovery process.
(iii) One of the major differences between LGD from defaulted vs. non-default
exposures is that the former is estimated only as of the date of default, while the
latter is estimated at several point in time, depending on the needs of risk
management and financial reporting.
4.5.3 Properties: At a minimum, LGD models should meet the following properties.
(i) The modelled LGD should be based upon the historical realised LGD observations
previously estimated.
(ii) The methodology should avoid excessive and unreasonable generalisations to
compensate for a lack of data.
(iii) The model performance should be validated based on clear performance
measurement criteria. For instance, model predictions should be compared against
individual observations and also against segment average.
(iv) There should be enough evidence to demonstrate that in-sample fit and out-of-
sample performance are reasonable.
(v) The choice of parameters should be justified and documented.
(vi) The model inputs should be granular and specify enough to generate a LGD
distribution that is a fair and accurate reflection of the observed LGDs.
4.5.4 Functional form: Institutions are free to use LGD models with any functional form
provided that the model output is an accurate reflection of the observed LGD. Institutions
should aim to build LGD models that incorporate the suitable drivers enabling the model
to reflect the main possible outcomes of the workout process.
4.5.5 Parameters: Institutions should aim to incorporate the following drivers in their LGD
models. This means that any model using less granular inputs should be considered as a
first generation model that requires improvement as further data becomes available.
(i) The probability of cure without restructuring,
(ii) The probability of cure through restructuring,
(iii) The collateral coverage,
(iv) Direct and indirect recovery costs,
(v) Collateral liquidation values including haircuts, and
(vi) Recovered cash flows
4.5.6 The parameters listed above should drive the estimation of LGD. The mathematical
construction of the LGD model can take several forms, that institutions are free to choose.
The form presented below serves as illustration. Institutions are not expected to use this
expression literally; rather, they should ensure that their final LGD model respects the
principles of this expression with a suitable estimation of each component. If institutions
employ a different functional form, they are encouraged to use the following expression
as a challenger model.
The loss arising from the unsecured portion (𝐿4 ) is often called “unsecured LGD”. The
final LGD after taking collateral into accounts is often referred to as the “secured LGD”.
Irrespective of the semantic employed, LGD models should reconcile conceptually
against the expression above.
4.5.7 Granularity: Institutions should aim to develop TTC LGD models to estimate LGD at a
low level of granularity. The following minimum expected practices apply:
(i) Institutions should aim to model LGD at facility level, i.e. the LGD should
incorporate facility characteristics. If this is not possible for practical reasons, LGD
should be computed at obligor level. This means that LGD should be driven by
parameters specific to each obligor and the associated collaterals if any.
(ii) If several facilities are secured by one or several collaterals, institutions should
implement a clear collateral allocation mechanism from the obligor to each facility.
(iii) If institutions do not have the required data to build such granular models, they
should put in place a formal project in order to collect the necessary data as a
stepping stone towards accurate LGD modelling.
4.5.9 Collateral haircuts: The last valuation of an asset is unlikely to reflect the resale value of
a repossessed asset. Consequently, institutions should estimate appropriate haircuts based
on the assumption that they intend to sell the repossessed asset as soon as reasonably
possible. Haircuts should be estimated based on historical data by type of collateral.
4.5.10 Bimodal distribution: Institutions should identify whether the distribution of observed
LGD is bimodal, i.e. a distribution with two peaks of high frequency. In this case, specific
modelling constraints apply. Institutions should be cautious when using an average value
between these two peaks. Such average can be misleading and should not be employed
to assign LGD at facility level since it does not correspond to an observable LGD at
facility level.
4.5.11 Logical features: Following on from the conceptual framework presented above, some
logical characteristics should be respected: (i) the final LGD should be equal or smaller
than the unsecured LGD, (ii) the LGD should decrease with the presence of collateral, all
other parameters being kept constant, and (iii) the longer the recovery period, the higher
the recovery, the lower the LGD. The logical features should be tested as part of the model
validation process.
4.6.1 There is general evidence that LGD levels tend to be higher during economic downturns.
This intuitive relationship is supported by numerous publications from academics and
practitioners based on data from the US and European markets. In the UAE, whilst this
relationship is more difficult to prove, there are objective reasons to believe it exists. In
any case, this should be investigated as part of current accounting requirements.
Consequently, institutions should implement a process to analyse the relationship
between LGD and macro factors. This should be done at a relevant level of granularity.
4.6.2 This analysis should drive the modelling strategy of PIT LGD. Several modelling options
can be envisaged and institutions should articulate explicitly their approach based on their
preliminary analysis. While making a strategic decision, institutions should remain
conservative. A portfolio may not be large enough to capture this relationship despite the
existence of such relationship at larger scale. In doubt, it is preferable to include some
degree of correlation between LGD and macro factors for the estimation of ECL. Once a
mechanism is in place, the strength of the relationship can be adjusted in calibration
exercises, upon further evidence proving or refuting it.
4.6.3 The objective of PIT LGD models is to estimate LGD as a function of the economic
circumstances at the time of default and during the recovery process. Therefore, these
models should depend on macroeconomic variables. Institutions are free to choose the
most suitable methodology, provided that it meets the minimum expected practices
articulated in this section.
4.6.4 PIT LGD models can be constructed by (i) adjusting TTC LGD or (ii) developing models
independently from TTC LGD. For consistency purpose, the former is recommended over
the latter. If institutions chose the second approach, they should ensure that both PIT LGD
output and TTC LGD outputs are coherent.
4.6.5 The properties of the PIT LGD models should be similar to that of TTC LGD models. At
a minimum, these models should meet the following:.
(i) The modelled LGD should be based upon the historical realised LGD observations
previously estimated.
(ii) The methodology should avoid excessive and unreasonable generalisations to
compensate for a lack of data.
4.6.6 PIT LGD models can take several forms depending on the data available and the type of
portfolio. Several broad categories of models can be defined as follows, ranked by
increasing granularity and accuracy:
(i) Option 1: Most granular approach. The LGD parameters are directly linked to the
macro forecasts and used as inputs to compute the losses (𝐿1 , 𝐿2 , 𝐿3 , 𝐿4 ). The final
LGD is subsequently computed based on these losses, as defined in the TTC LGD
section. For instance, collateral values at facility level can be directly linked to the
macro forecasts, then secured LGDs are derived.
(ii) Option 2: Intermediate granular approach. The losses (𝐿1 , 𝐿2 , 𝐿3 , 𝐿4) are linked to
the macro forecasts and used as input to estimate the final LGD, as defined in the
TTC LGD section. For instance, the segment level secured and unsecured LGDs
can be linked to the macro forecasts.
(iii) Option 3: Non-granular approach. The final LGD is directly linked to the macro
forecasts. In this case the PIT LGD models does not use the LGD parameters.
(iv) Option 4: Alternative approach. The final LGD is linked to the obligor’s PD, itself
linked to macro forecasts. In this case, the LGD response to macroeconomic
shocks is constructed as a second order effect through correlation rather than
structural causation.
4.6.7 Institutions should articulate and document explicitly their preferred modelling option.
All these options are acceptable; however institutions should be aware of their theoretical
and practical limitations, in particular the potential accuracy issues arising from options
3 and 4. Institutions should aim to model PIT LGD via option 1. Consequently,
institutions should understand and assess the financial implications of their modelling
choice. This choice should be approved by the Model Oversight Committee.
4.6.8 If the PIT LGD model uses PIT PD as a sole driver of macro adjustment, then the model
segmentation should be identical between PIT LGD and PIT PD. If institutions decide to
develop dedicated PIT LGD-macro models, those should follow the minimum
expectations articulated in the section of the MMG dedicated to macro models.
4.7.1 Institutions should validate all LGD models according to the validation principles
articulated in the MMS. Both qualitative and quantitative assessments should be
conducted for an appropriate validation.
4.7.2 Institutions should ensure that segment-level LGD values represent the risk profile of
their books. Statistical tests alone are not sufficient to conduct appropriate validation of
LGD at segment level. Consequently, institutions should combine statistical tests,
judgement and checks across comparable metrics to ensure that the calibration of these
metrics are meaningful and accurate.
4.7.3 The scope of the validation should be comprehensive. If the validation is performed by a
third party consultant, institutions should ensure that the validation effort is
comprehensive in scope and substance.
4.7.4 The validation scope should include, at a minimum, the following components:
(i) The data quality, comprehensiveness and collection process. This should cover the
analysis of unusual features observed in historical data and their subsequent
treatment for modelling.
(ii) The definition of default. This should cover the treatment of technical defaults and
restructured accounts.
(iii) The methodology employed to compute historical LGD. This should cover in
particular the possible outcomes as described earlier in this section. Partial
replication of the historical LGD for a sample of facilities should be performed.
(iv) The methodology employed to estimate TTC LGD and subsequent PIT
adjustments. This should cover the model fit, functional form, robustness,
properties and sensitivities.
(v) The treatment of collateral. The treatment of LGD segmentation granularity. The
quality of the model output in terms of economic and business meaning. This can
rely on comparables based on data available outside of the institution.
(vi) The existence of spurious accuracy and excessive generalisation. In particular, the
validation process should report the excessive usage of average LGD employed
across a large population of heterogeneous obligors.
(vii) Back-testing of modelled LGD, estimated separately for defaulted and non-
defaulted obligors.
5 MACROECONOMIC MODELS
5.1 Scope
5.1.1 Macroeconomic models (“macro models”) are primarily employed by UAE institutions
for the estimation of Expected Credit Loss (“ECL”) and for internal and regulatory stress
testing purpose. The objective of this section is to provide guidance and set the Central
Bank’s expectation applicable to all macroeconomic models used by institutions. The
practices described in this section are in compliance with current accounting principles.
5.1.2 In this context, macro models are defined as statistical constructions linking macro
variables (“independent variables”) to an observed risk or business metrics, typically
PD, Credit Index, LGD, cost of funds, or revenues (“dependent variables”). Several
types of macro models exist. A common approach relies on time series regression
techniques, which is the main focus of this section on macro models. Other approaches
include (i) for retail clients, vintage-level logistic regression models using directly
macroeconomic drivers as inputs and (ii) for corporate clients, structural models using
macro variables as inputs.
5.1.3 Irrespective of the methodology employed, institutions should use judgement and
critical thinking, where statistical techniques are coupled with causality analysis.
Institutions should justify and balance (i) statistical performance, (ii) business intuition,
(iii) economic meaning, and (iv) implementation constraints. Statistical methods and
procedures will only provide part of the solution. Therefore, rigorous modelling
techniques should be coupled with sound economic and business judgement in order to
build and choose the most appropriate models. The key modelling choices and the thought
process for model selection should be rigorously documented and presented to the Model
Oversight Committee.
5.1.4 The modelling decision process should be driven by explorations, investigations, and
comparisons between several possible methods. Note that time series regression models
have been proven to yield the most intuitive results over other techniques.
5.1.5 When developing macro models, institutions should follow a clear sequential approach
with a waterfall of steps. Depending on the outcome, some steps may need to be repeated.
Each step should be documented and subsequently independently validated. In particular,
for time series regression models, the process should include, at a minimum, the steps
presented in the table below.
Table 8: Sequential steps for the development of macro models
# Step
1 Data collection
2 Analysis of the dependent variables
3 Analysis of the macro variables
4 Variable transformations
5 Correlation analysis
6 Model construction
7 Statistical tests
8 Model selection
9 Monitoring and validation
10 Scenario forecasting
5.2.1 In order to proceed with macroeconomic modelling, institutions should collect several
types of time series. This data collection process should follow the requirements
articulated in the MMS.
5.2.2 At a minimum, these time series should be built with monthly or quarterly time steps over
an overall period of five (5) years covering, at least, one economic cycle. Institutions
should aim to build longer data series. The following data should be collected.
5.2.3 Macro variables: Institutions should obtain macro variables from one or several external
reliable sources.
(i) The scope of variables should be broad and capture appropriately the evolution of
the economic environment. They will typically include national accounts (overall
and non-oil, nominal and real), oil production, real estate sector variables, CPI,
crude oil price and stock price indexes.
(ii) Institutions should collect macro data pertaining to all jurisdictions where they
have material exposures (at least greater than ten percent (10%) of the total
lending book, excluding governments and financial institutions).
(iii) Institutions should document the nature of the collected variables, covering at a
minimum, for each variable, a clear definition, its unit, currency, source,
frequency, and extraction date.
(iv) Institutions should ensure that all variables employed for modelling will also be
available for forecasting.
5.2.4 Historical default rates: Macro-PD models (or macro-to-credit index models) stand at the
end of a chain of models. They are employed to make adjustments to the output of TTC
PD models, themselves linked to rating models. Therefore the default data used for
macro-PD modelling should reflect the institution’s own experience. If external default
data points are used, justification should be provided. Finally, institutions are encouraged
to also include restructuring and/or rescheduling events in their data to better capture the
relationship between obligor creditworthiness and the economic environment.
5.2.5 Historical recovery rates: Macro-LGD models are designed to adjust the output of TTC
LGD models. Consequently, the recovery data employed for macro-LGD modelling
should reflect the institution’s own experience. If external recovery data points are used,
justification should be provided.
5.2.6 Macro data series are mostly available with quarterly or even annual data points and rarely
with monthly time intervals. Consequently, interpolation techniques may need to be
developed. Institution should include interpolation methodology as part of the data
transformation step. Such interpolation should be documented and included in the
validation process.
5.3.1 Institutions should demonstrate that default series are suitable for modelling and are
representative of their current portfolio. For that purpose, they should employ judgement
and critical thinking when analysing the data. At a minimum, they should perform an
analysis of the dependent variables through descriptive statistics, covering the distribution
followed by each dependent variable and the identification of outliers, if any. Upon this
analysis, a clear statement should be made regarding the suitability of the data for macro
modelling. Consideration should be given to (i) the data quality, (ii) length, and (iii)
representativeness. This analysis should be fully described in the model development
documentation.
Business consistency: Institutions should pay attention to the business significance of the
historical data related to the dependent variable. One possible conclusion is that historical
data of a given variable is no longer an appropriate representation of the current
institution’s portfolio because the segment business strategy has changed materially. In
the case of default and recovery rates, conservatism prevails.
(i) The institution may believe that its current portfolio is less risky than its historical
portfolio and that it expects to experience lower default rates and/or losses in the
future. In that case, the existing historical default series should be used for a
reasonable period until there is enough evidence supporting the new risk profile.
Subsequently, adjustments are permitted on the forecasted values, for instance in
the form of scalers.
(ii) The institutions may believe that its current portfolio is more risky than its
historical portfolios and that it will consequently experience higher default rates in
the future. In that case, forecasts should be immediately adjusted, i.e. forecasted
PDs and LGDs should be shifted upward.
5.3.2 Regime shifts: Institutions should identify the presence of regime shifts in all times series.
These can be clearly identified by the presence of sudden permanent jumps in the data.
Regime shifts tend to occur in default and recovery series due to changes in the data
collection process, definition of default, recovery process or business strategies. For
modelling, it is strongly recommended to avoid using time series with regime shifts as
direct model inputs. Instead, adjustments should be implemented such as a truncation of
the series or the use of specific econometrics techniques (the introduction of a dummy
variable in the model).
5.3.3 Segmentation consistency: Segmentation means splitting a statistical sample into several
groups in order to improve the accuracy of modelling. This concept applies to any
population of products or customers. In particular, for the construction of PD and LGD
macro models, the choice of portfolio, customer and/or product segmentation has a
material impact of the quality of macro models. The economic behaviours of obligors
and/or products should be homogeneous within each segment in order to build appropriate
models. As mentioned in the data collection section, a degree of consistency is required
between macro models and other models. For macro-PD models in particular, such
consistency should be analysed and documented as follows:
(i) The granularity of segments for macro modelling should be equal or greater than
the granularity of segments employed for (i) rating models, and (ii) PD term
structures models. If this alignment is not possible, institutions should provide
robust justifications and document them accordingly.
(ii) Institutions may decide to increase the segmentation granularity of macro models.
An increase in the number of segments will lead to a reduction in segment size and
in the number of observed defaults, could, in turn, reduce the statistical significance
of the default rate. Therefore, increasing the segmentation granularity is permitted,
provided that there is no material loss in the representativeness of the default rates.
5.3.4 Institutions should analyse and assess the impact of segmentation choices as part of the
development of macro models. Several segmentation options should be considered and
subject to the entire model development process described hereby. Institutions should
then choose the best segmentation by assessing the quality and robustness of the macro
models across several segmentation options.
5.4.1 Institutions should perform a robust analysis of the macro variables through descriptive
statistics. At minimum, this analysis should examine, amongst others, the shape of
distribution to identify outliers, shape of tails, multimodality. Upon this analysis, a clear
statement should be made regarding the suitability of the data for macro modelling. In
particular, the analysis should consider the data quality, length and representativeness.
This analysis should be fully described in the model development documentation.
5.4.2 Regime shift: Institutions should identify the presence of regime shifts in all macro time
series. Regime shifts can occur in macro time series due to economic decisions such as
the introduction of VAT or a large shift in interest rates. Similarly to the dependent
variables, macro time series with regime shifts should be avoided or adjusted accordingly.
5.4.3 Economic consistency: Institutions should pay attention to the economic significance of
the macro variables. Some macro variables provide consistently better explanatory power
of risk metrics in the banking book. Conversely some variables are more challenging to
interpret, consequently institutions should be cautious when using those variables for PD
and LGD macro models. Particular attention is needed for the following:
(i) Employment rate: A large proportion of employees leave the UAE upon losing their
employment. Consequently, the UAE employment rate incorporates a material bias,
hence it is preferable to avoid this variable to model business or risk metrics.
(ii) Interest rates: The relationship between interest rates and default rates is ambiguous.
Institutions should ensure that an appropriate interpretation of the estimates is
provided upon modelling PDs and LGDs.
(iii) Abu Dhabi and Dubai stock indices: These indices can suffer from a lack of liquidity
therefore institutions should ensure that an appropriate interpretation of the estimates
is provided upon modelling PDs and LGDs.
(iv) UAE account balances (e.g. fiscal balance, current account): By construction these
variables can oscillate between positive and negative values. Consequently, a
relative time differencing can lead to very high returns and uncontrollable spikes.
Instead, it is recommended to normalise these variables by nominal GDP prior to
using them for modelling.
5.5.1 The principles of variable transformation articulated in the MMS also apply to macro
models. Variable transformations have a material impact on macro models and on ECL.
Therefore, institutions should test, choose and document the most appropriate
transformations applied to both the macro variables and to the dependent variable.
5.5.2 Stationarity: Modelled economic relationship should be stable over time. In the context
of time series regression model, variables should be stationary in order to construct robust
and meaningful econometric models. Stochastic trends, seasonality and structural breaks
are most common sources of non-stationarity. This property should be tested for both the
dependent and independent variables, according to the following principles:
(i) Macroeconomic variables should meet stationarity criteria prior to be used for
modelling. The absence of stationarity has adverse material consequences on
macro models because it often leads to spurious correlations. Macro variables that
are not stationary should either be transformed to obtain stationary series or should
be excluded from the modelling process.
(ii) Even after transformations, in some circumstances full stationarity is challenging
to obtain if series are short and data is scarce. In this case, institutions should use
judgement and critical thinking to balance economic significance and stationarity
requirement in order to assess if modelling can proceed. In this assessment,
institutions should pay particular attention to the presence of trends, that often leads
to spurious correlations.
(iii) To test for stationarity, standard unit root test may be used, including the
Augmented Dickey-Fuller test, the Phillips-Perron test, the Kwiatkowski–Phillips–
Schmidt–Shin (KPSS test). In case there is evidence of the presence of stochastic
trend, standard transformations can be applied such as quarter-on-quarter or year-
on-year log differencing.
(iv) Seasonality may be checked using X12 or X13 seasonal adjustment algorithms.
Year-on-year differencing could also be used to remove stable seasonal patterns.
Formal structural breaks tests (e.g. Chow test) may be employed if there is visual
evidence of break in the series.
(v) Common stochastic trends between two variables may be explicitly modelled using
the rigorous application of standard co-integration models (e.g. Engle-Granger two
step method or Johansen approach).
(vi) The absence of stationarity of the dependent variable can also be addressed by a
first order time differencing or by autoregressive models. However, this can
potentially lead to further complexity in implementation. Institutions should use
judgement in this choice provided that it is justified and clearly documented.
5.5.3 Differencing: Time differencing should be based upon the following principle. Let 𝑋𝑡 be
a time series of the macroeconomic variable 𝑋 at regular time steps 𝑡. Formally we can
define two types of changes: (i) backward looking returns that estimate the change of the
variable over a previous horizon ℎ and (ii) forward looking returns that estimate the
change of the variable over a coming horizon ℎ. Formally:
𝑋𝑡 − 𝑋𝑡−ℎ 𝑋𝑡+ℎ −𝑋𝑡
∆𝑋 𝐵𝑎𝑐𝑘𝑤𝑎𝑟𝑑 = 𝑎𝑛𝑑 ∆𝑋 𝐹𝑜𝑟𝑤𝑎𝑟𝑑 =
𝑋𝑡−ℎ 𝑋𝑡
It is recommended to build macro models based on backward looking returns as these are
more intuitive to interpret and easier to implement. It is also recommended to compute
backward looking default rates in order to ensure that both the dependent and independent
variables are homogeneous.
5.5.4 Horizon of differencing: Institutions should choose carefully the horizon of return applied
to macro variables, i.e. the period used to compute the change of a variable through time.
Institutions should take notes of the following principles:
(i) For macroeconomic models, common return horizons include quarterly, half-
yearly and yearly. The most appropriate return horizon should be chosen to
maximize the explanatory power of the macro variables.
(ii) Note that the return horizon is not necessarily equal to the granularity of the time
series. For instance, rolling yearly returns can be computed on quarterly time steps.
(iii) Institutions should be aware of the degree of noise in high frequency data.
Consequently judgement should be used when using high frequency returns.
5.5.5 Lags: Variable lags should be considered in the modelling process to capture delayed
effects of macro drivers. The use of large lags (more than 6 quarters) should be justified
since long lags delay the impact of macro shocks on the dependent variable. For each
macro variable, the choice of the most appropriate lag should be based on its statistical
performance and economic meaning.
5.5.6 Smoothing: This means reducing the presence of spikes and outliers in times series. This
is commonly addressed by the usage of moving average. Such practice is permitted but
should be employed with caution and documented. The right balance of smoothing needs
to be found. No smoothing (too much noise) in time series can lead to weak models.
Alternatively, too much smoothing can dilute the strength of correlations. Smoothing
techniques should be documented when applied.
5.5.7 Standard and intuitive transformations should be used. For example, the growth rate of a
variable that can be zero or negative is not a meaningful measure.
5.6.1 The objective of the correlation analysis is to assess the strength of the relationship
between (i) each of the transformed dependent variable (e.g. PD) and (ii) each of the
transformed macro variables, on a bilateral basis. Consequently such univariate analysis
should be performed for each obligor segment. This analysis should include both
statistical and causality perspectives. Relationships should make economic sense and
meet business intuitions.
5.6.2 Institutions should pay attention to the strength and sign of correlations (positive vs.
negative relationships) and assess whether they meet business intuitions. At a minimum,
the following components should be documented: (i) the top macro variables ranked by
correlation strength and (ii) comments and analysis on the observed statistical
relationships vs. expected business intuitions.
5.6.3 A cut-off should be establish to eliminate the transformed macro variables that display
weak and/or incoherent correlations with the independent variables. This reduced
population of transformed macro variables should be used to perform multivariate
analysis.
5.7.1 The objective of this step is to construct relevant and robust relationships between a single
transformed dependent variable (e.g. PD) and several macro variables. The choice of the
macro variables entering each model should be based upon the results of the correlation
analysis. This process results in the construction of a range of multifactor models for each
dependent variable.
5.7.2 In the context of time series regressions, institutions should choose the most appropriate
methodology to perform multifactor regressions. Amongst others, it is recommended to
perform multifactor regressions with or without autoregressive terms. It is recommended
that institutions include several modelling forms as part the pool of possible model
candidates.
5.7.3 The estimation of model coefficients should be performed with recognised professional
statistical software and packages. The entire process should be fully documented and
replicable by an independent party.
5.7.4 Several performance metrics should be used to rank and choose models. As these metrics
depend on the type of models, institutions should use judgements to employ the most
appropriate performance metrics per model type. At a minimum, the adjusted R-square
should be used for multifactor regression models. For models based on the ARIMA form,
a pseudo R-square should be employed as the square of the correlation between the fitted
variable and the original dependent variable.
5.8.1 Standard post-estimation tests should be used to check that the underlying assumptions
are appropriate for all types of macro models. The set of appropriate tests should be based
on best practices in the relevant field / literature. The model development documentation
should clearly indicate (i) the chosen test for each property, (ii) the nature of the 𝐻0
hypothesis, (iii) the cut-off values chosen upfront to determine the rejection or non-
rejection.
5.8.2 In the context of time series regression, regression coefficients should be significant and
residuals should be tested for autocorrelation and normality. The table below indicates
properties that should be tested, at a minimum. Other tests may be considered, if
necessary.
Table 9: Typical statistical tests for models based on time series regression
Property to be Description of the property to be Suggested test
tested rejected (others may exist)
Stationarity Absence of stationarity in each time Augmented Dickey-
series Fuller (ADF)
Co-integration Absence of stationarity in a linear Engle-granger two-
combination of the dependent variable step method
and each independent variable
Multicolinearity High correlation between the Variance Inflation
independent variables Factor
Coefficient The coefficients are not statistically Coefficient p-value
significance significantly different from zero on a t-distribution
Autocorrelation High correlation between the error Ljung-Box test
terms of the model
Heteroscedasticity Absence of relationship between Breusch-Pagan or
independent variables and residuals White test
Normality Normal distribution of the residuals Shapiro Wilk
5.9.1 The model and macroeconomic variable selection should be based on clearly defined
performance criteria using a transparent selection algorithm. The final model should be
able to (i) generate values that fit the historical values of the dependent variable and (ii)
generate accurate predictions.
5.9.2 For each segment, institutions should choose a final model from the list of candidate
models generated from the model construction step. Statistical performance should not
be the only decisive factor to choose a model. Instead, the choice of the final model
should be based upon the combination of various factors. At a minimum, institutions
should use the criteria outlined below. It is essential that institutions include all these
criteria in the selection process. The absence of one criteria could be materially
detrimental to the choice of the most relevant model.
(i) Statistical performance:
a. The chosen model should meet minimum requirements of performance,
statistical stability and robustness as shown by the statistical indicators and
their associated thresholds. Model parameters and forecasts should remain
stable over time.
b. In addition, at the model development stage, it is important to examine the
stability of models: out-of-sample performance and in-sample fit should be
tested and compared across candidate models. A common metric employed
to express model performance is the root mean square error, for which limits
should be established.
(ii) Model sensitivity: Quantitative response of the dependent variable to independent
variables should be meaningful and statistically significant - both quantitatively
and qualitatively. This can be examined through simulating one standard deviation
change in individual dependent variables or by considering the forecast differences
across alternative scenarios.
(iii) Business intuition: The chosen model should be constructed with variables and
relationships that meet logical business and economic intuitions. This means that
the model should be explained by causal relationships.
(iv) Realistic outcomes: Projected values should be consistent with historical
observations and meet economic intuition. Any material jump and/or disconnect
between historical values and forecasted should be explained.
(v) Implementation: When choosing a model, institutions should be mindful of the
implementation and maintenance constraints, which should form part of the choice
of the most appropriate models. For instance, some variables may not be available
as frequently as expected for forecasting. Also, some model formulations may
require autoregressive terms that need specific treatment during implementation.
5.9.3 In order to test the business intuition, for each candidate model, institutions should
forecast the dependent variables (e.g. PD, Credit Index) under a severe downside
scenario. The outcome will therefore be a range of projected dependent variables (one for
each model) under the same scenario. It may become apparent that some candidate
models should be excluded as they generate outputs that deviate too much from economic
and business expectations.
5.9.4 Forecast Uncertainty: Projected forecast are based on mean or median values, around
which uncertainty (i.e. confidence interval) inherently exists. Institutions should ensure
that the model forecast uncertainty are clearly estimated, documented and reported to the
Model Oversight Committee. In the context of time series regression, the confidence
interval around the mean can be estimated empirically or based on the standard deviation
of the residuals under the assumption of normally distributed residuals.
5.10.1 The validation of macro models should be performed by a different and independent party
from the development team, according to the validation principles articulated in the
MMS. If macro models are developed by a third party consultant, then a team within the
institution or another consultant should therefore perform these independent validations.
5.10.2 Monitoring macro models may be challenging due to the low frequency of
macroeconomic data. Institutions are expected to monitor the performance of macro
models once a year. However, exceptional monitoring is required in the case of
exceptional macroeconomic events.
5.10.3 Pre-implementation validation: This step involves the validation of the chosen macro
models immediately after their development, prior to using them in production. The
objective is to ensure that macro models meet a minimum level of quality and that they
are fit for purpose. At a minimum, the following validation steps should be performed.
(i) Development process: The validator should review the development process as per
the principles articulated in the MMS.
(ii) Replication: The validator should replicate the final chosen model per segment and
ensure that the coefficients are correctly estimated.
(iii) Statistical tests: The validator should ensure that statistical tests are correct, that
cut-offs are reasonable and that statistical assumptions are correctly interpreted.
This may necessitate partial replication. Additional statistical tests may be needed.
(iv) Model sensitivity: The validator should measure the elasticity of the model output
to changes in each input variable. The model user and validator should be aware
of the presence of independent variables that dominates other variables in a given
model.
(v) Model stability: The validator should test the model stability, for instance by
removing data points from the original time series (at the start or the end), re-run
the regressions and re-project the dependent variable. The validator should also
examine the stability of the model coefficients.
(vi) Conclusion: When deemed appropriate, the validator can make suggestions for
defect remediation to be considered by the development team.
5.10.4 Post-usage validation: This is otherwise referred to as back-testing, whereby the validator
should compare the realized values of the dependent variable (e.g. PD, LGD, Credit
Index) against the forecasted values based on the macroeconomic scenarios employed at
the time of the forecast. A conclusion should be made based pre-defined confidence
intervals.
5.10.5 Upon the post-usage validation, the validator should make a clear statement regarding the
suitability of the model to be used for another cycle. When deemed appropriate, the
validator can make suggestions for defect remediation to be considered by the
development team.
5.11.1 The expected practices articulated in this section relate to the regular estimation of ECL.
As per current accounting requirements, institutions should estimate an unbiased and
probability weighted ECL by evaluating a range of possible outcomes. Consequently,
institutions should forecast economic conditions over the lifetime of their portfolio. All
the macroeconomic variables employed as input in macro models should be forecasted
until the longest maturity date of the institutions’ portfolio.
5.11.2 Institutions are encouraged to disclose macroeconomic scenarios in their annual reports.
For this information to be meaningful, institutions should provide the values of the main
economic drivers over the next three (3) years with the weight of each scenario.
5.11.3 Institutions should use the most recent set of models to forecast PD and LGD. If the most
recent models are not used in the computation of ECL, this should be mentioned in
monitoring and validation reports and reported to internal and external auditors because
it has direct impacts on financial reporting.
5.11.4 Governance: Institutions can either develop macroeconomic forecasts internally or rely
on third party providers. In both cases, a governance process should be put in place to
guarantee the quality of forecasts.
(i) If scenarios are developed internally, they should be constructed by subject matter
experts with robust economic knowledge, within the institution. The scenarios
should be reviewed by the Model Oversight Committee and the committee in
charge of ECL oversight.
(ii) If scenarios are developed externally, institutions should put in place an internal
validation process, by which the forecasts are checked, errors are adjusted and
economic consistency is ensured. Even if scenarios are provided by an external
party, each institution remains the owner of the economic forecasts and therefore
remains accountable for inconsistencies present in those scenarios.
(iii) To support the adequate estimation of ECL, institutions should produce regular
reports to present the aspects of macro scenario calibration discussed in this
section. The report should address the source of scenarios, their economic
consistency, their severity, weights and potential adjustments.
5.11.5 Weights and severity: As per current accounting requirements, institutions should use
several weighed scenarios. At a minimum, institutions should employ one base, one
upside and one downside scenario for each macro variable. In order to obtain an unbiased
estimation of ECL, both the scenario weights and their associated severity should be
jointly calibrated. For each variable, institutions should pay attention to the relative
weight and severity of the downside scenario vs. the weight and severity of the upside
scenario. Finally, it is recommended to estimate the ECL under each scenario in order to
convey possible ECL volatility and support appropriate risk management.
5.11.6 Forward looking information: As per current accounting requirements, ECL should be
based on forward looking information that is relevant reasonable and supportable. This
should be understood as follows:
(i) The economic information should be based on consensus, when possible, rather
than a single source.
(ii) The economic forecasts should be realistic in terms of trend, level and volatility.
For instance, economic forecasts assuming a constant positive inflation should not
push asset prices to excessive and unrealistic levels in the long term. This feature
is particularly relevant for real estate collaterals.
(iii) The divergence between the scenarios (base, upside, downside) should meet
economic intuition and business sense. Such divergence should follow a logical
economic narrative.
5.11.7 Benchmarks: Aside from ECL measurement, institutions employ existing scenarios for
planning purposes, with varying severity and probability of occurrence. Amongst others,
dedicated scenarios are used for the ICAAP, the recovery and resolution plan, and for
stress testing purpose. These scenarios should not be employed as input for ECL
computation because they do not represent an unbiased probability-weighted set of
scenarios. Similarly, the macroeconomic scenarios provided by the CBUAE as part of
regulatory enterprise-wide stress testing exercises should not be used as input for ECL
computation. All these alternative scenarios can only be used as distant comparables for
the ECL scenarios. However, this comparison should be made with caution because the
calibration (severity, likelihood) of the ECL scenarios is likely to be different.
5.11.8 The construction of the scenarios should be economically coherent and therefore should
follow a set of rules to be consistent. Let 𝑋𝑡 be a time series of the macroeconomic
variable 𝑋 at regular time steps 𝑡. For the purpose of articulating this point, we will use
three scenarios. The time series of 𝑋 corresponding to each scenario are noted 𝑋𝑡 𝐵𝑎𝑠𝑒 ,
𝑋𝑡 𝑈𝑝 and 𝑋𝑡 𝐷𝑜𝑤𝑛 .
(i) Scenarios should be constructed in such way that their values diverge after a given
date, called the forecasting date, noted 𝑇. The time series for the three scenarios
should be identical prior to the date of forecast and diverge after the date of
forecast. Formally:
𝑋𝑡 𝐵𝑎𝑠𝑒 = 𝑋𝑡 𝑈𝑝 = 𝑋𝑡 𝐷𝑜𝑤𝑛 𝑓𝑜𝑟 𝑡 ≤ 𝑇
𝑋𝑡 𝐵𝑎𝑠𝑒 ≠ 𝑋𝑡 𝑈𝑝 ≠ 𝑋𝑡 𝐷𝑜𝑤𝑛 𝑓𝑜𝑟 𝑡 > 𝑇
(ii) The portfolio date noted 𝐾 employed in loss computation should be close to the
forecasting date. Therefore, institutions should keep updating the macroeconomic
forecasts along with the portfolio date, in order to minimize the time difference
between 𝑇 and 𝐾. It may happen that ECL reporting is done at a higher frequency
than the update of macroeconomic forecasts. In this case, the time step at which
scenarios start diverging occurs before the portfolio date 𝐾. Formally, for 𝑇 < 𝐾:
𝑋𝑡 𝐵𝑎𝑠𝑒 ≠ 𝑋𝑡 𝑈𝑝 ≠ 𝑋𝑡 𝐷𝑜𝑤𝑛 𝑓𝑜𝑟 𝑡 ∈ [𝑇, 𝐾].
This misalignment is likely to create unwarranted effects in particular if scalers are
used in PD modelling. Therefore, the maximum delay between the two dates
should be no more than three months: 𝐾 − 𝑇 ≤ 3 𝑚𝑜𝑛𝑡ℎ𝑠. If this difference is
greater than three (3) months, the impact on the forecasted PD and LGD should
be analysed and documented by the independent model validation team.
(iii) Beyond the forecasting start date, the upside and downside scenarios should not be
constructed by a constant parallel shift (or scaling) of the base scenarios. Rather,
the upside and downside scenarios should be constructed such that they display a
gradual divergence from the base scenario through time (in growth rate terms if
growth rates are applied). This property should apply to the stationary transformed
macroeconomic variable. Consequently, the forecasted PD and LGD should follow
the same pattern. Formally:
𝑋𝑡 𝑈𝑝 − 𝑋𝑡 𝐵𝑎𝑠𝑒 ≠ 𝑋𝑡+1 𝑈𝑝 − 𝑋𝑡+1 𝐵𝑎𝑠𝑒 ∀ 𝑡 > 𝑇
𝑋𝑡 𝐷𝑜𝑤𝑛 − 𝑋𝑡 𝐵𝑎𝑠𝑒 ≠ 𝑋𝑡+1 𝐷𝑜𝑤𝑛 − 𝑋𝑡+1 𝐵𝑎𝑠𝑒 ∀𝑡 >𝑇
(iv) Any scaler subsequently constructed based on these scenarios should follow the
same pattern: a gradual divergence from the base scenario.
5.11.9 The principles articulated in this section about scenario forecast should also be included
in the scope of review of the validation process. The validation process should test, assess
and document practices for scenarios forecasts, including the governance, scenario
construction and calibration.
6.1 Scope
6.1.1 For the purpose of this section, and in order to simplify technical considerations, both
interest rate risk (for conventional products) and profit rate risk (for Islamic products)
will be referred to as Interest Rate Risk in the Banking Book (“IRRBB”). Both lead to a
similar structural risk on institutions’ balance sheet.
6.1.2 Institutions should implement models to address the requirements articulated in “Interest
rate and rate of return risk in the banking book Standards” issued by the CBUAE in 2018
(notice 165/2018), hereby referred to as the “CBUAE Standards on IRRBB”. In
addition to the CBUAE Standards, institutions should refer to the Standards articulated
by the Basel Committee on Banking Supervision in April 2016: “Interest rate risk in the
banking book”, hereby referred as the “Basel Standards on IRRBB”.
6.1.3 According to the CBUAE Standards on IRRBB, interest rate risk should be captured
through changes in both (i) expected earnings and (ii) the economic value of the balance
sheet. In order to ensure more homogeneity in the methodology employed by institutions
across the UAE, the MMG hereby presents some guidance on IRRBB modelling. The
IRRBB requirements related to governance, management, hedging and reporting are
covered in a separate CBUAE Standards on IRRBB.
6.2 Metrics
6.2.1 Institutions should identify all positions in interest sensitive instruments including:
(i) All assets, which are not deducted from Common Equity Tier 1 (“CET1”) capital,
and which exclude (a) fixed assets such as real estate or intangible assets and (b)
equity exposures in the banking book.
(ii) All liabilities, including all non-remunerated deposits, other than CE1 capital ; and
(iii) Off-balance sheet items.
6.2.2 Institutions should reconcile their exposures against their general ledger and their
published financials. Differences may arise for valid reasons, which should be
documented. This reconciliation process should be included in the model documentation
and should be verified by the finance team on a yearly basis.
6.2.3 Changes in expected earnings and economic value can be captured through several
possible metrics. At a minimum, the following metrics should be computed. These are
referred as “IRRBB metrics”:
(i) Gap risk: It is defined as the difference between future cash in-flows and cash-
outflows generated by both assets and liabilities. The cash in-flows and out-flows
are derived from the allocation of all relevant interest rate sensitive instruments
into predefined time buckets according to their repricing or their maturity dates.
These dates are either contractually fixed or based upon behavioural assumptions.
The resulting metrics is the net position (gap) of the bank per future time bucket.
(ii) Gap risk duration: Also referred to as partial duration or partial “PV01”. It is
defined as the modified duration of the gap per maturity bucket. The modified
duration is the relative change in the present value of the position caused by a 1
basis point change in the discount factor in a specific maturity bucket. The resulting
metrics is a term structure of PV01 per maturity bucket.
(iii) Economic value of equity: Also referred to as “EVE”. It is defined as the difference
between the present value of the institution’s assets minus the present value of
liabilities. The change in EVE (“∆EVE”) is defined as the difference between the
EVE estimated with stressed discount factors under various scenarios, minus the
EVE estimated with the discount factors as of the portfolio reporting date.
(iv) Net interest income: For the purpose of the MMG, and in order to simplify
notations, both Net Interest Income (for conventional products) and/or Net Profit
Income (for Islamic Products) are referred to as “NII”, defined as the difference
between total interest (profit) income and total interest (profit) expense, over a
specific time horizon and taking into account hedging. The change in NII (“∆NII”)
is defined as the difference between the NII estimated with stressed interest rates
under various scenarios, minus the NII estimated with the interest rates as of the
portfolio reporting date. ∆NII is also referred to as earning at risk (“EAR”).
6.3.1 The models employed to compute the metrics above should follow the principles
articulated in the MMS. In particular all IRRBB models should follow the steps in the
model life-cycle. The assumptions and modelling choices surrounding IRRBB models
should not be the sole responsibility of the ALM function nor the market risk function.
Rather, these assumptions should be presented to and discussed at the appropriate
governance forum reporting to the Model Oversight Committee.
6.3.2 The modelling sophistication of the EVE should depend upon the size and complexity of
institutions. For that purpose, different requirements are defined in function of their
systemically important nature. The modelling requirements presented hereby should be
regarded as minimum standards. To remain coherent with Basel principles, higher
standards apply to large and/or sophisticated institutions (“LSI”). However, the other
institutions may choose to implement models with higher standards than the one
prescribed for them. This proportionality is an exception to the MMG due to the
prescriptive nature of the Basel methodology surrounding IRRBB.
6.3.3 The requirements below refer to the methodology articulated in section IV (“The
standardised framework”) of the Basel Standards on IRRBB. All institutions are
requirements to fully understand this framework.
6.3.4 The estimation of EVE should be based upon the following principles: (a) it includes all
banking book assets, liabilities and off-balance sheet exposures that are sensitive to
interest rate movements, (b) it is based on the assumption that positions roll off, and (c)
it excludes the institution’s own equity. The approach subsequently depends on the type
of institutions.
(i) LSIs should compute EVE as the difference between discounted assets and
liabilities at a granular level. Institutions should aim to perform this computation
at a facility level. For practical reasons, some facilities could be aggregated,
provided that they are homogeneous and share the same drivers and features. All
inputs including, but not limited to, cash-flows, time buckets, risk-free rates, option
risk and basis risk should also be estimated at a granular level.
(ii) It should be noted that the Gap risk and the Gap risk duration are not directly used
to estimate EVE in the context of a granular full revaluation. However, the Gap
risk and Gap risk duration should be estimated and reported in order to manage
IRRBB.
(iii) Non-LSI can compute EVE at a higher level of granularity, according to the
principles outlined in the Basel Standards on IRRBB and in particular according
to article 132. The methodology is based upon the summation of discounted Gap
risk across time buckets, rather than a granular NPV estimation at facility level.
Institutions should pay particular attention to the cash flow allocation logic in each
time bucket.
(iv) Irrespective of their size, all institutions should compute ∆EVE as the difference
between EVE estimated under interest rate scenarios and the EVE under the current
risk-free rates. The final EVE loss and the standardised risk measure employed in
Pillar II capital should be computed according to the method explain in the article
132 (point 4) of the Basel Standards on IRRBB, whereby EVE loss should be
aggregated across currencies and scenarios in a conservative fashion.
6.3.5 The estimation of NII should be based upon the following principles: (a) it includes all
assets and liabilities generating interest rate revenue or expenses, (b) it includes
commercial margins and (c) no discounting should be used when summing NII across
time buckets. The approach subsequently depends on the type of institutions.
(i) LSIs should compute NII at a granular level, both for facilities and maturity time
steps. NII should be based on expected repricing dates upon institutions’ business
plan of future volume and pricing. Therefore LSIs should estimate ∆NII as the
difference in NII under the base and the stress scenarios. Such granular
computation should include option risk and basis risk.
(ii) Non-LSIs can compute ∆NII by allocating interest revenue and interest expenses
in the standardised time buckets used for EVE. Non-LSI institutions can compute
∆NII by estimating directly their earning at risk on each expected repricing date.
(iii) For the purpose of risk management, institutions are free to model NII based on
static or dynamic balance sheet assumptions (although LSIs are recommended to
employ the latter). Institutions can also choose the NII forecasting horizon.
However, for Pillar II assessment as part of the ICAAP and for reporting to the
CBAUE, the following, institutions should compute NII over 1 year; in addition
LSIs should also compute NII over 3 years.
6.3.6 Institution’s own equity: For NII estimation, institutions should include interest-bearing
equity instruments. For EVE, in the context of the MMG, institutions should compute
two sets of metrics by first including and then excluding instruments related their own
equity. These two types of EVE will be used for different purposes.
(i) CET1 instruments should be excluded at all times to avoid unnecessary
discrepancies related to the choice of behavioural maturity associate to this
component.
(ii) Some institutions have a large proportion of interest-sensitive instruments, in
particular in the AT1 components. Consequently, these institutions should estimate
and report a first set of EVE sensitivities by including these instruments. This type
of EVE is useful for proactive management of IRRBB.
(iii) Conversely, one of the objectives of assessing IRRBB is to ensure that institutions
hold enough capital to cover such risk, which is articulated through the ICAAP.
Institutions should not use part of their capital to cover a risk that is itself generated
from capital. Therefore, institutions should also compute and report EVE by
excluding their own equity entirely. This type of EVE is useful to estimate the
Pillar II capital charge arising from IRRBB.
6.3.7 Commercial margins: The treatment of commercial margins is different between NII and
EVE. However, the recommendation is similar for both LSIs and non-LSIs.
(i) All institutions should include commercial margins in NII estimation. Margins
should be adjusted based on business plans and expected customer behaviour in a
given interest rate environment. For instance, it might be assumed that margins
will be increased to retain depositors in a falling interest rate environment.
(ii) All institutions have the option to include or exclude commercial margins in EVE
estimation. However, institutions should also aim to estimate the impact of
commercial margins on EVE. For consistency, if margins are included in the cash
flows (numerator), then discount factors should also reflect the corresponding
credit spread of the obligors (denominator). Such estimation should be done at
homogeneous pools obligors with similar credit risk profiles.
6.3.8 Basis risk: This risk arises when assets and liabilities with the same tenor are discounted
with different ‘risk-free’ interest rates. Potential credit risk embedded in these rates makes
them not entirely risk-free, hence the existence of bases. A typical example is an asset
priced with the US LIBOR curve but funded by a liability priced with the US Overnight
Index Swap (“OIS”) curve, thereby creating an LIBOR-OIS basis leading to different
NPV and NII from both the asset and the liability. Another example is the recent
introduction of USD Secured Overnight Financing Rate (“SOFR”) creating a LIBOR-
SOFR basis. LSIs are required to fully identify and assess basis risk. They should employ
the appropriate risk-free rate for each instrument type, thereby capturing basis risk in all
the IRRBB metrics. While non-LSIs are not expected to fully quantify basis risk on a
regular basis, they should perform an approximation of this risk to assess whether further
detailed quantification is necessary.
6.3.9 Currency risk: The currencies of assets and liabilities have a material impact on the
resulting IRRBB, therefore this dimension should be fully addressed by institutions’
modelling practice.
(i) All the IRRBB metrics should be estimated for each currency in which the
institution has material exposures, i.e. when the gross exposure accounts for more
than five percent (5%) of either the gross banking book assets or gross liabilities.
For those, the interest rate shocks should be currency-specific.
(ii) For the estimation of the capital charge, the Basel Standards on IRRBB suggests
to sum the maximum change in EVE across currencies without offsetting. While
the CBUAE recognises that no offsetting is conservative for pegged currencies,
(typically USD/AED), institutions should manage basis risk appropriately since
material bases have been observed between USD rates and AED rates.
Consequently, each institution has the option to offset ∆EVE between pegged
currencies, only if it can demonstrate that it does capture the basis risk between
these currencies with dedicated stress scenarios.
6.3.10 Non-performing assets (“NPA”): Institutions should define clearly the treatment of non-
performing assets in their modelling practice, according to the following principles.
(i) NPA (net of provisions) should be included in the estimation of EVE. In most default
cases, LGD>0% therefore a recovery is estimated at some point in the future. The
LGD is estimated by discounting expected recoveries with a discount rate generally
based on the effective interest rate of the facility. In the context of IRRBB, a change
in the interest rate environment should have an impact the present value of discounted
recoveries and therefore on LGD. This effect could likely impact EVE. Finally,
consideration should also be given to rescheduled facilities and/or forbearance with
payment holidays where interests are accrued. The postponement could results in
lower PV under scenarios with increasing rates.
(ii) The treatment of NPA (net of provisions) for NII computation is left to the discretion
of banks. Under a static balance sheet assumption, non-performing assets will not
generate cash inflows. A change in rates would have no impact the NII from such
assets. However, under dynamic a balance sheet assumption, some NPA could return
to a performing status and therefore impact NII.
6.4.1 Option risk constitutes a fundamental building block of IRRBB. Option risk is defined as
the potential change of the future flows of assets and liabilities caused by interest rate
movements. In the context of the MMG, option risk refers to deviations from either
contractual maturity or expected behavioural maturity. Consequently, these options can
be explicit or implicit. The exercise of these options are a function of the contractual
features of the product, the behaviour of the parties, the current interest rate environment
and/or the potential interest shocks. All institutions should capture option risks,
irrespective of their size and sophistication.
6.4.2 In order to model option risk appropriately, all institutions should, at a minimum,
undertake the following steps:
(i) Identify all material products subject to embedded options,
(ii) Ensure that assumptions employed in modelling are justified by historical data,
(iii) Understand the sensitivity of the IRRBB metrics to change in the assumptions
related to option risk, and
(iv) Fully document the method and assumptions used to model option risk.
6.4.3 LSIs should incorporate option risks at a granular level and undertake the necessary
analysis to substantiate their assumptions. Option risk can be modelled and estimated at
an aggregated level that displays similar behavioural characteristics, but the model results
should be applied as a granular level. For that purpose, LSIs can use the standardised
approach as a starting point and elaborate on it, in such a way that the approach meets the
size and complexity of the institution. Ultimately, cash flows from assets and liabilities
should be conditional upon the level of interest rates in each scenario. The
methodology and assumptions employed to model optionality should be fully
documented.
6.4.4 Non-LSIs should use the EVE approach articulated in the Basel Standards on IRRBB,
whereby option risk is incorporated via the dependency of cash flows on interest rate
levels by using conditional scalers. Subsequently, under each stress scenario with specific
interest rate shocks, institutions should employ a different set of netted cash flows per
bucket to compute EVE. In other words and using the Basel formulation, the cash flow
𝐶𝐹𝑖,𝑐 (𝑡𝑘 ) should vary for each interest rate scenario, where 𝑖, 𝑐 and 𝑡𝑘 are respectively the
interest rate scenario, the currency and the time bucket. The below steps explain further
the standardised approach.
6.4.5 Non-maturity Deposits (“NMD”): All institutions should model option risk for NMD, as
described in the Basel Standards on IRRBB, from article 110 to 115. The objective is to
assess the behavioural repricing dates and cash flow profiles of NMD. In particular,
institutions should undertake the following steps:
(i) Segregate NMD into categories of depositors, considering at a minimum, retail
clients, wholesale clients and Islamic products.
(ii) Identify stable and core deposits, defined as those that are unlikely to be repriced,
even under significant changes in the interest rate environment. For that purpose,
institutions should analyse historical patterns and observe the change in volume of
deposits over long periods. Institutions should describe the data sample and the
statistical methodology used for this analysis.
(iii) For each segment, apply the caps mentioned in Table 2 of the Basel Standards on
IRRBB and allocate the cash flows in the appropriate time bucket based on their
estimated maturity.
(iv) Construct assumptions regarding the proportion of core deposits and their
associated maturity under each interest rate scenario and in particular the potential
migrations between NMD and other types of deposit. These assumptions should
reflect the most likely client behaviour but with a degree of conservatism.
Institutions should bear in mind the importance of portfolio segmentation on
behavioural modelling.
6.4.6 Fixed rate loans: Such instruments are subject to prepayment risk because a drop in
interest rates is susceptible to accelerate their early prepayment. In addition, restructuring
events can also change their expected cash flow profiles. Consequently, all institutions
should implement the approach mentioned in articles 120 to 124 of the Basel Standards
on IRRBB. In particular, institutions should proceed as follows.
(i) Business-as-usual prepayment ratios should be estimated per product type and per
currency.
(ii) These ratios should be multiplied by the scalers in Table 3 of the Basel Standards
on IRRBB, that depend on the interest rate shock scenarios, in order to derive
adjusted prepayment rates. If the institution has already defined prepayment rates
under each scenario based on its own internal historical data, then it can use these
rates, provided that they are fully documented and justified. Portfolio concentration
and segmentation should be taken into account when performing such behavioural
modelling.
(iii) The adjusted prepayment rates should be employed to construct the repayment
schedule under a given scenario. The choice of the time buckets where the
prepayments are made, should also be justified and documented.
6.4.7 Term deposits: Such instruments are subject to redemption risk because an increase in
interest rates is susceptible to accelerate their early withdrawal. Consequently, all
institutions should implement the approach mentioned in the articles 125 to 129 of the
Basel Standards on IRRBB. In particular, institutions should proceed as follows:
(i) Business-as-usual redemption ratios should be estimated per product type and per
currency.
(ii) These ratios should be multiplied by the scalers in Table 4 of the Basel Standards
on IRRBB, that depend on the interest rate shock scenarios, in order to derive
adjusted redemption rates.
(iv) The adjusted redemption rates should be used to derive the proportion of
outstanding amount of term deposits that will be withdrawn early under a given
scenario. If the institution has already defined redemption rates under each scenario
based on its own internal historical data, then it can use these rates, provided that
they are fully documented and justified. Portfolio concentration and segmentation
should be taken into account when performing such behavioural modelling.
(iii) That proportion is finally allocated to the overnight time bucket, per product type
and per currency, as per article 127 of the Basel Standards on IRRBB.
(iv) Finally, institutions should take into consideration off-balance sheet exposures in
the form of future loans and expected drawings on committed facilities.
6.4.8 Automatic interest rate options: All institutions should follow the methodology
articulated in the Basel Standards on IRRBB in articles 130 to 131. Automatic interest
rate options should be fully taken into account in the estimation of both EVE and NII.
6.5.1 All institutions should compute ∆EVE and ∆NII under the six scenarios prescribed in
Annex 2 of the Basel Standards on IRRBB and pasted in the following table. The interest
rate shocks for AED can be directly derived from those corresponding to USD. For
convenience, the AED shocks have been computed and provided below. For other
currencies, all institutions should compute themselves the corresponding interest shocks
based on the methodology outlined in the Basel Standards on IRRBB. The six interest
rate shocks are as follows:
(i) Parallel shock up,
(ii) Parallel shock down,
(iii) Steepener shock (short rates down and long rates up),
(iv) Flattener shock (short rates up and long rates down),
(v) Short rates shock up, and
(vi) Short rates shock down.
6.5.2 In addition to the standard shocks prescribed by the Basel Standards on IRRBB, LSIs
should define other scenarios combining shift of yield curves with changes in basis and
commercial margins in order to comprehensively capture the risk profile of their balance
sheet structure. These institutions should ensure that scenarios are commensurate with
the nature, and complexity of their activities.
The choice of scenarios should be supported by an appropriate governance and fully
documented. All institutions should integrate the IRRBB scenarios and results in their
stress testing framework and in enterprise-wide stress testing exercises.
Table
Standard12: Standard
shocks (bps) pershocks
scenario per scenario
for AED (bp)
prescribed forBIS
by the AED prescribed by the BIS method
methodology
Tenors
Time Buckets (M: months ; Y: Years) (i) (ii) (iii) (iv) (v) (vi)
(years)
t = Overnight (O/N) 0.0028 200 -200 -195 240 300 -300
O/N < t <= 1M 0.0417 200 -200 -192 237 297 -297
1M < t <= 3M 0.1667 200 -200 -182 227 288 -288
3M < t <= 6M 0.375 200 -200 -165 210 273 -273
Short-Term
6M < t <= 9M 0.625 200 -200 -147 192 257 -257
9M < t <= 1Y 0.875 200 -200 -130 175 241 -241
1Y < t <= 1.5Y 1.25 200 -200 -106 151 219 -219
1.5Y < t <= 2Y 1.75 200 -200 -78 123 194 -194
2Y < t <= 3Y 2.5 200 -200 -42 87 161 -161
3Y < t <= 4Y 3.5 200 -200 -3 48 125 -125
Medium-Term 4Y < t <= 5Y 4.5 200 -200 28 17 97 -97
5Y < t <= 6Y 5.5 200 -200 52 -7 76 -76
6Y < t <= 7Y 6.5 200 -200 70 -25 59 -59
7Y < t <= 8Y 7.5 200 -200 84 -39 46 -46
8Y < t <= 9Y 8.5 200 -200 96 -51 36 -36
9Y < t <= 10Y 9.5 200 -200 104 -59 28 -28
Long-Term
10Y < t <= 15Y 12.5 200 -200 121 -76 13 -13
15Y < t <= 20Y 17.5 200 -200 131 -86 4 -4
t > 20Y 25 200 -200 134 -89 1 -1
6.5.3 Institutions should consider the possibility of negative interest rates and understand the
impact on their balance sheet and business models. For each asset and liability, if the legal
documentation of the contract stipulates a certain treatment of negative rates, then this
treatment should be used. If the legal documentation is silent on the treatment of negative
rates, then such negative rates should be used to price assets, but they should be floored
at 0% for deposits (liabilities) because there is little evidence supporting the assumption
that both retail and corporate clients would accept being charged for depositing their
funds in UAE banks.
6.6.1 Institutions should validate all EVE and NII models according to the principles articulated
in the MMS and in particular related to model life cycle management.
6.6.2 The validation of EVE and NII models should be based upon the principles articulated
for both deterministic and statistical models. The validation exercise should ensure that
modelling decisions are justified and documented and cover all the model components
presented in the previous sections. In particular, the appropriate use of data input should
also be reviewed by the validator.
(i) The validator should ensure that the mechanistic construction of these models is
sound. This should be tested with partial replication and internal consistency
checks.
(ii) The validator should ensure that the financial inputs are correctly flowing into
these models. This step may require the join work between several teams including
the risk and finance teams.
(iii) The validator should ensure that the results produced by these models are coherent.
For that purpose sensitivity analysis can be performed.
(iv) Finally, some of the inputs are derived from statistical models, including the
behavioural patterns observed for non-maturity deposits, fixed rate loans and term
deposits. Consequently, the validation should consider the robustness, stability and
accuracy of the ancillary statistical models employed to derived inputs to EVE and
NII models.
6.6.3 Overall, the validation process of EVE and NII models should focus on the economic
meaning and business intuition of the model outputs. The development and validation
processes should not be dominated by the mechanistic aspect of these models, but also
ensure that those are suitably designed to support robust decision making and the
appropriate management of interest rate risk in the banking book.
7.1 Scope
7.1.1 The concept of Net Present Value (“NPV”) is commonly used to estimate various metrics
for the purpose of financial accounting, risk management and business decisions. This
section focuses on standalone NPV models employed for the purpose of general asset
valuation, covering, amongst others, investment valuation, collateral valuation and
financial modelling to estimate the cost of facility rescheduling (or restructuring). The
discounting component embedded in ECL, LGD and CVA models is discussed in other
sections.
7.2 Governance
7.2.1 Standalone NPV models should be included in the scope of models managed by
institutions. These models should be included in the model inventory and subject to the
life-cycle management articulated in the MMS. The management of these models should
be governed by the Model Oversight Committee.
7.2.2 These models are deterministic in substance as they do not need statistical calibrations.
Therefore the recalibration step of the life-cycle does not apply to them. However, the
suitability of inputs and the assumptions embedded in the model construction should be
reviewed on a regular basis, or whenever there is a significant change in assumptions.
7.2.3 Institutions should establish a general consistent methodology for standalone NPV
computation that follows the entire model life-cycle management, including the
validation step. Independent model validation should follow the principles articulated in
the MMS.
7.2.4 In addition to the regular generic validation of the approach and usage, institutions should
define materiality thresholds, beyond which the valuation of specific transactions should
be subject to modelling review by an independent model validator. The threshold of
materiality should be clearly documented.
7.2.5 For NPV computation in the context of facility rescheduling (restructuring), the choice
of methodology, inputs and assumptions should follow accounting principles.
7.3 Methodology
7.3.1 The methodology surrounding NPV computation can be split into two parts: (i) the
mathematical mechanistic considerations and (ii) the choice of inputs. The mathematical
considerations surrounding NPV computation are well documented in accounting
rulebooks, practitioner guidelines and academic literature. Consequently, institutions
have limited room to deviate from these rules and are expected to apply rigorously these
principles in a transparent fashion. Institutions can exercise some judgement regarding
the choice of inputs, although a tight framework is generally provided by accounting
standards.
7.3.3 Inputs: For a given set of mechanistic rules in place, the choice of inputs has a material
impact on the NPV values produced by the model. In particular:
(i) The discount factor should be chosen to reflect the opportunity cost of lending or
investing the same notional elsewhere at a similar level of risk. It should reflect the
contractual obligations of all parties involved in the transaction.
(ii) In the context of facility rescheduling (or restructuring), the discount factor
employed to compute the NPV of the original and the new facilities should be
based on the same effective interest rate as the contractual obligations of the
original facility.
(iii) In addition, if there is evidence that the creditworthiness of the obligor has
deteriorated, a credit premium should be added to the discount factor of the newly
rescheduled facility. The calibration of this credit premium should be substantiated
by market analysis and comparables. If no credit premium is added, justification
should be provided.
(iv) In the context of facility rescheduling (or restructuring), the cash-flows of the
original and new facilities should reflect the original and the new contractual
obligations, respectively. This is applicable to the principal repayment flows and
interest rate payments. In particular, if the interest of a restructured facility has
been dropped, the received cash-flows should include lower interest payments.
(v) In the case of assets and facilities with floating interest rates or resetting rates, the
source of the input rates should be clearly identified. Assumptions regarding
forward rates should be based upon the term structure or interest rate at the date of
valuation.
(vi) In the context of facility restructuring (or rescheduling) with floating rates or
resetting rates, the reference interest rates should be identical for both the original
facility and the new facility.
(vii) If several choices of inputs are envisaged for the same asset, institutions should
perform several valuations under a different set of inputs and choose the most
appropriate one. This choice should be clearly justified, documented and validated.
The chosen set of assumptions are not necessarily those leading to the lower P&L
impact.
7.4 Documentation
7.4.1 All standalone NPV models should be fully documented. The documentation should
address the methodology, the assumptions and the principles behind the choice of inputs.
7.4.2 For each valuation exercise deemed material, institutions should produce dedicated
documentation at the time of the exercise in order to justify the model inputs. Institutions
should provide the business rationale, the economic context and the background for such
valuations in order to justify the choice of inputs. This is particularly relevant for facility
rescheduling (or restructuring).
7.5.1 All NPV models should be included in the scope of the validation exercise undertaken by
the institution and in line with the principles articulated in the MMS and, in particular,
the validation checks related to deterministic models.
7.5.2 The validation process should cover, at a minimum, the assumptions, inputs and usage of
the general NPV methodology. In addition, the review should cover specific valuations
deemed material on a sample basis to ensure that the choice of inputs are coherent with
the principles articulated in the general NPV methodology documentation.
7.5.3 The general principles of the NPV computation methodology should be reviewed on a
regular basis. The choice of inputs and assumptions in the context of material valuations
should be reviewed for each material restructuring events.
7.5.4 The validation exercise should ensure that the model inputs reflect accurately the legal
terms of both the original agreement and the new agreement. It should also ensure that
the model outputs meet economic and business intuitions. This is particularly relevant for
restructurings over a long time horizon where material uncertainty arises. Finally, the
validation exercise should pay particular attention to the calibration of the credit spread
premium in the context of a deterioration in the obligor’s creditworthiness.
APPENDIX
The MMG contains several numerical thresholds that institutions should follow.
The following table indicates the relevant Articles to facilitate their implementation.