Customer Logins

Obtain the data you need to make the most informed decisions by accessing our extensive portfolio of information, analytics, and expertise. Sign in to the product or service center of your choice.

Customer Logins

FRTB Decision-Making: The Data Quality Paradigm

29 October 2018 Gil Shefi

The idiomatic phrase "Garbage In, Garbage Out" still rings true 60 years after it was first coined. It has expanded from the world of computer science to encompass the wider remit of decision-making and the poor decisions made due to inaccurate, incomplete or inconsistent data. It is a concept that goes right to the heart of the main challenges associated with the upcoming Fundamental Review of the Trading Book (FRTB) requirements.

When we talk about "data quality" in the context of the decision-making process, is it just about having access to good quality data sources or is it also about how data is processed? As Ronald Coase says: "If you torture the data long enough, it will confess"; in other words, bad data processed by the "proper" model may result in an answer, but it won't necessarily be a very reliable answer or indeed the answer expected.

Another way to look at this "data quality paradigm" is as follows:

Bill Coen, Secretary General of the Basel Committee on Banking Supervision (BCBS), recognized the data quality issue in the context of FRTB during the ISDA conference in April 2018 when he said:

".… The Committee has conducted many quantitative exercises on market risk, both before and after the publication of the market risk framework, with a QIS exercise currently under way. While the quality of data submitted by banks has improved over time, data quality concerns remain. Thus, a significant proportion of banks' data has been excluded from the Committee's analysis. … as a result, the Committee has in some areas been left with a small sample of observations to finalise certain outstanding revisions. ….".

When Coen (and others) talk about data in the FRTB context and more specifically in the context of non-modellable risk factors (NMRFs), they most likely mean Real Price Observation (RPO) data. However, RPOs alone do not portray the full picture. Through the omission of reference data and relevant market data, an incomplete data set is created. Enrichment of the transactional data is therefore vital to enable better classification and enhance accuracy, which in turn unlocks meaningful modellability test results.

And what did Coen mean when he mentioned quality concerns? Data quality has multiple dimensions: accuracy, completeness, consistency, uniqueness, timeliness, staleness and conformance. It is most likely that the BCBS encountered a myriad of quality issues likely touching upon most of the dimensions mentioned above, reinforcing the point that quality is never a one-dimensional definition.

It is therefore not surprising that the BCBS had to exclude "a significant proportion of banks' data" when conducting quantitative exercises on market risk. Clearing data is a good example of just one element of the data reporting conundrum. It is no secret that there is no standard for the reporting of clearing data. This has a knock-on effect on downstream reporting and probably explains why the BCBS struggled to make sense of the data reported by the banks. It is just one manifestation of the data aggregation and reporting issues within the banking system, identified by regulators in the post-crisis era.

For some use cases which require higher level reporting or trend analysis, clearing and trade repository data can be of great value. However, the strict modellability criteria for FRTB (the 24 RPO count and one-month gap or whatever the rules stipulate in the final text) and the clear requirement to provide evidence to supervisors cannot be based on subjective assumptions; both the input data and the data models must be robust and traceable.

So, by putting data and quality back together in the context of FRTB/NMRF, the solution for the NMRF challenge should not just be a measurement of RPO count but a more holistic data governance approach covering the following areas:

  • RPO coverage from the perspective of unique underlying assets as well as total count;
  • Reference data comprehensiveness and accuracy of the pricing data used to enrich and better classify the RPO data;
  • A data model to validate and normalize the data as well as deal with events such as errors in deal capture (fat finger), early terminations (partial or full) etc.;
  • Data dictionary / transaction taxonomy to ensure consistency, uniqueness and completeness of the data to allow accurate mapping of the RPOs to risk factors;
  • Committed Quotes: Are RPOs, both pre and post, treated homogeneously or are committed quotes being treated differently;
  • Timeliness of the service: Is it an EOD service or does it take T+2/3 to reconcile and process the data?
  • Is the pool of RPOs based on inventory or single evidence of trade?

The pursuit of good quality data is an ongoing struggle for most banks; this pain is amplified tenfold when trying to pool and amalgamate multiple sources (such as bank submissions, exchanges and trade repositories). Banks realize that relying on just their own trading activity can only take them so far, leaving them with a large universe of NMRFs. The BCBS acknowledged as much in its March 2018 Consultative Document by emphasizing the importance of pooling in mitigating the NMRF challenge. Participating in a pooling scheme is no longer a nice-to-have but a must-have. Therefore, banks requiring third party data to help them with their NMRF challenge must look not only for "good data" but also for good data models and processes as they have become ever more paramount.

In conclusion, with the final FRTB text due to be published shortly, banks are now re-energizing their FRTB programs and starting to work in earnest on their IMA plans. This will require them to make countless decisions and in this decision-making process, the scarcest commodity will be good quality data. The data quality paradigm of Garbage In, Garbage Out will force banks to make the acquisition and processing of reliable data a top priority. And time is short: too long a delay and the paradigm will mutate and become "Garbage data in, Garbage decisions out".

Posted 29 October 2018 by Gil Shefi, Managing Director-FRTB, IHS Markit

IHS Markit provides industry-leading data, software and technology platforms and managed services to tackle some of the most difficult challenges in financial markets. We help our customers better understand complicated markets, reduce risk, operate more efficiently and comply with financial regulation.

Follow Financial Services


Follow Us

{"items" : [ {"name":"share","enabled":true,"desc":"<strong>Share</strong>","mobdesc":"Share","options":[ {"name":"facebook","url":"","enabled":true},{"name":"twitter","url":"","enabled":true},{"name":"linkedin","url":"","enabled":true},{"name":"email","url":"?subject=FRTB Decision-Making: The Data Quality Paradigm | IHS Markit &","enabled":true},{"name":"whatsapp","url":"","enabled":true}]}, {"name":"rtt","enabled":true,"mobdesc":"Top"} ]}
Filter Sort