BCBS 239: 7 years of lessons learned

on Jun 29, 2020
BCBS 239 is summed up by the Basel Committee on Banking Supervision as: “ the right information needs to be presented to the right people at the right time”. In the global financial crisis of 2007/8 (remember that one?), regulators found some “globally systemically important” banks did not have a firm grip on their risk reporting and could not react quickly enough. The Principles for effective risk data aggregation and risk reporting were developed to try to ensure this did not happen again.

What do the principles require?

In many ways, BCBS 239 principles require application of ‘good practice’ data governance to ensure all the right data gets to decision-makers. The principles can apply to any organisation that depends on accurate reports for decision-making (are there any that don’t?). If applied, a bank should have better, faster decision-making.

With a 3 year deadline set in 2013 to implement BCBS 239, how are things looking now?

2 steps forward, 1 step back

According to the latest progress report (April 2020), “ In general, banks require more time to ensure that the Principles are effectively implemented.” So why is it taking so long? Why are banks with (supposedly) limitless resources not succeeding - and what can be learnt from their mistakes?

What is taking most of the time?

Perhaps surprisingly, over half the resources of a typical BCBS 239 project are taken up by determining and documenting data flows as data lineage. This is a critical part of the process – after all, if a bank can’t describe what data flows where, or document all of the sources of data used in a regulatory report, a regulator will not be satisfied. Seeing data in context is essential for a full high-level view.

The worst scenario for a bank is to spend vast quantities of money on a manual data lineage exercise, only for results to be discarded because they do not link together with other metadata and have gone stale. This is ‘dead’ data lineage.

The ideal scenario is for system owners to publish updated lineage information when it changes, and for that to be merged in and change-managed with all other metadata (like data catalog, data quality metrics, etc). This is ‘live’ data lineage.

Most banks are between these two extremities, able to extract lineage and metadata programmatically in some cases, and needing to merge this with data curated by system owners and subject matter experts. It is imperative to make this process as efficient as possible.

What takes up the rest of the time? Implementing data quality controls is the next largest resource demand. To fully benefit from data quality controls, issues in data need to be reflected in quality metrics of datasets – and consumers of affected data (not just immediate – all downstream consumers) need to be aware. With many quality assessment implementations for different architectures possible, the main aim for BCBS 239 should be to capture quality metrics for datasets, visualise data quality in context, and maintain it along with other metadata.

Data dictionaries and catalogs - Data dictionaries/glossaries/catalogs – providing business meaning for data and identifying critical data elements – is next for resources. Materiality (or impact of getting it wrong) of data can be used to prioritise governance. Data glossaries play a crucial role in completeness of reporting – classifying all relevant data for a report across multiple systems. Integrated with lineage flows and quality metrics, they help to ensure all data for a report is present with correct quality.

Harsh lessons in sustainable governance

The main mistake is to have a short-term view: ramp up a big push, hire some consultants to collect data and ‘get across the line’, without thinking of the bigger picture and a sustainable process. This approach ends up with wasted effort and the need to do it all again – properly this time.

To focus on the longer term, think about what you want to be able to do when you have all this regulation-required information. You want to see quality issues along the flows for data to your regulatory reports, and anywhere else. You want to find all the creditor data in your systems and make sure it is being reported. You want to help your team find the data they need to run new initiatives. You want to do ‘what-if’ experiments to see what would change if systems move to the cloud.

There are 2 key aspects to get right:
• A sustainable process to collect up-to-date metadata
• An integrated metadata platform

A sustainable process requires federating subject matter and system owners work across the organisation, and to automate where possible. Consider how business as usual will look.

Users need to easily apply their subject matter expertise and carry out manual tasks quickly and consistently, with intelligent suggestions for lineage and relationships from machine learning.

Connectors and open APIs, and integrations through expert partners, should be used to extract and load metadata and keep it up-to-date – including properties like dataset quality metrics.

An integrated platform for metadata treats metadata as a first-class asset, with all the change control and ownership that implies. The underlying architecture needs to be simple and adaptable, without artificial constraints. It needs to include lineage, quality metrics, glossaries, high level views, search, queries and reports – seamlessly.

Without integration, organisations face a synchronisation and maintenance nightmare and inability to see the end-to-end view.

Solidatus is designed by people who have been there and done it, and see the full picture. Why not use the regulatory imperative to deliver an asset of growing value to the whole organisation, where business leaders and regulators can see the full data picture?
Read More

Topics: Regulatory Compliance, Data Lineage, Metadata Management, Data Quality, Data Management, metadata governance, glossary, Data Dictionary, Business Glossary, Data Taxonomy, BCBS 239

How to turn a Regulatory Burden into an Operational Advantage

on Apr 30, 2020

On the 28th May 2018, the General Data Protection Regulation (GDPR) changed the data privacy landscape forever. What most organisations failed to grasp when the regulation came into force, was that GDPR should not have been viewed as a burden or to have been treated as a tick-box exercise. It was an opportunity to extract mandatory regulatory budget and use it to elevate and transform organisations’ data capabilities. Turning the 4% of the global turnover stick into the carrot to lead the organisations’ data journeys forward.

Read More

Topics: Regulatory Compliance, Data Lineage, GDPR, Data Protection, Regulation, PDPA, Data Sharing, Data Privacy, Regulatory Lineage, LGPD, CCPA

Are you at risk? - Solidatus for PDPA

on Apr 26, 2020

In my fourth post, I look at Singapore’s Personal Data Protection Act (PDPA). It broadly aligns with standards in other global legislation, albeit with some crucial differences such as control and access requests as opposed to the right to be forgotten, and Singapore’s limitations on transferring personal data overseas. Organisations face significant financial and reputational risk if they don't comply with PDPA and therefore need an effective data management solution to ensure they can track the personal data they have, what they are doing with it and how it moves through their systems. 

Read More

Topics: Regulatory Compliance, Singapore, Regulation, PDPA, Data Sharing, Data Privacy, Regulatory Lineage, Personal Data

IFRS 17 – Are You Prepared?

on Feb 19, 2020

As insurers and the insurance industry begin to implement the International Financial Reporting Standard 17 (IFRS 17), I wanted to take a moment to highlight the benefits that Solidatus can help organisations gain from this wholescale evolution of their digital landscape and the huge transformation it will undergo. 

Read More

Topics: Regulatory Compliance, Metadata Management, Data Management, Regulation, IFRS 17, Insurance, Insurance Contracts

EDM Council Goes Digital with Solidatus

on Sep 27, 2019

“Regulators and industry participants are leveraging data more comprehensively than ever, so it’s imperative that they have the necessary modeling tools to analyse it in order to make informed decisions,” Philip Dutton, Co-Founder of Solidatus.

Solidatus is committed to helping shape the future of data engineering, data management and data governance. Data engineering must evolve if it is to reach its full potential, much like other engineering disciplines have. With the exponential growth in data and the systems that hold it, data management is critical to ensuring standards are defined and structures implemented to support the ever-increasing rate of change. Without control and good governance, the already massively complex interconnectivity of systems will cause organisational paralysis without an urgent paradigm shift.

Read More

Topics: Regulatory Compliance, Data Lineage, Data Governance, Chief Data Officer, Partners, Data Management, EDMC, DCAM