Flood of Factors
Q&A with Deloitte's Dilip Krishna about risk data aggregation
Does it make sense to divide up risk data and evaluate or inspect it before aggregating it?
Risk data usually originates elsewhere in the organization, as booked trades, originated and serviced loans, etc. It is enriched in a number of ways, most pertinently by adding risk metrics to it. To ensure high levels of risk data quality, it is essential to ensure the raw input itself has high fidelity. Additionally, high
quality requires the aggregation process to be free from corruption, so both of these are necessary conditions to ensure the ultimate accuracy of risk data.
How should risk data be divided and organized to those ends?
Risk data has several components. The base input is the current actual financial state of the organization as represented by trading positions and loan balances. Risk metrics also depend on other important information such as client, facility and collateral information. In addition, to develop models for risk management, it is critical to have a sufficiently long historical record of such data (e.g., five years of loan history). Finally, external data may also be required to supplement internal historical data (e.g. operational loss history data).
Are the stress-testing requirements of CCAR and BCBS 239 driving more attention to risk data aggregation and getting more done in that regard?
Stress-testing requirements are driving significant changes in risk data aggregation infrastructures. These requirements go well beyond generating risk reports, and demand that banks perform a meaningful analysis on both inputs and outputs of stress tests. In addition, there is a timeliness requirement that is hard to meet. These requirements are usually difficult for banks to meet with existing infrastructures, prompting their focus on risk data aggregation systems. Since BCBS 239 is consistent with these requirements but states them more explicitly, both requirements are together driving more coherence in risk data aggregation infrastructures.
Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.
To access these options, along with all other subscription benefits, please contact info@waterstechnology.com or view our subscription options here: https://subscriptions.waterstechnology.com/subscribe
You are currently unable to print this content. Please contact info@waterstechnology.com to find out more.
You are currently unable to copy this content. Please contact info@waterstechnology.com to find out more.
Copyright Infopro Digital Limited. All rights reserved.
As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (point 2.4), printing is limited to a single copy.
If you would like to purchase additional rights please email info@waterstechnology.com
Copyright Infopro Digital Limited. All rights reserved.
You may share this content using our article tools. As outlined in our terms and conditions, https://www.infopro-digital.com/terms-and-conditions/subscriptions/ (clause 2.4), an Authorised User may only make one copy of the materials for their own personal use. You must also comply with the restrictions in clause 2.5.
If you would like to purchase additional rights please email info@waterstechnology.com
More on Data Management
Deutsche Börse invests $200M in Kraken, DTCC advances cloud strategy, and more
A recap of this week’s major tech and data news in the capital markets.
Data industry spend hits $50B for first time in new report
A new product by BCG Expand will track market data vendor size and market share as it seeks to show data users where the market is heading.
TNS integrates Radianz, Exegy reduces latency, BondXN allies with BlackRock, and more
A recap of this week’s major tech and data news in the capital markets.
Re-engineering reconciliations: User-initiated AI cuts recs from days to minutes
Reconciliations have long been tied to batch scheduling. Prasanna Anandan explains how one bank broke down bottlenecks by embedding an AI-driven, user-initiated interface.
The public market data formula is coming to private markets
As interest in private markets grows, S&P Global, Bloomberg, and ICE are including increasingly valuable data in their offerings.
Tradefeedr pairs with BMLL to expand FX offering into equities, futures
Tradefeedr will also use BMLL’s historical data to help build out an LLM-powered chatbot.
Equity data plans eye Dec. 6 for overnight trading launch
The US SIPs are looking to launch near 24-hour operations as exchanges seek to extend their hours.
After the shuttering of Wilshire Indexes, the indexes space is a little tighter
The IMD Wrap: Max analyzes the winding up of Wilshire Indexes, a venture not yet three years old, and what the move means for the index industry and its consumers.