Panel: Firms Face Increased Complexity When Measuring Data Quality

Following the financial crisis, it has become more complex to measure data quality, with firms implementing advanced metrics to avoid a dramatic rise in exceptions, according to a panel of speakers at the Paris Financial Information Summit in June.

Panelists said regulators are still focused on data quality and completeness of data, but ensuring data is good quality has become more challenging with high levels of volatility and different ways to define data quality. Paris-based Philippe Rozental

Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.

To access these options, along with all other subscription benefits, please contact [email protected] or view our subscription options here:

You are currently unable to copy this content. Please contact [email protected] to find out more.

To continue reading...

You need to sign in to use this feature. If you don’t have a WatersTechnology account, please register for a trial.

Sign in
You are currently on corporate access.

To use this feature you will need an individual account. If you have one already please sign in.

Sign in.

Alternatively you can request an individual account here: