Panel: Firms Face Increased Complexity When Measuring Data Quality

Following the financial crisis, it has become more complex to measure data quality, with firms implementing advanced metrics to avoid a dramatic rise in exceptions, according to a panel of speakers at the Paris Financial Information Summit in June.

Panelists said regulators are still focused on data quality and completeness of data, but ensuring data is good quality has become more challenging with high levels of volatility and different ways to define data quality. Paris-based Philippe Rozental

Only users who have a paid subscription or are part of a corporate subscription are able to print or copy content.

To access these options, along with all other subscription benefits, please contact info@waterstechnology.com or view our subscription options here: http://subscriptions.waterstechnology.com/subscribe

You are currently unable to copy this content. Please contact info@waterstechnology.com to find out more.

Sorry, our subscription options are not loading right now

Please try again later. Get in touch with our customer services team if this issue persists.

New to Waterstechnology? View our subscription options

Register for free

Access two articles, our IMD and Waters Wraps, plus a member newsletter. Find out more.

All fields are mandatory unless otherwise highlighted.

Back to basics: Taxonomies, lineage still stifle data efforts

Voice of the CDO: While data professionals are increasingly showing their value when it comes to analytics and AI adoption, their main job is still—crucially—getting a strong data foundation in place. That starts with taxonomies and lineage.

Most read articles loading...

You need to sign in to use this feature. If you don’t have a WatersTechnology account, please register for a trial.

Sign in
You are currently on corporate access.

To use this feature you will need an individual account. If you have one already please sign in.

Sign in.

Alternatively you can request an individual account here