skip to Main Content

Data Quality

Last Updated:
Analyst Coverage:

Data quality is about ensuring that data is fit for purpose; that it is accurate, timely and complete enough relative to the use to which it is put. As a technology, data quality can either be applied after the fact or in a preventative manner.

Data quality products usually cover a range of areas including data profiling, data cleansing. merge/matching (discovering duplicated records) and data enrichment (adding, say, geocoding, postal codes or assorted business data like credit data to records).

Data quality also provides one of the sets of functionality required for data governance and master data management (MDM). Some data quality products have specific capabilities to support, for example, data stewards and/or facilities such as issue tracking.

Data quality products provide tools to perform various automated or semi-automated tasks that ensure that data is as accurate, up-to-date and complete as you need it to be. This may, of course, be different for different types of data: you want your corporate financial figures to be absolutely accurate but a margin of error is probably acceptable when it comes to mailing lists.

Data quality provides a range of functions. A relevant tool might simply alert you that there is an invalid postal code and then leave you to fix that; or the software, perhaps integrated with a relevant ERP or CRM product, might prevent the entry of an invalid post code altogether, prompting the user to re-enter that data. Some functions, such as adding a geocode to a location, can be completely automated while others will almost always require manual intervention to a degree. For example, when identifying potentially duplicate records the software can do this for you, and calculate the probability of a match, but it may require a business user or data steward to actually approve the match. Vendors are increasingly using artificial intelligence to help here by training expert systems to observe the actions of human domain experts, and then improve their suggestions based on these observations.

Poor data quality can be very costly indeed and there have been numerous studies examining, and proving, this point. The CFO should care. Conversely, good data quality ensures that your information about your customers is as complete and as accurate as it can be, which means, as we move more into a world of one-to-one marketing, that the CMO will also be interested in data quality. For companies that recognise that data is a corporate asset then data quality will be important for line of business managers and everybody up to the CEO level.

Further, data quality is of particular importance for compliance officers and data governance (which overlap) and CIOs. We discuss the relevance to compliance in the section on emerging trends but for CIOs data quality is important in a number of technical environments such as data migration, where poor data quality can adversely affect the success of the project and extend both costs and duration. This also applies to data warehousing where unsuccessful or delayed implementations have frequently been ascribed to poor data quality.

Data quality is what you might call a slow-burner. It has been an issue since the mid to late 90s and there are still companies that either refuse to recognise that they have an issue with data quality or don’t think it is worth the cost of fixing. This is gradually changing as people get better educated but the uptake of data quality technology remains on a slow growth path and we don’t expect that to change unless and until compliance requires it.

Traditionally, the adoption of data quality methods and tools has been a choice. The biggest emerging trend is that it is starting to become mandatory. Regulations such as Solvency II and MiFID II in Europe, and Dodd-Frank in the United States are starting to mandate that data is accurate, in which case good data quality will no longer be a choice. It is our belief that other regulations will increasingly focus on the accuracy of data in addition to the existing emphasis on process.

The data quality market is mature and there has been little change over the last several years, though some merges and acquisition activity. One notable feature has been that a number of smaller companies, such as Uniserv and Ataccama, have emerged as credible suppliers from non-English speaking environments.

In general, the market is split between those companies that focus purely on data quality and those that also offer either ETL (extract, transform and load) functionality or MDM (master data management) or both. Some of these “platforms” have been built from the ground up, such as that from SAS, while some others consist more of disparate bits that have been loosely bolted together. There is also a distinction between those that can provide specialist facilities for product matching (which is more complex than name and address matching), such as Oracle, and those that cannot.


  • Alex Solutions (logo)
  • ATACCAMA logo
  • DATA LADDER logo
  • DATACTICS logo
  • DQ GLOBAL logo
  • Experian logo
  • Global IDs logo
  • Pitney Bowes (logo)
  • PRECISELY logo
  • SAP (logo)
  • SYNITI logo

These organisations are also known to offer solutions:

  • Actian
  • Active Prime
  • BDQ
  • Clavis
  • FICO (InfoGlide)
  • IBM
  • Informatica
  • Innovative Software
  • InterSystems
  • iWay
  • Melissa Data
  • Microsoft
  • Oracle
  • Pervasive Software
  • SAS
  • Talend
  • Uniserv


IRI Voracity Healthcare InContext cover thumbnail

Data Preparation Challenges in Healthcare and Voracity from IRI

Voracity’s key strength is its ability to not just deliver broad data integration and transformation functionality, but to deliver it within a unified, centralised, integrated platform.
00002744 - IRI Voracity InContext cover thumbnail

Data Preparation Challenges in the Telecommunications Industry and Voracity from IRI

Voracity’s key strength is its ability to not just deliver broad data integration and transformation functionality, but to deliver it within a unified, centralised, integrated platform.
SAP (Data Quality) InBrief cover thumbnail

SAP Data Quality

Bloor Evaluates the latest SAP Data Quality offering, a cloud-based solution that will gradually supersede its previous data quality products.
DATA QUALITY 2022 Market Update cover thumbnail

Data Quality (2022)

Market Update of the Data Quality Enterprise Software Market in 2022, including data profiling, merge matching, data enrichment, data cleansing, monitoring.
DATA QUALITY Market Update (thumbnail)

Pure-play Data Quality

This Market Update focuses on vendors specialising in data quality that do not also provide data integration capabilities.
DATACTICS InBrief (thumbnail)


Datactics is a data quality platform oriented around self-service. It provides solutions for data quality, data matching, and single customer view.
DATA LADDER InBrief (thumbnail)

Data Ladder

Data Ladder provides a data quality management suite that includes all the sorts of capabilities that one would expect.
DQ GLOBAL InBrief (thumbnail)

DQ Global

DQ Global offers a number of different products, all of which are focused on data quality and all of which are targeted at customer data.
Back To Top