The general vision for any organisation is to establish a “trusted information” environment which means high data quality and accurate information, reconciled where there is a risk or requirement. Trusted and relevant information means high data quality and relevance; one version of the truth and accurate information.
Data quality and reconciliation then becomes a key concern especially when integrating data from different source systems to a central Enterprise Data Warehouse. It is thus critical to ensure that the source data is of high quality and the extraction transformation and loading (ETL) process is fail proof before data is loaded into the target system. Back to reality, it’s seldom to have such a perfect scenario and there is no guarantee that the ETL process always brings accurate data.
Here are some pain points of what you may have experienced:
Experiencing deja vu :shock: ? To address these pain points we need to break them into two categories ie Data Quality and Data Reconciliation.
Data Quality
From a data quality strategy point of view the general recommendation is to have well defined data quality processes and to establish toolsets that deliver the following outcomes:
Data Quality - Key Approach
It is recommended to setup an Operation Change Management (OCM) initiative to identify and integrate data stewards into the data governance framework of your enterprise. This helps data stewards to be empowered to take action to improve the data quality in their respective data domains.
Some of the key responsibilities of data stewards should be to:
Data Quality - Key Challenges
Data Quality - Toolset
A data quality profiling and validation toolset (such as SAP Information Steward) for data stewards, business users and IT support teams is a way to measure and improve data quality across the enterprise. This will aid users to get insights into their data through dashboards, analyse business rules and improve the overall quality of the data.
(Figure shows a Data Quality Scorecard - SAP reference image)
Data Reconciliation
The approach to establishing trusted data would be to reconcile metrics thus ensuring the consistency of data that has been moved between source and target (BI) system. Data reconciliation for a data source metric allows for checking the integrity of the loaded data into BI, for example, comparing the totals of a metric in BI with the corresponding totals against the source system.
In general, it is recommended that a well defined reconciliation process and toolset be established that delivers the following outcomes:
Candidates for data reconciliation would be based on the factors below:
The ranking of the above factors would be evaluated against a score of low, medium and high.
Assessment questions should be provided to evaluate the risk factors for each metric during the requirements gathering workshops. If the result of the assessment for a metric is above the threshold level (medium) in all three areas then the metric is recommended as scope for reconciliation. The scope may be enhanced by IT Support team if they experience reconciliation issues in problem areas.
Data Reconciliation - Key Approach
Signal data reconciliation issues
Frequency of data reconciliation:
Level of data to reconcile:
Step 1 – Reconcile data at the highest level of data aggregation (eg company code level etc)
Step 2 – If step 1 does not provide the required information, then reconcile data at medium level aggregation (eg business unit etc) with additional selection criteria
Step 3 – If step 2 does not provide the required information, then reconcile data at low-level aggregation (eg document etc) with additional selection criteria
Data Reconciliation - Key Challenges
Implementation effort could depend on the complexity of the data, number of source systems and the number of metrics to reconcile. It is thus advised to prioritise the metrics based on business requirements.
Processing time largely depends on the volume of data that has to be read by the database and transferred; the prerequisite for using this process is that users have set meaningful filter selections in order to keep the volume of data that is to be transferred as small as possible.
Data quality at the source system is also a contributing factor for data reconciliation. If the quality of data is poor then data reconciliation will not add value.
Maintainability, supportability and developing reconciliations to problem areas are some key matters that should be address by the IT support team.
Data Reconciliation - Toolset
Key drivers for selecting an enterprise data reconciliation tool (such as SAP certified iRekon – Data Accuracy and Reconciliation) should be based on the above key approach, challenges and deliver the following features:
(Above figure are screen captures of iRekon – Data Accuracy and Reconciliation Dashboards)
Summary
In all, addressing both Data Quality and Data Reconciliation as above should enable the enterprise to achieve best in class data quality and most importantly improve business user’s confidence in data integrity and accuracy.
ℹ For more information about Data Quality and Data Reconciliation solutions you can get in touch at http://goo.gl/TWrEGK