A Process Pattern Model for Tackling and Improving Big Data Quality

Journal Article (2018)
Author(s)

A. Wahyudi (TU Delft - Information and Communication Technology)

George Kuk (Nottingham Trent University)

Marijn Janssen (TU Delft - Information and Communication Technology)

Research Group
Information and Communication Technology
Copyright
© 2018 A. Wahyudi, George Kuk, M.F.W.H.A. Janssen
DOI related publication
https://doi.org/10.1007/s10796-017-9822-7
More Info
expand_more
Publication Year
2018
Language
English
Copyright
© 2018 A. Wahyudi, George Kuk, M.F.W.H.A. Janssen
Research Group
Information and Communication Technology
Pages (from-to)
1-13
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Data seldom create value by themselves. They need to be linked and combined from multiple sources, which can often come with variable data quality. The task of improving data quality is a recurring challenge. In this paper, we use a case study of a large telecom company to develop a generic process pattern model for improving data quality. The process pattern model is defined as a proven series of activities, aimed at improving the data quality given a certain context, a particular objective, and a specific set of initial conditions. Four different patterns are derived to deal with the variations in data quality of datasets. Instead of having to find the way to improve the quality of big data for each situation, the process model provides data users with generic patterns, which can be used as a reference model to improve big data quality.