Forchhammer, BenediktPapenbrock, ThorstenStening, ThomasViehmeier, SvenDraisbach, UweNaumann, FelixMarkl, VolkerSaake, GunterSattler, Kai-UweHackenbroich, GregorMitschang, BernhardHärder, TheoKöppen, Veit2018-10-242018-10-242013978-3-88579-608-4https://dl.gi.de/handle/20.500.12116/17320With the ever increasing volume of data and the ability to integrate different data sources, data quality problems abound. Duplicate detection, as an integral part of data cleansing, is essential in modern information systems. We present a complete duplicate detection workflow that utilizes the capabilities of modern graphics processing units (GPUs) to increase the efficiency of finding duplicates in very large datasets. Our solution covers several well-known algorithms for pair selection, attribute-wise similarity comparison, record-wise similarity aggregation, and clustering. We redesigned these algorithms to run memory-efficiently and in parallel on the GPU. Our experiments demonstrate that the GPU-based workflow is able to outperform a CPU-based implementation on large, real-world datasets. For instance, the GPU-based algorithm deduplicates a dataset with 1.8m entities 10 times faster than a common CPU-based algorithm using comparably priced hardware.enDuplicate detection on GPUsText/Conference Paper1617-5468