Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Acosta, Maribela; * | Zaveri, Amrapalib | Simperl, Elenac | Kontokostas, Dimitrisb | Flöck, Fabiand | Lehmann, Jensb
Affiliations: [a] Institute AIFB, Karlsruhe Institute of Technology, Germany. E-mail: maribel.acosta@kit.edu | [b] Institut für Informatik, AKSW, Universität Leipzig, Germany. E-mails: zaveri@informatik.uni-leipzig.de, kontokostas@informatik.uni-leipzig.de, lehmann@informatik.uni-leipzig.de | [c] Web Science and Internet Research Group, University of Southampton, United Kingdom. E-mail: e.simperl@soton.ac.uk | [d] Computational Social Science Group, GESIS – Leibniz Institute for the Social Sciences, Germany. E-mail: fabian.floeck@gesis.org
Correspondence: [*] Corresponding author. E-mail: maribel.acosta@kit.edu.
Abstract: In this paper we examine the use of crowdsourcing as a means to detect Linked Data quality problems that are difficult to uncover automatically. We base our approach on the analysis of the most common errors encountered in the DBpedia dataset, and a classification of these errors according to the extent to which they are likely to be amenable to crowdsourcing. We then propose and study different crowdsourcing approaches to identify these Linked Data quality issues, employing DBpedia as our use case: (i) a contest targeting the Linked Data expert community, and (ii) paid microtasks published on Amazon Mechanical Turk. We secondly focus on adapting the Find-Fix-Verify crowdsourcing pattern to exploit the strengths of experts and lay workers. By testing two distinct Find-Verify workflows (lay users only and experts verified by lay users) we reveal how to best combine different crowds’ complementary aptitudes in Linked Data quality issue detection. Empirical results show that a combination of the two styles of crowdsourcing is likely to achieve more effective results than each of them used in isolation, and that human computation is a promising and affordable way to enhance the quality of DBpedia.
Keywords: Quality assessment, quality issues, Linked Data, crowdsourcing, microtasks, experts
DOI: 10.3233/SW-160239
Journal: Semantic Web, vol. 9, no. 3, pp. 303-335, 2018
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
sales@iospress.com
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
info@iospress.nl
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office info@iospress.nl
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
china@iospress.cn
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
如果您在出版方面需要帮助或有任何建, 件至: editorial@iospress.nl