DBpedia – A large-scale, multilingual knowledge base extracted from Wikipedia
Issue title: Regular Papers
Article type: Research Article
Authors: Lehmann, Jens; | Isele, Robert | Jakob, Max | Jentzsch, Anja | Kontokostas, Dimitris | Mendes, Pablo N. | Hellmann, Sebastian | Morsey, Mohamed | van Kleef, Patrick | Auer, Sören; | Bizer, Christian
Affiliations: University of Leipzig, Institute of Computer Science, AKSW Group, Augustusplatz 10, D-04009 Leipzig, Germany. E-mail: {lehmann,kontokostas,hellmann,morsey,auer}@informatik.uni-leipzig.de | University of Mannheim, Research Group Data and Web Science, B6-26, D-68159 Mannheim, Germany. E-mail: chris@informatik.uni-mannheim.de | OpenLink Software, 10 Burlington Mall Road, Suite 265, Burlington, MA 01803, USA. E-mail: pkleef@openlinksw.com | Hasso-Plattner-Institute for IT-Systems Engineering, Prof.-Dr.-Helmert-Str. 2-3, D-14482 Potsdam, Germany. E-mail: mail@anjajentzsch.de | Neofonie GmbH, Robert-Koch-Platz 4, D-10115 Berlin, Germany. E-mail: max.jakob@neofonie.de | Kno.e.sis – Ohio Center of Excellence in Knowledge-enabled Computing, Wright State University, Dayton, USA. E-mail: pablo@knoesis.org | Brox IT-Solutions GmbH, An der Breiten Wiese 9, D-30625 Hannover, Germany. E-mail: mail@robertisele.com | Enterprise Information Systems, Universität Bonn & Fraunhofer IAIS, Römerstraße 164, D-53117 Bonn, Germany. E-mail: auer@cs.uni-bonn.de
Note: [] Corresponding author.
Abstract: The DBpedia community project extracts structured, multilingual knowledge from Wikipedia and makes it freely available on the Web using Semantic Web and Linked Data technologies. The project extracts knowledge from 111 different language editions of Wikipedia. The largest DBpedia knowledge base which is extracted from the English edition of Wikipedia consists of over 400 million facts that describe 3.7 million things. The DBpedia knowledge bases that are extracted from the other 110 Wikipedia editions together consist of 1.46 billion facts and describe 10 million additional things. The DBpedia project maps Wikipedia infoboxes from 27 different language editions to a single shared ontology consisting of 320 classes and 1,650 properties. The mappings are created via a world-wide crowd-sourcing effort and enable knowledge from the different Wikipedia editions to be combined. The project publishes releases of all DBpedia knowledge bases for download and provides SPARQL query access to 14 out of the 111 language editions via a global network of local DBpedia chapters. In addition to the regular releases, the project maintains a live knowledge base which is updated whenever a page in Wikipedia changes. DBpedia sets 27 million RDF links pointing into over 30 external data sources and thus enables data from these sources to be used together with DBpedia data. Several hundred data sets on the Web publish RDF links pointing to DBpedia themselves and make DBpedia one of the central interlinking hubs in the Linked Open Data (LOD) cloud. In this system report, we give an overview of the DBpedia community project, including its architecture, technical implementation, maintenance, internationalisation, usage statistics and applications.
Keywords: Knowledge extraction, Wikipedia, multilingual knowledge bases, Linked Data, RDF
DOI: 10.3233/SW-140134
Journal: Semantic Web, vol. 6, no. 2, pp. 167-195, 2015