In computing, linked data (often capitalized as Linked Data) describes a method of publishing structured data so that it can be interlinked and become more useful through semantic queries. It builds upon standard Web technologies such as HTTP, RDF and URIs, but rather than using them to serve web pages for human readers, it extends them to share information in a way that can be read automatically by computers. This enables data from different sources to be connected and queried.
Use HTTP URIs so that these things can be looked up (interpreted, "dereferenced").
Provide useful information about what a name identifies when it's looked up, using open standards such as RDF, SPARQL, etc.
Refer to other things using their HTTP URI-based names when publishing data on the Web.
Tim Berners-Lee gave a presentation on linked data at the TED 2009 conference. In it, he restated the linked data principles as three "extremely simple" rules:
All kinds of conceptual things, they have names now that start with HTTP.
If I take one of these HTTP names and I look it up [..] I will get back some data in a standard format which is kind of useful data that somebody might like to know about that thing, about that event.
When I get back that information it's not just got somebody's height and weight and when they were born, it's got relationships. And when it has relationships, whenever it expresses a relationship then the other thing that it's related to is given one of those names that starts with HTTP.
The goal of the W3C Semantic Web Education and Outreach group's Linking Open Data community project is to extend the Web with a data commons by publishing various opendatasets as RDF on the Web and by setting RDF links between data items from different data sources. In October 2007, datasets consisted of over two billion RDF triples, which were interlinked by over two million RDF links. By September 2011 this had grown to 31 billion RDF triples, interlinked by around 504 million RDF links.
There are a number of European Union projects[definition needed] involving linked data. These include the linked open data around the clock (LATC) project, the PlanetData project, the DaPaaS (Data-and-Platform-as-a-Service) project, and the Linked Open Data 2 (LOD2) project. Data linking is one of the main goals of the EU Open Data Portal, which makes available thousands of datasets for anyone to reuse and link.
DBpedia – a dataset containing extracted data from Wikipedia; it contains about 3.4 million concepts described by 1 billion triples, including abstracts in 11 different languages
GeoNames provides RDF descriptions of more than 7,500,000 geographical features worldwide.
UMBEL – a lightweight reference structure of 20,000 subject concept classes and their relationships derived from OpenCyc, which can act as binding classes to external data; also has links to 1.5 million named entities from DBpedia and YAGO
FOAF – a dataset describing persons, their properties and relationships
Linked Data Is Merely More Data – Prateek Jain, Pascal Hitzler, Peter Z. Yeh, Kunal Verma, and Amit P. Sheth. In: Dan Brickley, Vinay K. Chaudhri, Harry Halpin, and Deborah McGuinness: Linked Data Meets Artificial Intelligence. Technical Report SS-10-07, AAAI Press, Menlo Park, California, 2010, pp. 82–86.
Ontology Alignment for Linked Open Data – Prateek Jain, Pascal Hitzler, Amit Sheth, Kunal Verma, Peter Z. Yeh. In proceedings of the 9th International Semantic Web Conference, ISWC 2010, Shanghai, China