Metadata, literally "data about data" -- specifically, descriptive metadata -- is structured data about anything that can be named, such as Web pages, books, journal articles, images, songs, products, processes, people (and their activities), research data, concepts, and services. Now a mainstream concept, metadata first trended in 1995, closely following World Wide Web in 1994. ("Big data" metadata about actions and transactions such as Facebook likes, phone calls, tweets, and the like, which trended in 2013, is out of scope for this brief introduction.)
Dublin Core™ metadata, or perhaps more accurately metadata "in the Dublin Core™ style", is metadata designed for interoperability on the basis of Semantic Web or Linked Data principles. Metadata in this style uses Uniform Resource Identifiers (URIs) as global identifiers both for the things described by the metadata and for the terms used to describe them (vocabularies). This style is distinguished by the application profile -- a specification detailing how well-known generic vocabularies such as the Dublin Core are used, constrained, or combined with more specialized vocabularies to meet the requirements of specific applications. Application profiles have been the focus of the Dublin Core™ community since they first trended in 2000.
The Dublin Core, a set of fifteen generic, widely used elements -- Creator, Contributor, Publisher, Title, Date, Language, Format, Subject, Description, Identifier, Relation, Source, Type, Coverage, and Rights -- was first drafted at a 1995 meeting in Dublin, Ohio, initially to facilitate information discovery on an explosively growing Web by embedding simple, card-catalog-like metadata in its pages. A diverse community of librarians, technologists, and researchers rallied to the idea, pursued and refined through a series of lively workshops and conferences, of achieving rough interoperability across languages and disciplines through a core of shared semantics. Successive developments in Web technology pulled this community in two directions:
Metadata based on record formats. Mainstream developers have used, and continue to use, vocabularies such as Dublin Core™ in the context of relational databases and repositories, many of which are based on XML, an Extensible Markup Language for specifying the contents of metadata records as structured documents. Implementers of record formats favor text values, closed-world quality control, top-down conformance, and reliance on well-understood, tried-and-true software solutions. Interoperability across applications is seen in terms of adherence to fixed formats such as the fifteen-element Simple Dublin Core™ and Qualified Dublin Core™ (2003), with several dozen additional DCMI metadata terms, to dozens of other formats published over the years. While the record-based approach may be relatively easy to deploy, interoperability across differently structured formats relies on ad-hoc "crosswalks" (mappings) that are hard to maintain and to use.
Metadata based on recombinant statements. Starting in the late 1990s, working groups of the World Wide Web Consortium pursued the vision of a web of data, or Semantic Web. This vision was enabled by Resource Description Framework (RDF) and by a global Domain Name System (DNS) that could resolve URIs to resources on the Web. The first W3C Recommendation for RDF in 1999 featured annotated examples of metadata using Dublin Core, which in 2000 became one of the first vocabularies to be published in RDF with persistent URIs. In the face of the messiness and complexities of the open Web, RDF implementers aim at achieving partial interoperability. In the RDF mindset, metadata consisted not of discrete, bounded records (documents) of a known structure, but of unbounded, schema-less graphs composed of atomic statements that could be recombined, or "mashed up", by merging multiple sources into the graph. In statement-based metadata, interoperability among multiple sources results from using, or mapping to, shared URIs, preferably from well-known vocabularies such as Dublin Core™.
Dublin Core-style application profiles. Where XML implementers saw application profiles as blueprints for creating validatable metadata records within a specific application, RDF implementers saw profiles as a basis for designing metadata that would compatibly fit into data graphs spanning multiple applications. To bridge this gap, DCMI's Singapore Framework (2007) the ideal application profile as the sum of several best-practice design components. At its core was the notion of a description as a set of statements about a single resource. Descriptions of multiple resources, such as Book and Author, could be bundled into a description set which, in turn, could either be stored directly as an RDF graph or encoded in a format designed to be convertible into RDF. A well-designed application profile would be based on available RDF vocabularies, well-articulated entity models, and explicitly defined functional requirements. Metadata in this style need not be based on the Dublin Core™ but can draw on a diversity of RDF vocabularies, such as Friend of a Friend (FOAF), the Bibliographic Ontology (BIBO), and Schema.org.
The uptake of metadata, and specifically of metadata based on RDF vocabularies, has been propelled by the evolution of technology:
DCMI's Linked Data Competency Index breaks down many of these areas into sets of skills and concepts ("competencies") usable by teachers, trainers, professors, or independent learners in designing courses or for self-directed study.
In his keynote talk at DC-2016, Bradley Allen discussed the role of metadata in the Second Machine Age. "While the user experience of discovery has come to be dominated by search engines such as Google," he noted, "metadata standards are pervasive in the infrastructure of content curation and management, and underpin search infrastructure". In his view, a single thread runs from the establishment of the Dublin Core™ through Open Linked Data to the emergence of Knowledge Graphs -- graph-structured databases extracted from content with the help of machine intelligence not only for helping people find, filter, and organize information, but also for constructing answers to questions. In his vision, the design of metadata should evolve in ways that can help machines read and learn from the Web and, in turn, help make its resources easier for both machines and people to discover and use.