Hi,
I would like to share my personal opinion.
The German National Library (DNB) has released the GND in an RDF Turtle Dump under a CC0
license. More information:
http://www.dnb.de/DE/Service/DigitaleDienste/LinkedData/linkeddata_node.html
-> Download der Linked Data Dumps
What does that mean? Well, each of us can download a GND base set in RDF, put it into a search
engine, for example into Elasticsearch (it took only a few minutes for indexing 9.493.987 subject
URI-based documents consisting of a total of 97.267.642 triples), or into a triple store like 4store,
and start locally using GND as a source for authority control and for mixing up and building
mashups with other bibliographic and non-bibliographic data.
Setting up an OAI client - for example - completes the scenario. By fetching RDF/XML updates on a
regular basis from DNB you will always have the most recent authoritative data. This is not a vision,
it is reality.
A central triple store would be a major drawback. It does not scale. Each library has a lot of users
and applications, a single triple store would soon collapse. A good strategy is to put URI-based
authorities under an open data license and encouraging and enabling everybody around the world
to use it, too.
With RDF you can organize the data not only in records, but also in a graph of bibliographic
entities. Such a graph has a wealth of sub-graphs, attributes, and other properties. Well, if you
prefer, you can interpret the RDF graph of the GND as a record sequence of subject URIs as you
would do with MARC record collections, for example, to build searchable documents. But you are
no longer restricted to the record model. An RDF graph has an abstract semantic interpretation
and follows the rules of the W3C, describing statements about resources and facts (literals), having
rules in ontologies that are also part of the Semantic Web etc.
By using one of the many RDF serializations, bibliographic data can be packaged for transport
purposes. If you need to transport such packages over the wire, you can choose between formats
such as N-Triples, N3, Turtle, or RDF/XML. You are no longer restricted to the record-centered ISO
2709 format family with ancient character encodings, or XML wrappers around ISO 2709 that are
inheriting all the weaknesses from ISO 2709, since they are not aware of how to link to external
bibliographic entities or to reference them in a stable, reliable manner.
We all know, with the Internet, the massive number of mobile devices, and broadband connectivity,
transporting records in file packages from one place to another - like our elders need to do on
magnetic tapes due to the lack of affordable online transport capacity - is becoming more and
more the exception. The typical read access on catalog entities today is performed as lookups via a
growing number of web browsers and other web clients. These clients need to search documents,
traverse links and reference related information in many not foreseeable ways. So, methods for
bibliographic file packaging should be seamlessly connected to such popular use cases, i.e. how
the data is used later on the web.
Technologically, an RDF-based framework is a remarkable difference, it means that libraries are
joining Tim Berners-Lee's effort to interpret the World Wide Web as a global database where
everyone (even machines) can use (bibliographic) entities automatically just because they are part
of the Web, and not just being exposed to the Web.
Best regards,
Jörg
|