Marja Haapalainen schrieb:
> We are therefore looking for other, library or consortia driven, initiatives regarding the collection and aggregation of metadata
In Germany there is the DFG Nationallizenzen funding programme (some
information in English is on
http://www.dfg.de/en/research_funding/programmes/infrastructure/lis/digital_information/library_licenses/index.html).
Though the primary goal of this programme is the acquisition of digital
scientific content, the Nationallizenzen license agreement requires
publishers to deliver metadata about the purchased content as well.
We (Verbundzentrale des GBV, a German library consortium) collect that
data, analyze, convert and aggregate it. So far we have about 22 million
records.
This data is available through some interfaces (SRU, Z39.50, unAPI) and
as download for entitled users (institutions participating in the DFG
Nationallizenzen). Some libraries use it in their metasearch portals,
few actually do something with the data downloads (eg. index it in their
own search interfaces, like ELIB Bremen:
http://elib.suub.uni-bremen.de/). And we use it ourselves in the
Nationallizenzen discovery interface
(http://finden.nationallizenzen.de/) we just started.
The whole data aggregation workflow really is a pain. Much effort goes
into analyzation and conversion of all the funny things you get from
publishers (few deliver bibliographic record formats like MARC21, but
most provide some homebrewn XMLish formats, Excel sheets, formatted text
files... worst thing was a >10000 pages Word document containing
"records" as formatted paragraphs, weird...).
Till
--
Till Kinstler
Verbundzentrale des Gemeinsamen Bibliotheksverbundes (VZG)
Platz der Göttinger Sieben 1, D 37073 Göttingen
kinstler@gbv.de, +49 (0) 551 39-13431, http://www.gbv.de
Received on Tue Feb 16 2010 - 04:59:50 EST