We're sorry but this page doesn't work properly without JavaScript enabled. Please enable it to continue.
Feedback
16 Ergebnisse
Vorschaubild
22:51
4Frosterus, Matias Mikael et al.
Controlled vocabularies and IT systems enabling their use have been in the forefront of library work for decades. In the National Library of Finland the national bibliography has been indexed using the YSA general thesaurus since the 1980s. A dedicated browser called VESA was developed in 1999 in order to eliminate the need to publish YSA as a printed document. In user surveys, VESA continually ranked as our most loved service. However, as years went on it became more difficult to integrate VESA’s old code to new environments. When the time came to renew VESA, library world was already buzzing with open linked data, semantic web etc. So it was decided that the new system should provide YSA and other vocabularies as open linked data with the ability to integrate the vocabularies to other systems using modern APIs. In 2013 work begun on the national ontology and thesaurus service Finto slated to replace VESA. Due to VESA being so well-liked, Finto was developed in deep collaboration with the users. Regular usability tests were conducted during the development and in all aspects and features care was taken in order to not put any extra burden on the daily tasks of the annotators. Finto provides the functionalities that VESA did, but also offers various new features and possibilities. An example of an auxiliary feature is the new suggestions system streamlining the process of gathering suggestions for new concepts into Finto vocabularies. Furthermore, the modular design of Finto also allowed us to utilize open APIs in other systems to, e.g., provide direct links to content annotated using a given concept in a vocabulary. We present the lessons learned during the development of a replacement for an extremely well-loved core service of a national library. A particular focus will be on the collaboration with the users during the development process and the migration.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
21:21
10Suominen, Osma et al.
Many libraries are experimenting with publishing their metadata as Linked Data in order to open up bibliographic silos, usually based on MARC records, and make them more interoperable, accessible and understandable to developers who are not intimately familiar with library data. The libraries who have published Linked Data have all used different data models for structuring their bibliographic data. Some are using a FRBR-based model where Works, Expressions and Manifestations are represented separately. Others have chosen basic Dublin Core, dumbing down their data into a lowest common denominator format. The proliferation of data models limits the reusability of bibliographic data. In effect, libraries have moved from MARC silos to Linked Data silos of incompatible data models. Data sets can be difficult to combine, for example when one data set is modelled around Works while another mixes Work-level metadata such as author and subject with Manifestation-level metadata such as publisher and physical form. Small modelling differences may be overcome by schema mappings, but it is not clear that interoperability has improved overall. We present a survey of published bibliographic Linked Data, the data models proposed for representing bibliographic data as RDF, and tools used for conversion from MARC. We also present efforts at the National Library of Finland to open up metadata, including the national bibliography Fennica, the national discography Viola and the article database Arto, as Linked Data while trying to learn from the examples of others.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
23:40
1Manguinhas, Hugo et al.
In the past years, the number of references to places, peoples, concepts and time in Europeana’s metadata has grown considerably and with it new challenges have arisen. These contextual entities are provided as references as part of the metadata delivered to Europeana or selected by Europeana for semantic enrichment or crowdsourcing. However their diversity in terms of semantic and multilingual coverage and their very variable quality make it difficult for Europeana to fully exploit this rich information. Pursuing its efforts towards the creation of a semantic network around cultural heritage objects and intending in this way to further enhance its data and retrieval across languages, Europeana is now working on a long term strategy for entities. The cornerstone of this strategy is a “semantic entity collection” that acts as a centralised point of reference and access to data about contextual entities, which is based on the cached and curated data from the wider Linked Open Data cloud. While Europeana will have to address the technical challenges of integration and representation of the various sources, it will also have to define a content and curation plan for its maintenance. This presentation will highlight the design principles of the Europeana Entity Collection and its challenges. We will detail our plans regarding its curation and maintenance while providing the first examples of its use in Europeana users' services. We will also reflect on how our goals can fit our partners' processes and how can organizations like national cultural heritage portals and smaller institutions contribute to (and benefit from) such a project as a network.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
24:17
2Deliot, Corine et al.
The British Library published the first Linked Open Data iteration of the British National Bibliography (BNB) in 2011. Since then it has continued to evolve with regular monthly updates, addition of new content (e.g. serials) and new links to external resources (e.g. International Standard Name Identifier (ISNI)). Data is available via deferenceable URIs, a SPARQL endpoint and RDF dataset dumps. There has been clear value to the Library in its linked data work, e.g. learning about RDF modelling and linked data. However, like many linked open data publishers, the Library has found it challenging to find out how the data has been used and by whom. Although basic usage data are captured in logs, there is currently no widely available tool to extract Linked Open Data insights. This makes it challenging to justify continued investment at a time of limited resourcing. This talk will report on collaboration between Fujitsu Laboratories Limited, Fujitsu Ireland and the British Library in the development of a Linked Open Data Analytics platform. The aim of the project was twofold: to examine Linked Open BNB usage and to potentially develop a tool of interest to the wider Linked Open Data community. We will describe the analytics platform and the functionality it provides as well as demonstrate what we found out about the usage of our data. Over the period under consideration (April 2014-April 2015) usage of the Linked Open BNB increased, and there was a discernible growth in the number of SPARQL queries relative to HTTP queries. Usage patterns were traced to the addition of new metadata elements or to linked data tuition sites or events.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
22:08
1Janssen, Olaf et al.
During the second World War some 1.300 illegal newspapers were issued by the Dutch resistance. Right after the war as many of these newspapers as possible were physically preserved by Dutch memory institutions. They were described in formal library catalogues that were digitized and brought online in the ‘90s. In 2010 the national collection of underground newspapers – some 200.000 pages – was full-text digitized in Delpher, the national aggregator for historical full-texts. Having created online metadata and full-texts for these publications, the third pillar ''context'' was still missing, making it hard for people to understand the historic background of the newspapers. We are currently running a project to tackle this contextual problem. We started by extracting contextual entries from a hard-copy standard work on Dutch illegal press and combined these with data from the library catalogue and Delpher into a central LOD triple store. We then created links between historically related newspapers and used Named Entity Recognition to find persons, organisations and places related to the newspapers. We further semantically enriched the data using DBPedia. Next, using an article template to ensure uniformity and consistency, we generated 1.300 Wikipedia article stubs from the database. Finally, we sought collaboration with the Dutch Wikipedia volunteer community to extend these stubs into full encyclopedic articles. In this way we can give every newspaper its own Wikipedia article, making these WW2 materials much more visible to the Dutch public, over 80% of whom uses Wikipedia. At the same time the triple store can serve as a source for alternative applications, like data visualizations. This will enable us to visualize connections and networks between underground newspapers, as they developed over time between 1940 and 1945. SWIB16 Conference, 28 - 30 November 2016, Bonn, Germany http://swib.org/swib16/ #swib16 Licence: CC-BY-SA https://creativecommons.org/licenses/by-sa/3.0/
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
22:02
4Bensmann, Felix et al.
The project linked.swissbib.ch aims to integrate the Swiss library metadata into the semantic web. A Linked Data infrastructure has been created to provide on the one hand a data service for other applications and on the other hand an improved interface for the end user (e.g. a searcher). The workflow for the development of this infrastructure involves basically five steps: (1) data modeling and transformation in RDF, (2) data indexing, (3) data interlinking and enrichment, (4) creation of a user interface and (5) creation of a RESTful API. The project team would like to highlight some challenges faced during these stages, and the means found to solve them. This includes for example the conception of various use cases of innovative semantic search functionalities to give specifications for data modelling, data enrichment and for the design of the search index. Data processing operations such as transformation and interlinking must be highly scalable, with the aim of an integration in the workflow of the already existing system. Wireframes have been made to realize early usability evaluations. Finally, negotiations have been undertaken with the various Swiss library networks to adopt a common open license for bibliographic data.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
21:35
9Alexiev, Vladimir
RDF is a graph data model, so the best way to understand RDF data schemas (ontologies, application profiles, RDF shapes) is with a diagram. Many RDF visualization tools exist, but they either focus on large graphs (where the details are not easily visible), or the visualization results are not satisfactory, or manual tweaking of the diagrams is required. We describe a tool *rdfpuml* that makes true diagrams directly from Turtle examples using PlantUML and GraphViz. Diagram readability is of prime concern, and rdfpuml introduces various diagram control mechanisms using triples in the puml: namespace. Special attention is paid to inlining and visualizing various Reification mechanisms (described with PRV). We give examples from Getty CONA, Getty Museum, AAC (mappings of museum data to CIDOC CRM), Multisensor (NIF and FrameNet), EHRI (Holocaust Research into Jewish social networks), Duraspace (Portland Common Data Model for holding metadata in institutional repositories), Video annotation. If the example instances include SQL queries and embedded field names, they can describe a mapping precisely. Another tool *rdf2rdb* generates R2RML transformations from such examples, saving about 15x in complexity.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
26:28
3Chapman, John W.
Continuing the longstanding research program by OCLC in the field of linked data, recent projects have focused on creating sets of entities of high interest for any organization wanting to utilize linked data paradigms. Through intensive mining and clustering of WorldCat bibliographic data, name and subject authority files, and other related data sets, OCLC has produced over 300 million entity representations. These clusters pull together and represent creative works, and persons related to those works. OCLC has engaged with a number of libraries and organizations to create and experiment with this data. A pilot project during October 2015-February 2016 to explore new methods of providing access to Person entities provided a number of new directions and insights. The core purpose of the work is to understand how these entities might best be leveraged to make library workflows more efficient, and to improve the quality of metadata produced in the library sector. This presentation will provide a background on data used in the project, as well as the development of services and APIs to provision the data. It will address challenges and opportunities in the area of creating and managing entities, and ways in which they could be improved and enriched over time.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
24:27
3Beck, Julia et al.
Imagine a theatre play. There are contributors such as the playwright, director, actors, etc. The play may have several performances with changing casts while actors may contribute to other plays. The play might be based on a drama which also has a screen adaption. All this is documented in manuscripts, photos, videos and other materials. The more relations you find among these performance-related objects, the more it emerges as a perfect use case for linked data. At the University Library Frankfurt am Main, the Specialised Information Service Performing Arts aggregates performing arts-related metadata of artefacts gathered by German-speaking cultural heritage institutions. It is funded by the German Research Foundation and aims to give researchers access to specialized information by providing a VuFind-based search portal that presents the metadata modeled as linked and open data. The Europeana Data Model (EDM) offers a universal and flexible metadata standard that is able to model the heterogeneous data about cultural heritage objects resulting from the data providers’ variety of data acquisition workflows. Being a common aggregation standard in digitization projects a comprehensive collection of mappings already exists. With the amount of delivered manuscript data in mind, the DM2E-extension of EDM was used and further extended by the ECLAP-namespace covering the specific properties for the performing arts domain. The presentation will show real life examples and focus on the modeling as linked data and the implementation within the VuFind framework.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
25:12
8Han, Myung-Ja K. et al.
Emblematica Online allows humanities scholars to seamlessly discover and link to items in a unique virtual emblem corpus distributed across six institutions in the US and Europe. The site supports multi-granular discovery of 1,400+ digitized emblem books and 25,000+ individual emblems from selective emblem books. To better integrate with related digital images and textual resources elsewhere, and to provide additional context for users, the site exploits linked open data (LOD) in two ways. First, as a producer of LOD, it publishes emblem and emblem book metadata as HTML+RDFa with schema.org semantics, making emblem resources more visible and useful in a linked open data context. Second, as a consumer of LOD, it enhances user experiences by utilizing LOD services and resources. For example, using the Iconclass LOD service, Emblematica Online supports multi-lingual browsing of the Iconclass vocabulary and connects users to digital sources elsewhere that share Iconclass descriptors. Also, it provides additional context about authors and contributors, including gender, nationality, and occupation, by reconciling names appearing in emblem metadata with LOD sources, such as the VIAF, DNB, and Wikipedia. This presentation discusses how Emblematica Online publishes its metadata as LOD and improves user experience using LOD sources as well as Emblem ontology development and plans for new services that allow possible reuse of Emblem LOD.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
20:12
1Schreur, Philip Evan
The Mellon Foundation recently approved a grant to Stanford University for a project called Linked Data for Production (LD4P). LD4P is a collaboration between six institutions (Columbia, Cornell, Harvard, Library of Congress, Princeton, and Stanford University) to begin the transition of technical services production workflows to ones based in Linked Open Data (LOD). This first phase of the transition focuses on the development of the ability to produce metadata as LOD communally, the enhancement of the BIBFRAME ontology to encompass multiple resource formats, and the engagement of the broader academic library community to ensure a sustainable and extensible environment. As its name implies, LD4P is focused on the immediate needs of metadata production such as ontology coverage and workflow transition. In parallel, Cornell also has been awarded a grant from the Mellon Foundation for Linked Data for Libraries-Labs (LD4L-Labs). LD4L-Labs will in turn focus on solutions that can be implemented in production at research libraries within the next three to five years. Their efforts will focus on the enhancement of linked data creation and editing tools, exploration of linked data relationships and analysis of the graph to directly improve discovery, BIBFRAME ontology development and piloting efforts in URI persistence, and metadata conversion tool development needed by LD4P and the broader library community. The presentation will focus on a brief description of the projects, how they interrelate, and what has been accomplished to date. Special emphasis will be given to extensibility and interactions with the broader LOD community.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
55:55
2Yelton, Andromeda
They say “build it, and they will come”, but what happens if you build it and they don’t? Getting people involved with open source projects takes more than good software or even a compelling use case: it’s about infrastructure, governance, and culture. This talk will cover research, current thinking, and real-world strategies for increasing and diversifying participation in open source projects.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
1:05:40
2Baker, Tom et al.
The publication of EPUB3 has been a major step forward digital publishing. Relying on Web Technologies like HTML, CSS, SVG, and others, EPUB3 offers a solid basis to publish not only digital books, but all sorts of digital publications in a portable, adaptable and accessible manner. However, it is possible to bring the publishing and the Web world even closer together, making the current format- and workflow-level separation between offline/portable and online (Web) document publishing eventually disappear. These should be merely two dynamic manifestations of the same publication: content authored with online use as the primary mode can easily be saved by the user for offline reading in portable document form. Content authored primarily for use as a portable document can be put online, without any need for refactoring the content. Essential features flow seamlessly between online and offline modes; examples include cross-references, user annotations, access to online databases, as well as licensing and rights management. W3C and IDPF have recently started to work towards this vision, exploring the technical challenges to make it a reality; this talk will give an overview on where we are.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
28:29
Király, Péter et al.
Europeana aggregates metadata from a wide variety of institutions, a significant proportion of which is of inconsistent or low quality. This low-quality metadata acts as a limiting factor for functionality, affecting e.g. information retrieval and usability. Europeana is accordingly implementing a user- and functionality-based framework for assessing and improving metadata quality. Currently, the metadata is being validated (against the EDM XML schema) prior to being loaded into the Europeana database. However, some technical choices with regard to the expressions of rules impose limitations on the constraints that can be checked. Furthermore, Europeana and its partners sense that more than simple validation is needed. Finer-grained indicators for the 'fitness for use' of metadata would be useful for Europeana and its data providers to detect and solve potential shortcomings in the data. Beginning 2016, Europeana created a Data Quality Committee to work on data quality issues and to propose recommendations for its data providers, seeking to employ new technology and innovate metadata-related processes. This presentation will describe more specifically the activities of the Committee with respect to data quality checks: - Definition of new data quality requirements and measurements, such as metadata completeness measures; - Assessment of (new) technologies for data validation and quantification, such as SHACL for defining data patterns; - Recommendations to data providers, and integration of the results into the Europeana data aggregation workflow.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
25:18
6Johnson, Christopher Hanna
"The IIIF Presentation API specifies a web service that returns JSON-LD structured documents that together describe the structure and layout of a digitized object or other collection of images and related content." IIIF website The dynamic serialization of IIIF JSON-LD structured manifests via SPARQL CONSTRUCT is an interesting possibility that has great potential for cross-domain discovery and rendering of digitized objects with variable criteria. I have explored this possibility by implementing a data model in the Fedora Commons Repository that matches the specifications of the IIIF Presentation API. Fedora has the facility to index objects via Apache Camel directly to a triplestore. With SPARQL CONSTRUCT, the triplestore can serialize normalized JSON-LD as a graph. The use of "ordered lists" (aka collections) is a fundamental component of JSON-LD and necessary feature of the IIIF manifest sequence which is represented in a canonical RDF graph as a cascade of blank nodes. In order to dynamically create the sequence with SPARQL requires that the data is modelled identically to the IIIF specification. This gist is a representation of a compacted and framed JSON-LD graph that was serialized from a SPARQL query of Fedora metadata. The ability to assemble parts of distinct, disparate and disassociated digital objects on demand in one cohesive presentation becomes a real possibility. For example, the "range" object is equivalent to a part of a sequence, like a chapter in a book. With SPARQL, it is possible to target ranges from different "editions" based on a metadata specification (i.e. a person, place, or date) and unify them in a manifest object which is then rendered by a client viewer like OpenSeadragon.
2016ZBW - Leibniz-Informationszentrum Wirtschaft et al.
Vorschaubild
25:44
Harlow, Christina et al.
One of the key goals of Cornell University Library (CUL) is to ensure preservation of the scholarly works being published by the Cornell faculty members and other researchers. VIVO is an open source and semantic technologies driven application that enables the preservation and open access of the scholarship across institutions. Driven by different needs, users look at VIVO implementation at Cornell from different viewpoints. The college requires the structure data for reporting needs. The library is interested in preservation of the scholarship data. University executives are interested in identifying the areas where they should invest in the forthcoming future. First, these viewpoints do not completely overlap with each. Second, current user interface represents the scholarship data in the list view format. Such representation of the scholarship data is not easy to use and consumable by the users. In this presentation, we present our ongoing work of integration of D3 visualizations into the VIVO pages. Such visualizations are constructed on the fly based on the underlying RDF data. A visualization-driven approach provides an efficient overview of the huge linked data network of interconnected resources. These visualizations are intuitive for the users to interact and offer the ability to visualize and navigate through the large linked data network. We discuss the performed (data) gap analysis as well as a few of the visualizations in detail and their integration into the VIVO framework.
2017ZBW - Leibniz-Informationszentrum Wirtschaft et al.