ALIADA, an Open Source Solution to Easily Publish Linked Data of Libraries and Museums
This is a modal window.
The media could not be loaded, either because the server or network failed or because the format is not supported.
Formal Metadata
Title |
| |
Title of Series | ||
Number of Parts | 16 | |
Author | ||
License | CC Attribution 3.0 Unported: You are free to use, adapt and copy, distribute and transmit the work or content in adapted or unchanged form for any legal purpose as long as the work is attributed to the author in the manner specified by the author or licensor. | |
Identifiers | 10.5446/47541 (DOI) | |
Publisher | ||
Release Date | ||
Language |
Content Metadata
Subject Area | ||
Genre | ||
Abstract |
|
3
8
10
00:00
Lecture/Conference
00:18
Computer animation
02:25
Computer animation
03:59
Computer animation
05:58
Computer animation
06:48
Computer animation
08:13
Computer animation
09:50
Computer animation
10:28
Computer animation
12:33
Computer animation
13:01
Computer animationDiagram
13:19
Computer animation
13:42
Computer animation
14:06
Computer animation
14:28
Computer animation
15:01
Computer animation
15:48
Computer animation
16:18
Computer animation
Transcript: English(auto-generated)
00:07
Okay. Morning to everybody. I am going to present the European Project Ariada on behalf of the consortium. Perhaps most of you know the project and this presentation is going to be a demo.
00:27
And I would like to highlight the technology used in the tool just to show you that Ariada follows the standards and also the trends you have been talking about.
00:49
Okay. Yes, there is a challenge for libraries and museums because they are information professionals and the challenge should be a good opportunity to be there in the semantic web.
01:12
And the link data technology is also an opportunity because the libraries and museums could have a
01:26
global pool of shared data that can be reused to describe resources and to reduce the cataloging effort. And the use of the web and the web-based identifiers will be used by catalogers also to enrich the descriptions and so on.
01:52
And the link data technology is more durable and robust than metadata formats used in libraries and museums.
02:02
Developers will also no longer have to work with library-specific data formats, Mark and Lido, and with linked open data libraries can increase their presence on the web. Because they are very important information providers and they should be there.
02:24
Okay. I will do the presentation from a librarian point of view because I am librarian. I'm not an engineer. And if you are interested in more technicalities, please contact the consortium and we will provide you a lot of information about the technology.
02:48
And how to start. Libraries and museums, first of all, they have to follow the standards. The standard conceptual models and the standard metadata schemes for biographical and museum information.
03:09
And they normally or usually do it, but now they have also to select an ontology.
03:20
They have to convert their metadata into RBF statements and they have to link their dataset to other datasets. And finally, they have to publish their data as a five-star linked open data.
03:41
Libraries and curators are experts in cataloguing and making accessible their resources, but they don't know about linked data technology. So they need an ally. This is the starting point of the Aliada project. Okay. The Aliada is ally in Spanish and is a tool to publish linked open data of libraries and museums.
04:19
It's an open source Java application to automatically publish as linked data and metadata created by a library or museum management system.
04:31
It supports the most used metadata types for describing biographical
04:42
objects and documents, authorities, museum objects and other information resources. It's compliant with the Mark XML, Lido and Dublin Core formats. And the conversion to RBF is done according to the Aliada ontology and a specific ontology for the project.
05:07
That is mainly based on FRVROO ontology, SCOS, WAF and two ontologies for time and geolocalization.
05:29
The linking to other datasets, Aliada has a predefined list of external datasets to select and to use during the conversion and during the linking.
05:45
Finally, Aliada will provide the possibility of publishing the DAMs, the URIs, and the Sparkling Point on Data Hub. And a short overview of the project. The project started in 2013 and the project has finished this October.
06:16
The partners are art museums, libraries, library system vendors, researchers on semantic web technology.
06:25
And we are from Spain, Italy and Hungary. The final release has been published in October and there is an open source community around the project.
06:42
Aliada is free software and you can download it and use it. This is the timeline of the project. As I have said before, we have published the final release in October.
07:01
And the main features or the main functions of the tool are the following. In the first prototype, we include the user interface module in Spanish and English. We implemented the validation of imported records, MARC and LIDO.
07:26
We created the mapping templates according to the FRDR 00 ontology, mainly. We created the art divisor, the conversion to RDF of the tool.
07:45
And we added some data sets, the data sets that can be queried with the SPARQL queries. And we implemented the linked data server creation, the SPARQL endpoint and the URIs difference.
08:07
And yes, we did the validation mainly. And for this second prototype, we have implemented the functionality for validating the DABLING core and for authorities.
08:28
We implemented also a validation in the art divisor. We added a kind of name entity recognition for some test free elements.
08:47
We added more external data sets to the list, the data sets that can be queried using an API. We added to the linking module the disambiguation function.
09:07
We did a lot of things. We implemented an advanced URI difference in the creation of our web page for the generated data set. And the publication module to publish the data set and the links on data hub.
09:31
And we translated the user interface to Italian and Hungarian languages. And finally, we integrated the Aliada tool with a library system using REST services.
09:50
This is more or less the architecture for Aliada's second prototype. As you can see in the slide, there are some modules.
10:08
The user interface module, the validation module, the divisor, the linking. That includes the linked data server and the discovery of links.
10:22
And finally, the publisher. I will highlight some features of the tool. For example, in the Aliada art divisor, we wanted to achieve these features.
10:56
The scalability of the module.
11:01
The art divisor is built as a RESTful application and is compliant with the Java web applications. The conversion templates and the mapping is configurable and extensible.
11:22
We can modify the existing mapping to adapt it to the changes in the models and in the cataloging techniques.
11:45
The usability is also an important aspect. It's possible to have the module as a standalone installation.
12:01
It's very easy to use and maintain. The conversion job is controlled by the conversion templates, as I said before. It's also easy to extend because of the same argument. Finally, we have added a validation to validate the conversion after starting the process.
12:32
The natural language processing has been done over some free test fields such as the Lido descriptive node value
12:48
and for Mark over some 500 tags.
13:01
This is an overview of how Aliada links the own dataset to the external dataset. For example, in the case of the external datasets that have a sparkling point. This is the case of the external datasets that don't provide a sparkling point.
13:30
Aliada includes also a link disambiguation to delete the generated links that are ambiguous about the URI.
13:49
We tried to create a cool URI. This is the format of the URI created by Aliada.
14:06
We include also an end convention to the URI structure to make it easier for a human reader.
14:21
I will show you all these things during the demo. Before the publication of the dataset, Aliada shows to the user this screen just to view what information will be published on the data hub.
14:45
For example, the description of the organization, the source, the license, the sparkling point, the vocabulary, the number of triples, the list of resources, and the subset used.
15:01
This is the aspect of the dataset generated by Aliada on the SICAM data hub platform. Finally, for the institutions that are interested in publishing their dataset in the linked data cloud,
15:26
they have to ask for a circle to the institution, to the load cloud net page. Just to say that Aliada allows the institution to reach the level three.
15:49
This is an example of how we integrated Aliada in an existing library management system.
16:00
We did that with a RESTful API. This is the aspect. Finally, I will show you the demo of Aliada in a few minutes. This is the main aspect of the tool.
16:24
The usability is a key aspect, a very important thing in the project. As you can see, the user has the possibility of the site, the configuration, to see the previous uploads.
16:43
They can go to the GitHub, to the wiki, to see the manual. In this screen, the user can decide what type of metadata they are going to convert.
17:06
Normally, the expected metadata has to be in MARC XML, Lido XML, and WCOR XML. After importing the metadata file, the user will run the module artificer to convert automatically the files to RDF.
17:32
The RDF statements are stored in a virtual data store, and they can be queried by the user through the Aliada screen.
17:47
The user can see the URI, and they can change the extension of the URI just to see the data as JSON, as Tardel, as HTML.
18:06
Also, the URI is ready to show the work, for example, in the opaque, in the catalog of the library of the museum, to have also a human-readable interface for the data.
18:27
All the process is very automatic and very easy to the final user. In the case of the project, the target user is a librarian, a non-expert user, so you can perform the process very easily.
18:45
You will get the URIs, and you will publish the URIs automatically on the data hub. This is the main feature of the tool, the simplicity of the process.
19:04
So, I think that I will end my presentation here, and if you are interested in using and testing the application, we are at your disposition for all that you think. Thank you very much.
19:23
Thank you, Cristina.