Tools by Functionality

In this section we categorize collected tools by their functionality. Moreover, this categorization will also identify missing functionality in currently available tools and services, which provide a gap analysis to develop or extend new functionality for specific usecases. There are 5 categories:

P roduce: tools fallen in this category have ability to generate new data.
Pu blis h: tools fallen in this category have ability to publish data on web platforms.
C onsu me: tools fallen in this category have ability to process the data for specific applications; e.g. data mining, data compression.
Provis ionin g: tools fallen in this category have ability to manipulate data such as data conversion and data extraction.
Data M anage ment: tools fallen in this category have ability to manage data in large-scale scenarios; e.g. storage, indexing, and querying.

back to Tool Catalogue page  ↑

Navigate between tabs to filter tools by Functionality:

geometry2rdf

  • Version:
    0.0.3
    Status:
    Completed
    Creation Date:
    01/02/2011
    Last Modification:
    01/2013
    Publisher:
    OEG
    Language:
    English
    Keywords:
    geometry rdf transformation
    Representation:
    Java, RDF
    License Document:
    N/A
    Documentation:
    Data Providers:
    It helps to transform geographical information contained in DDBBs to RDF format. This will increase the possible reuse of that information
    Data Consumers:
    It converts the information into a format, RDF, that it's easier to consult and more reusable that the information available in the DDBB

OKKAM - Enabling the Web of Entities

  • Description:

    The OKKAM project aims at enabling the Web of Entities, namely a virtual space where any collection of data and information about any type of entities (e.g. people, locations, organizations, events, products, ...) published on the Web can be integrated into a single virtual, decentralized, open knowledge base (like the Web did for hypertexts, readhere what Tim Berners-Lee says on this parallel).

    OKKAM will contribute to this vision by supporting the convergence towards the use of a single and globally unique identifier for any entity which is named on the Web. The intuition of the project is that the concrete realization of the Web of Entities requires that we enable tools and practices for cutting to the root the proliferation of unnecessary new identifierss for naming the entities which already have a public identifier (the OKKAM's razor). Therefore, OKKAM will make available to content creators, editors and developers a global infrastructure and a collection of new tools and plugins which support them to easily find public identifiers for the entities named in their contents/services, use them for creating annotations, build new network-based services which make essential use of these identifiers in an open environment (like the Web or large Intranets).

    Status:
    Completed
    Creation Date:
    01/01/2008
    Last Modification:
    06/2010
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    web of entities, entity name system, identifiers
    Homepage:
    Representation:
    Java, XML, RDF
    Contact Name:
    Paolo Bouquet
    Contact Email:
    bouquet@disi.unitn.it
    License Document:
    GPL
    App. Developers:
    App developers can use OKKAM as an API, just by selecting the top-1 result (if the confidence is above an acceptable threshold)
    Data Consumers:
    It helps in the entity reconciliation step (i.e., after Named Entity Recognition, OKKAM can help to pinpoint the right entity instance)

ODEMapster

  • Description:

    A tool to transform database records into RDF instances using standard (R2RML) mapping language. This tool also allows to query the database using graph query language (SPARQL)

    Version:
    2.9.13
    Status:
    UnderDevelopment
    Creation Date:
    01/01/2007
    Last Modification:
    03/2013
    Publisher:
    Universidad Politécnica de Madrid
    Language:
    English
    Keywords:
    rdb2rdf, r2rml, rdf, sql, rdb
    Representation:
    Java, RDF, SPARQL
    Contact Name:
    Freddy Priyatna
    Contact Email:
    fpriyatna@fi.upm.es
    Data Providers:
    This tool can transform the legacy data stored in databases as RDF instances
    Data Consumers:
    This tool can help the user to pose SPARQL query to legacy data stored in databases

LDIF – Linked Data Integration Framework

  • Description:

    The Web of Linked Data grows rapidly and contains data from a wide range of different domains, including life science data, geographic data, government data, library and media data, as well as cross-domain data sets such as DBpedia or Freebase. Linked Data applications that want to consume data from this global data space face the challenges that:

    Data sources use a wide range of different RDF vocabularies to represent data about the same type of entity;
    The same real-world entity, for instance a person or a place, is identified with different URIs within different data sources;
    Data about the same real-world entity coming from different sources may contain conflicting value. For example the single value attribute population for a specific country can have multiple, different values after merging data from different sources. This usage of different vocabularies as well as the usage of URI aliases makes it very cumbersome for an application developer to write SPARQL queries against Web data which originates from multiple sources. In order to ease using Web data in the application context, it is thus advisable to translate data to a single target vocabulary (vocabulary mapping) and to replace URI aliases with a single target URI on the client side (identity resolution), before starting to ask SPARQL queries against the data.

    Up-till-now, there have not been any integrated tools that help application developers with these tasks. With LDIF, we try to fill this gap and provide an open-source Linked Data Integration Framework that can be used by Linked Data applications to translate Web data and normalize URI while keeping track of data provenance.

    Version:
    0.5.1
    Status:
    Completed
    Creation Date:
    29/06/2011
    Last Modification:
    02/2013
    Publisher:
    Universität Mannheim and MES Semantics
    Language:
    English
    Keywords:
    linked data, data integration, schema mapping, identity resolution, data quality assessment, data fusion
    Representation:
    JAVA
    Contact Name:
    Christian Bizer
    Contact Email:
    chris@bizer.de
    License Document:
    Apache
    Documentation:
    App. Developers:
    LDIF covers all necessary steps of the data integration process. It can be used to access and integrate Linked Data from various sources while keeping track of data provenance
    Data Providers:
    The tool targets data consumers
    Data Consumers:
    LDIF covers all necessary steps of the data integration process. It can be used to access and integrate Linked Data from various sources while keeping track of data provenance

Global Sensor Networks

  • Description:

    GSN is a Java environment that runs on one or more computers composing the backbone of the acquisition network. A set of wrappers allow to feed live data into the system. Then, the data streams are processed according to XML specification files. The system is built upon a concept of sensors (real sensors or virtual sensors, that is a new data source created from live data) that are connected together in order to built the required processing path. For example, one can imagine an anemometer that would sent its data into GSN through a wrapper (various wrappers are already available and writing new ones is quick), then that data stream could be sent to an averaging mote, the output of this mote could then be splited and sent for one part to a database for recording and to a web site for displaying the average measured wind in real time. All of this example could be done by editing only a few XML files in order to connect the various motes together.

    Version:
    1.1.2
    Version Description:
    New sensorscope listener server fixing problem of missing data Added count-based checkpoints to the CSV wrapper, logging line counts instead of latest timestamp Added new sensor types to sensorscope listener (NO2, CO, CO2, snow height, dendrometer, new temperature and air pressure sensors, data logger voltage, and GPS coordinates) Added /dynamicgeodata servlet, which allows making basic spatial queries for moving sensors Added new REST API for accessing data (beta) Fixed leak in unused connections (StorageManager pool configuration)
    Version Properties:
    Support for PostGIS for geospatial queries
    Status:
    UnderDevelopment
    Creation Date:
    11/11/2011
    Last Modification:
    02/2013
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    data stream, sensor network, distributed system
    Representation:
    Java XML XMLSchema RDF C++
    Contact Name:
    Sofiane Sarni
    Contact Email:
    sofiane.sarni@epfl.ch
    License Document:
    GPL
    License Description:
    GSN is distributed under GPL for the general public, in case GPL doesn't suite your needs, please contact us.
    Distribution Format:
    jar

OKKAM - Enabling the Web of Entities

  • Description:

    The OKKAM project aims at enabling the Web of Entities, namely a virtual space where any collection of data and information about any type of entities (e.g. people, locations, organizations, events, products, ...) published on the Web can be integrated into a single virtual, decentralized, open knowledge base (like the Web did for hypertexts, readhere what Tim Berners-Lee says on this parallel).

    OKKAM will contribute to this vision by supporting the convergence towards the use of a single and globally unique identifier for any entity which is named on the Web. The intuition of the project is that the concrete realization of the Web of Entities requires that we enable tools and practices for cutting to the root the proliferation of unnecessary new identifierss for naming the entities which already have a public identifier (the OKKAM's razor). Therefore, OKKAM will make available to content creators, editors and developers a global infrastructure and a collection of new tools and plugins which support them to easily find public identifiers for the entities named in their contents/services, use them for creating annotations, build new network-based services which make essential use of these identifiers in an open environment (like the Web or large Intranets).

    Status:
    Completed
    Creation Date:
    01/01/2008
    Last Modification:
    06/2010
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    web of entities, entity name system, identifiers
    Homepage:
    Representation:
    Java, XML, RDF
    Contact Name:
    Paolo Bouquet
    Contact Email:
    bouquet@disi.unitn.it
    License Document:
    GPL
    App. Developers:
    App developers can use OKKAM as an API, just by selecting the top-1 result (if the confidence is above an acceptable threshold)
    Data Consumers:
    It helps in the entity reconciliation step (i.e., after Named Entity Recognition, OKKAM can help to pinpoint the right entity instance)

Videk

  • Description:

    Videk currently uses four sources of sensor and linked data and relies on StreamSense engine for storage and processing. In this paper we present the architecture and current implementation of Videk as well as the lines along which we plan to extend and improve it.

    Status:
    UnderDevelopment
    Creation Date:
    01/01/2011
    Last Modification:
    01/2013
    Publisher:
    Unknown
    Language:
    English
    Keywords:
    Mash-up, sensors, web of things, real-time, data mining, semantic web.
    Representation:
    XML
    Contact Name:
    Carolina Fortuna
    Contact Email:
    carolina.fortuna@ijs.si
    License Document:
    Unknown
    App. Developers:
    by providing sensor data and meta-data
    Data Consumers:
    by providing sensor data and meta-data
    Market Analysts:
    by providing sensor data and meta-data
    Tech. Providers:
    by providing sensor data and meta-data
    Researchers:
    by providing sensor data and meta-data

Josef Stefan Institute's newsfeed

  • Description:

    Newsfeed provides a clean, continuous, real-time aggregated stream of semantically enriched news articles from RSS-enabled sites across the world.

    The pipeline performs the following main steps:
    1) Periodically crawl a list of RSS feeds and a subset of Google News and obtain links to news articles
    2) Download the articles, taking care not to overload any of the hosting servers
    3) Parse each article to obtain
    3a) Potential new RSS sources mentioned in the HTML, to be used in step (1)
    3b) Cleartext version of the article body
    4) Process articles with Enrycher (English and Slovene only)
    5) Expose two streams of news articles (cleartext and Enrycher-processed) to end users.

    Status:
    UnderDevelopment
    Creation Date:
    01/01/2012
    Last Modification:
    01/2013
    Publisher:
    JSI
    Language:
    English
    Keywords:
    text, news, data stream, enrichment
    Representation:
    XML
    Contact Name:
    Mitja Trampus
    Contact Email:
    mitja.trampus@ijs.si
    License Document:
    LGPL
    Documentation:
    App. Developers:
    They can help app developers that need enriched multilingual streams of news
    Data Consumers:
    Newsfeed produces a clean, continuous, real-time aggregated stream of semantically enriched news articles from RSS-enabled sites across the world

Datalift Platform

  • Description:

    The Datalift web platform is a tool suite for converting, structured data sources and publishing them as linked data on the web.

    Status:
    UnderDevelopment
    Creation Date:
    01/10/2010
    Last Modification:
    03/2013
    Publisher:
    INRIA
    Language:
    English, French
    Keywords:
    linked-data, structured data, interlinking, LOV, vocabulary mapping, ontologies, sql, shape, statistics, sdmx, datacube, CSV, SPARQL, JSON, Java, javascript, XML, RDF
    Representation:
    JAVA, XML, RDF, SPARQL, C++, OWL
    Contact Name:
    François Scharffe
    Contact Email:
    francois.scharffe@lirmm.fr
    License Document:
    Apache
    Documentation:
    App. Developers:
    By following web standards, Datalift offers languages and protocols shared by everyone.
    Data Providers:
    Datalift helps to convert structured data to RDF, to select appropriate vocabularies to describe the data, and to interlink these data with other datasets.
    Data Consumers:
    Datalift provides standard HTTP and SPARQL access to data in a variety of formats (HTML, JSON, RDF/n3, RDF/XML). It also has a graphical user interface including a graphical SPARQL query builder.

Global Sensor Networks

  • Description:

    GSN is a Java environment that runs on one or more computers composing the backbone of the acquisition network. A set of wrappers allow to feed live data into the system. Then, the data streams are processed according to XML specification files. The system is built upon a concept of sensors (real sensors or virtual sensors, that is a new data source created from live data) that are connected together in order to built the required processing path. For example, one can imagine an anemometer that would sent its data into GSN through a wrapper (various wrappers are already available and writing new ones is quick), then that data stream could be sent to an averaging mote, the output of this mote could then be splited and sent for one part to a database for recording and to a web site for displaying the average measured wind in real time. All of this example could be done by editing only a few XML files in order to connect the various motes together.

    Version:
    1.1.2
    Version Description:
    New sensorscope listener server fixing problem of missing data Added count-based checkpoints to the CSV wrapper, logging line counts instead of latest timestamp Added new sensor types to sensorscope listener (NO2, CO, CO2, snow height, dendrometer, new temperature and air pressure sensors, data logger voltage, and GPS coordinates) Added /dynamicgeodata servlet, which allows making basic spatial queries for moving sensors Added new REST API for accessing data (beta) Fixed leak in unused connections (StorageManager pool configuration)
    Version Properties:
    Support for PostGIS for geospatial queries
    Status:
    UnderDevelopment
    Creation Date:
    11/11/2011
    Last Modification:
    02/2013
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    data stream, sensor network, distributed system
    Representation:
    Java XML XMLSchema RDF C++
    Contact Name:
    Sofiane Sarni
    Contact Email:
    sofiane.sarni@epfl.ch
    License Document:
    GPL
    License Description:
    GSN is distributed under GPL for the general public, in case GPL doesn't suite your needs, please contact us.
    Distribution Format:
    jar

MonetDB

  • Description:

    A relational database management system for high-performance data warehouses for business intelligence and eScience. Since a few years column store technology as pioneered in MonetDB has found its way into the product offerings of all major commercial database vendors. The market for applications empowered by these techniques provide ample space for further innovation, e.g. as demonstrated by our ongoing projects. At the same time, the landscape for major innovations remain wide open. A peek preview is given in the award winning paper titled: The Researcher’s Guide to the Data Deluge: Querying a Scientific Database in Just a Few Seconds. MonetDB is actively used in our research and real life applications. Nightly builds and regression testing ensure its quality, bug tracking helps to collect experiences and feature requests. Browsing the source code repository is supported by the Mercurial web frontend. Contributions ranging from bug reports, cross-platform issues, patches and features are highly appreciated.

    MonetDB is actively used in our research and real life applications. Nightly builds and regression testing ensure its quality, bug tracking helps to collect experiences and feature requests. Browsing the source code repository is supported by the Mercurial web frontend. Contributions ranging from bug reports, cross-platform issues, patches and features are highly appreciated

    Status:
    UnderDevelopment
    Publisher:
    CWI
    Language:
    English
    Keywords:
    column-stored, XML, data management
    Representation:
    Java, XML, RDF, SQL, SPARQL
    License Document:
    MonetDB Public License

HDT

  • HDT
    Description:

    HDT (Header, Dictionary, Triples) is a compact data structure and binary serialization format for RDF that keeps big datasets compressed to save space while maintaining search and browse operations without prior decompression. This makes it an ideal format for storing and sharing RDF datasets on the Web. Some facts about HDT:

    - The size of the files is smaller than other RDF serialization formats. This means less bandwidth costs for the provider, but also less waiting time for the consumers to download.
    - The HDT file is already indexed. The users of RDF dumps want to do something useful with the data. By using HDT, they download the file and start browsing/querying in minutes, instead of wasting time using parsing and indexing tools that are difficult to setup and tune.
    - High performance querying. Usually the bottleneck of databases is slow disk access. The internal compression techniques of HDT allow that most part of the data (or even the whole dataset) can be kept in main memory, which is several orders of magnitude faster than disks.
    - Highly concurrent. HDT is read-only, so it can dispatch many queries per second using multiple threads.
    - The format is open and is acknowledged as W3C HDT Member Submission. This ensures that anyone on the Web can generate and consume files, or even write their own implementation.
    - The libraries are open source (LGPL). You can adapt the libraries to your needs, and the community can spot and fix issues

    Version:
    1
    Version Description:
    Implementation of the RDF/HDT W3C Member Submission providing triple pattern resolution and SPARQL evaluation
    Status:
    UnderDevelopment
    Creation Date:
    30/01/2013
    Last Modification:
    03/2013
    Publisher:
    DataWeb Research (University of Valladolid)
    Language:
    English
    Keywords:
    Binary RDF, compression, in-memory SPARQL, fast exchange
    Representation:
    Java, RDF, SPARQL, C++
    Contact Name:
    Miguel A. Martínez-Prieto
    Contact Email:
    migumar2@infor.uva.es
    License Document:
    GNU LESSER GENERAL PUBLIC LICENSE
    License Description:
    Open source code. Allows linking to the libraries from both open source and commercial software packages. However, the source code of any modification of the library itself must be made public
    Distribution Format:
    Source Code and Executables for several platforms (Windows, Linux and Mac, 32 and 64 bits)
    App. Developers:
    DF/HDT can be directly plugged into applications that generate or consume RDF. It provides a simple interface for generating HDT blocks with the application data, and an API for browsing, querying, and serializing the triples inside these blocks
    Data Providers:
    The main advantage for data providers is the compactness of the generated files. They are smaller, therefore minimizing disk space for storage, and network bandwidth for transmission. Also the generation process is highly efficient
    Data Consumers:
    RDF/HDT files are ready to be used, so data consumers can directly access them for browsing and querying without the need of heavy parsing and indexing. In addition, RDF/HDT compactness alleviates browsing and querying in limited computational config-s
    Market Analysts:
    A good analogy to show the benefits of RDF/HDT can be the success of other ad-hoc compression formats, such as JPEG, MPEG2 and MP3.
    Tech. Adopters:
    The main advantage of RDF/HDT is that it takes the most out of the available resources: both transfer and querying are very efficient. This means that the users can process higher amounts of data using the same hardware.
    Tech. Providers:
    Adopting RDF/HDT can be straightforward if the storage layer of the framework/application has already been abstracted (e.g. Using Jena API). In this case they just need to select the HDT as backing store and access it as usual
    Researchers:
    RDF/HDT can be an interesting format for sharing research data, such as biological datasets or social media interactions. It allows researchers to download it and directly start querying to extract valuable knowledge.

CKAN Extractor

  • Description:

    CKAN plugin for the automatic extraction of data sources. It enables administrator to upload transformation plugins written in Python which extract data from non-estructured sources. The extension provides a common framework for transformation development and periodic execution of tasks using celery

    Status:
    UnderDevelopment
    Creation Date:
    27/02/2013
    Last Modification:
    03/2013
    Publisher:
    University of Deusto
    Language:
    English
    Keywords:
    data extraction
    Representation:
    Java, RDF
    Contact Name:
    Unai Aguilera
    Contact Email:
    unai.aguilera@deusto.es
    License Document:
    AGPL

OOPS! - OntOlogy Pitfall Scanner!

  • Description:

    OOPS! is a web-based tool, independent of any ontology development environment, for detecting potential pitfalls that could lead to modelling errors. This tool is intended to help ontology developers during the ontology validation activity, which can be divided into diagnosis and repair. Currently, OOPS! provides mechanisms to automatically detect a number of pitfalls, thus helps developers in the diagnosis activity.

    Version:
    1.0.0
    Version Properties:
    RESTFul web service support. Bugs fixed.
    Status:
    UnderDevelopment
    Creation Date:
    14/11/2011
    Last Modification:
    03/2013
    Publisher:
    Ontology Engineering Group
    Language:
    English
    Keywords:
    ontology, ontology development, pitfall detection, ontology evaluation
    Representation:
    Java, RDF, OWL
    Contact Name:
    María Poveda-Villalon
    Contact Email:
    mpoveda@fi.upm.es
    License Document:
    GPLv.3
    App. Developers:
    Helping to find potential modelling problems in ontologies
    Tech. Providers:
    Integrating OOPS! pitfalls detection mechanism within their applications
    Researchers:
    It help them to evaluate ontologies they might create for research purposes

SPARQL endpoint analyzer and metadata generator for CKAN - ckanext-metadata

SPARQL Extension for CKAN - ckanext-sparql

  • Description:

    This CKAN plugin offers two main functionalities:
    - It allows the configuration of a SPARQL endpoint for the whole CKAN platform in which metadata about every dataset in the platform can be queried.
    - It allows dataset editors to configure a RDF store and manage the RDF data of the dataset directly from CKAN, enabling at the same time a SPARQL endpoint for querying this data

    Version:
    0.1.0
    Status:
    UnderDevelopment
    Creation Date:
    04/03/2013
    Last Modification:
    03/2013
    Publisher:
    MORElab
    Language:
    English
    Keywords:
    ckan, sparql, rdf store
    Representation:
    Java, RDF, SPARQL
    Contact Name:
    Jon Lazaro
    Contact Email:
    jlazaro@deusto.es
    License Document:
    AGPL
    App. Developers:
    This tool allows application developers to query data from any CKAN instance through SPARQL queries
    Data Providers:
    his tool allows data providers (using CKAN) to configure a RDF store and manage the RDF data of the dataset directly from CKAN, enabling at the same time a SPARQL endpoint for querying this data
    Data Consumers:
    This tool allows data consumers to query data from any CKAN instance through SPARQL queries
    Researchers:
    This tool allows researchers to query data from any CKAN instance through SPARQL queries

SILK extension for CKAN

  • Description:

    An extension for interlinking datasets uploaded on CKAN using SILK Link Discovery Framework.

    Version:
    beta
    Version Properties:
    Basic funtionalities and interface is done. Now, we are solving bugs and working on more complex interlinking rules
    Status:
    UnderDevelopment
    Creation Date:
    21/11/2012
    Last Modification:
    02/2012
    Publisher:
    DeustoTech - Internet
    Language:
    English
    Keywords:
    linked data, interlinking, CKAN, semantic web
    Representation:
    Java, RDF
    Contact Name:
    Mikel Emaldi
    Contact Email:
    m.emaldi@deusto.es
    License Document:
    Apache2.0
    Data Providers:
    This tool could help the Data Providers easing the process of using SILK to interlinking their datasets with others in the same CKAN instance

OKKAM - Enabling the Web of Entities

  • Description:

    The OKKAM project aims at enabling the Web of Entities, namely a virtual space where any collection of data and information about any type of entities (e.g. people, locations, organizations, events, products, ...) published on the Web can be integrated into a single virtual, decentralized, open knowledge base (like the Web did for hypertexts, readhere what Tim Berners-Lee says on this parallel).

    OKKAM will contribute to this vision by supporting the convergence towards the use of a single and globally unique identifier for any entity which is named on the Web. The intuition of the project is that the concrete realization of the Web of Entities requires that we enable tools and practices for cutting to the root the proliferation of unnecessary new identifierss for naming the entities which already have a public identifier (the OKKAM's razor). Therefore, OKKAM will make available to content creators, editors and developers a global infrastructure and a collection of new tools and plugins which support them to easily find public identifiers for the entities named in their contents/services, use them for creating annotations, build new network-based services which make essential use of these identifiers in an open environment (like the Web or large Intranets).

    Status:
    Completed
    Creation Date:
    01/01/2008
    Last Modification:
    06/2010
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    web of entities, entity name system, identifiers
    Homepage:
    Representation:
    Java, XML, RDF
    Contact Name:
    Paolo Bouquet
    Contact Email:
    bouquet@disi.unitn.it
    License Document:
    GPL
    App. Developers:
    App developers can use OKKAM as an API, just by selecting the top-1 result (if the confidence is above an acceptable threshold)
    Data Consumers:
    It helps in the entity reconciliation step (i.e., after Named Entity Recognition, OKKAM can help to pinpoint the right entity instance)

Linked Open Data Miner

  • Description:

    LOD Miner is a system for recommending missing properties for a given object. The input to the system is a set of objects or entities, each described with a set of properties. The system then tries to find the missing properties for a specified object based on similarity to other objects. Typical examples of datasets are RDF graphs from Linked Open Data (LOD).

    Status:
    UnderDevelopment
    Creation Date:
    01/01/2012
    Last Modification:
    01/2013
    Publisher:
    JSI
    Language:
    English
    Keywords:
    linked open data, graph, missing properties, prediction
    Representation:
    Java
    Contact Name:
    Klemen Simonic
    Contact Email:
    klemen.simonic@ijs.si
    License Document:
    N/A
    Documentation:

Yet-Another-SPARQL-GUI

  • Description:

    YASGUI is a web-based SPARQL client that can be used to query both remote and local endpoints. It integrates linked data services and web APIs to offer features such as autocompletion and endpoint lookup. It supports query retention – query texts persist across sessions – and query permalinks, as well as syntax checking and highlighting

    Version:
    13.03b
    Status:
    Completed
    Creation Date:
    01/07/2012
    Last Modification:
    03/2013
    Language:
    English
    Keywords:
    SPARQL, Semantic Web, Endpoints
    Representation:
    Java
    Contact Name:
    Laurens Rietveld
    Contact Email:
    laurens.rietveld@vu.nl
    License Document:
    MIT
    Data Providers:
    Enable users to navigate your endpoint in a more efficient/attractive, by using YASGUI as your main endpoint SPARQL client
    Data Consumers:
    Explore the Semantic Web and query any remote endpoint by using this robust, user friendly SPARQL client
    Researchers:
    YASGUI helps Web savvy developers as well: trying and testing SPARQL queries is often a cumbersome and painful experience: all who know the RDF namespace by heart raise their hands now!

D2RQ Plattform - Accessing Relational Databases as Virtual RDF Graphs

  • Description:

    The D2RQ Platform is a system for accessing relational databases as virtual, read-only RDF graphs. It offers RDF-based access to the content of relational databases without having to replicate it into an RDF store. Using D2RQ you can:
    1. query a non-RDF database using SPARQL
    2. access the content of the database as Linked Data over the Web
    3. create custom dumps of the database in RDF formats for loading into an RDF store
    4. access information in a non-RDF database using the Apache Jena API
    D2RQ is Open Source software and published under the Apache license. The source code is available on GitHub. You can contact the dev team on the D2RQ mailing list at d2rq-map-devel@lists.sourceforge.net.

    Version:
    0.8.1
    Status:
    Completed
    Creation Date:
    08/12/2004
    Last Modification:
    06/2012
    Publisher:
    University of Mannheim and DERi Galway
    Language:
    English
    Keywords:
    Database-to-RDF Mapping, Linked Data Publication, SPARQL-to-SQL Rewriting
    Homepage:
    Representation:
    Java, RDF, SQL
    Contact Name:
    Christian Bizer
    Contact Email:
    chris@bizer.de
    License Document:
    Apache
    App. Developers:
    Developers can use D2RQ to access relational databases
    Data Providers:
    They can use D2R Server to publish their relational databases on the Web
    Data Consumers:
    The tool targets data providers
    Tech. Providers:
    The can include it into their products as for instance TopQuadrant is doing (OEM)

LDIF – Linked Data Integration Framework

  • Description:

    The Web of Linked Data grows rapidly and contains data from a wide range of different domains, including life science data, geographic data, government data, library and media data, as well as cross-domain data sets such as DBpedia or Freebase. Linked Data applications that want to consume data from this global data space face the challenges that:

    Data sources use a wide range of different RDF vocabularies to represent data about the same type of entity;
    The same real-world entity, for instance a person or a place, is identified with different URIs within different data sources;
    Data about the same real-world entity coming from different sources may contain conflicting value. For example the single value attribute population for a specific country can have multiple, different values after merging data from different sources. This usage of different vocabularies as well as the usage of URI aliases makes it very cumbersome for an application developer to write SPARQL queries against Web data which originates from multiple sources. In order to ease using Web data in the application context, it is thus advisable to translate data to a single target vocabulary (vocabulary mapping) and to replace URI aliases with a single target URI on the client side (identity resolution), before starting to ask SPARQL queries against the data.

    Up-till-now, there have not been any integrated tools that help application developers with these tasks. With LDIF, we try to fill this gap and provide an open-source Linked Data Integration Framework that can be used by Linked Data applications to translate Web data and normalize URI while keeping track of data provenance.

    Version:
    0.5.1
    Status:
    Completed
    Creation Date:
    29/06/2011
    Last Modification:
    02/2013
    Publisher:
    Universität Mannheim and MES Semantics
    Language:
    English
    Keywords:
    linked data, data integration, schema mapping, identity resolution, data quality assessment, data fusion
    Representation:
    JAVA
    Contact Name:
    Christian Bizer
    Contact Email:
    chris@bizer.de
    License Document:
    Apache
    Documentation:
    App. Developers:
    LDIF covers all necessary steps of the data integration process. It can be used to access and integrate Linked Data from various sources while keeping track of data provenance
    Data Providers:
    The tool targets data consumers
    Data Consumers:
    LDIF covers all necessary steps of the data integration process. It can be used to access and integrate Linked Data from various sources while keeping track of data provenance

Global Sensor Networks

  • Description:

    GSN is a Java environment that runs on one or more computers composing the backbone of the acquisition network. A set of wrappers allow to feed live data into the system. Then, the data streams are processed according to XML specification files. The system is built upon a concept of sensors (real sensors or virtual sensors, that is a new data source created from live data) that are connected together in order to built the required processing path. For example, one can imagine an anemometer that would sent its data into GSN through a wrapper (various wrappers are already available and writing new ones is quick), then that data stream could be sent to an averaging mote, the output of this mote could then be splited and sent for one part to a database for recording and to a web site for displaying the average measured wind in real time. All of this example could be done by editing only a few XML files in order to connect the various motes together.

    Version:
    1.1.2
    Version Description:
    New sensorscope listener server fixing problem of missing data Added count-based checkpoints to the CSV wrapper, logging line counts instead of latest timestamp Added new sensor types to sensorscope listener (NO2, CO, CO2, snow height, dendrometer, new temperature and air pressure sensors, data logger voltage, and GPS coordinates) Added /dynamicgeodata servlet, which allows making basic spatial queries for moving sensors Added new REST API for accessing data (beta) Fixed leak in unused connections (StorageManager pool configuration)
    Version Properties:
    Support for PostGIS for geospatial queries
    Status:
    UnderDevelopment
    Creation Date:
    11/11/2011
    Last Modification:
    02/2013
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    data stream, sensor network, distributed system
    Representation:
    Java XML XMLSchema RDF C++
    Contact Name:
    Sofiane Sarni
    Contact Email:
    sofiane.sarni@epfl.ch
    License Document:
    GPL
    License Description:
    GSN is distributed under GPL for the general public, in case GPL doesn't suite your needs, please contact us.
    Distribution Format:
    jar

SPARQL Extension for CKAN - ckanext-sparql

  • Description:

    This CKAN plugin offers two main functionalities:
    - It allows the configuration of a SPARQL endpoint for the whole CKAN platform in which metadata about every dataset in the platform can be queried.
    - It allows dataset editors to configure a RDF store and manage the RDF data of the dataset directly from CKAN, enabling at the same time a SPARQL endpoint for querying this data

    Version:
    0.1.0
    Status:
    UnderDevelopment
    Creation Date:
    04/03/2013
    Last Modification:
    03/2013
    Publisher:
    MORElab
    Language:
    English
    Keywords:
    ckan, sparql, rdf store
    Representation:
    Java, RDF, SPARQL
    Contact Name:
    Jon Lazaro
    Contact Email:
    jlazaro@deusto.es
    License Document:
    AGPL
    App. Developers:
    This tool allows application developers to query data from any CKAN instance through SPARQL queries
    Data Providers:
    his tool allows data providers (using CKAN) to configure a RDF store and manage the RDF data of the dataset directly from CKAN, enabling at the same time a SPARQL endpoint for querying this data
    Data Consumers:
    This tool allows data consumers to query data from any CKAN instance through SPARQL queries
    Researchers:
    This tool allows researchers to query data from any CKAN instance through SPARQL queries

geometry2rdf

  • Version:
    0.0.3
    Status:
    Completed
    Creation Date:
    01/02/2011
    Last Modification:
    01/2013
    Publisher:
    OEG
    Language:
    English
    Keywords:
    geometry rdf transformation
    Representation:
    Java, RDF
    License Document:
    N/A
    Documentation:
    Data Providers:
    It helps to transform geographical information contained in DDBBs to RDF format. This will increase the possible reuse of that information
    Data Consumers:
    It converts the information into a format, RDF, that it's easier to consult and more reusable that the information available in the DDBB

SILK extension for CKAN

  • Description:

    An extension for interlinking datasets uploaded on CKAN using SILK Link Discovery Framework.

    Version:
    beta
    Version Properties:
    Basic funtionalities and interface is done. Now, we are solving bugs and working on more complex interlinking rules
    Status:
    UnderDevelopment
    Creation Date:
    21/11/2012
    Last Modification:
    02/2012
    Publisher:
    DeustoTech - Internet
    Language:
    English
    Keywords:
    linked data, interlinking, CKAN, semantic web
    Representation:
    Java, RDF
    Contact Name:
    Mikel Emaldi
    Contact Email:
    m.emaldi@deusto.es
    License Document:
    Apache2.0
    Data Providers:
    This tool could help the Data Providers easing the process of using SILK to interlinking their datasets with others in the same CKAN instance

Morph-Streams

  • Description:

    SPARQL-Stream is a language that extends SPARQL for continuous query processing over streaming data. The Morph-streams module for SPARQL-Stream is a a java library that enables the execution of SPARQL-Stream queries, using different underlying DSMS or CEP (e.g. Esper, GSN, Cosm, SNEE, etc.). This tool allows posing SPARQL-Stream queries to an existing datasource using R2RML mappings. The mappings provide a descriptive way of relating ontological concepts (e.g. classes and properties) to elements of the DSMS or CERP schema (streams, tables). Morph uses a query rewriting approach to transform the SPARQL-Stream queeries to native queries understandable and executable by the DSMS or CEP, using the R2RML mappings. Then, When morph executes the queries in the original datasources, it is capable of translating the responses to variable bindings or triples, depending on the type of query.

    Version:
    1.0.2
    Version Properties:
    Query rewritting for Esper and GSN Support for R2RL mappings Basic support for joins,filter,simple aggregation Basic push delivery
    Status:
    UnderDevelopment
    Creation Date:
    14/07/2011
    Last Modification:
    03/2013
    Publisher:
    UPM
    Language:
    English
    Keywords:
    data stream, sensor network, query rewriting, sparql, query processing, rdf stream
    Representation:
    Java, Scala
    Contact Name:
    Jean-Paul Calbimonte
    Contact Email:
    jp.calbimonte@upm.es
    License Document:
    GPL
    Distribution Format:
    .jar
    App. Developers:
    Developers can use the library to execute SPARQL-Stream queries, which hide the internal implementations, providing an integrated interface that uses standard ontologies such as SSN
    Data Providers:
    The library allows stream and sensor data providers to expose their existing datasources through a SPARQL-Stream interface
    Data Consumers:
    Data consumers can access and query the data using SPARQL-Stream, and using ontological concepts and vocabularies
    Market Analysts:
    The library can be used to show the potential use of semantic technologies for streaming and dynamic sensor data
    Tech. Adopters:
    SPARQL adopters can use morph-streams to integrate existing semantically enabled applications so that they can use data from streaming data sources
    Tech. Providers:
    Technology providers can use morph-streams to deploy a SPARQL interface with streaming extensions on top of existing Data stream management systems
    Researchers:
    Researchers can use morph-streams to compare existing RDF-streaming engines or to exploit existing streaming datasources through SPARQL

Videk

  • Description:

    Videk currently uses four sources of sensor and linked data and relies on StreamSense engine for storage and processing. In this paper we present the architecture and current implementation of Videk as well as the lines along which we plan to extend and improve it.

    Status:
    UnderDevelopment
    Creation Date:
    01/01/2011
    Last Modification:
    01/2013
    Publisher:
    Unknown
    Language:
    English
    Keywords:
    Mash-up, sensors, web of things, real-time, data mining, semantic web.
    Representation:
    XML
    Contact Name:
    Carolina Fortuna
    Contact Email:
    carolina.fortuna@ijs.si
    License Document:
    Unknown
    App. Developers:
    by providing sensor data and meta-data
    Data Consumers:
    by providing sensor data and meta-data
    Market Analysts:
    by providing sensor data and meta-data
    Tech. Providers:
    by providing sensor data and meta-data
    Researchers:
    by providing sensor data and meta-data

ODEMapster

  • Description:

    A tool to transform database records into RDF instances using standard (R2RML) mapping language. This tool also allows to query the database using graph query language (SPARQL)

    Version:
    2.9.13
    Status:
    UnderDevelopment
    Creation Date:
    01/01/2007
    Last Modification:
    03/2013
    Publisher:
    Universidad Politécnica de Madrid
    Language:
    English
    Keywords:
    rdb2rdf, r2rml, rdf, sql, rdb
    Representation:
    Java, RDF, SPARQL
    Contact Name:
    Freddy Priyatna
    Contact Email:
    fpriyatna@fi.upm.es
    Data Providers:
    This tool can transform the legacy data stored in databases as RDF instances
    Data Consumers:
    This tool can help the user to pose SPARQL query to legacy data stored in databases

Yet-Another-SPARQL-GUI

  • Description:

    YASGUI is a web-based SPARQL client that can be used to query both remote and local endpoints. It integrates linked data services and web APIs to offer features such as autocompletion and endpoint lookup. It supports query retention – query texts persist across sessions – and query permalinks, as well as syntax checking and highlighting

    Version:
    13.03b
    Status:
    Completed
    Creation Date:
    01/07/2012
    Last Modification:
    03/2013
    Language:
    English
    Keywords:
    SPARQL, Semantic Web, Endpoints
    Representation:
    Java
    Contact Name:
    Laurens Rietveld
    Contact Email:
    laurens.rietveld@vu.nl
    License Document:
    MIT
    Data Providers:
    Enable users to navigate your endpoint in a more efficient/attractive, by using YASGUI as your main endpoint SPARQL client
    Data Consumers:
    Explore the Semantic Web and query any remote endpoint by using this robust, user friendly SPARQL client
    Researchers:
    YASGUI helps Web savvy developers as well: trying and testing SPARQL queries is often a cumbersome and painful experience: all who know the RDF namespace by heart raise their hands now!

D2RQ Plattform - Accessing Relational Databases as Virtual RDF Graphs

  • Description:

    The D2RQ Platform is a system for accessing relational databases as virtual, read-only RDF graphs. It offers RDF-based access to the content of relational databases without having to replicate it into an RDF store. Using D2RQ you can:
    1. query a non-RDF database using SPARQL
    2. access the content of the database as Linked Data over the Web
    3. create custom dumps of the database in RDF formats for loading into an RDF store
    4. access information in a non-RDF database using the Apache Jena API
    D2RQ is Open Source software and published under the Apache license. The source code is available on GitHub. You can contact the dev team on the D2RQ mailing list at d2rq-map-devel@lists.sourceforge.net.

    Version:
    0.8.1
    Status:
    Completed
    Creation Date:
    08/12/2004
    Last Modification:
    06/2012
    Publisher:
    University of Mannheim and DERi Galway
    Language:
    English
    Keywords:
    Database-to-RDF Mapping, Linked Data Publication, SPARQL-to-SQL Rewriting
    Homepage:
    Representation:
    Java, RDF, SQL
    Contact Name:
    Christian Bizer
    Contact Email:
    chris@bizer.de
    License Document:
    Apache
    App. Developers:
    Developers can use D2RQ to access relational databases
    Data Providers:
    They can use D2R Server to publish their relational databases on the Web
    Data Consumers:
    The tool targets data providers
    Tech. Providers:
    The can include it into their products as for instance TopQuadrant is doing (OEM)

Datalift Platform

  • Description:

    The Datalift web platform is a tool suite for converting, structured data sources and publishing them as linked data on the web.

    Status:
    UnderDevelopment
    Creation Date:
    01/10/2010
    Last Modification:
    03/2013
    Publisher:
    INRIA
    Language:
    English, French
    Keywords:
    linked-data, structured data, interlinking, LOV, vocabulary mapping, ontologies, sql, shape, statistics, sdmx, datacube, CSV, SPARQL, JSON, Java, javascript, XML, RDF
    Representation:
    JAVA, XML, RDF, SPARQL, C++, OWL
    Contact Name:
    François Scharffe
    Contact Email:
    francois.scharffe@lirmm.fr
    License Document:
    Apache
    Documentation:
    App. Developers:
    By following web standards, Datalift offers languages and protocols shared by everyone.
    Data Providers:
    Datalift helps to convert structured data to RDF, to select appropriate vocabularies to describe the data, and to interlink these data with other datasets.
    Data Consumers:
    Datalift provides standard HTTP and SPARQL access to data in a variety of formats (HTML, JSON, RDF/n3, RDF/XML). It also has a graphical user interface including a graphical SPARQL query builder.

MonetDB

  • Description:

    A relational database management system for high-performance data warehouses for business intelligence and eScience. Since a few years column store technology as pioneered in MonetDB has found its way into the product offerings of all major commercial database vendors. The market for applications empowered by these techniques provide ample space for further innovation, e.g. as demonstrated by our ongoing projects. At the same time, the landscape for major innovations remain wide open. A peek preview is given in the award winning paper titled: The Researcher’s Guide to the Data Deluge: Querying a Scientific Database in Just a Few Seconds. MonetDB is actively used in our research and real life applications. Nightly builds and regression testing ensure its quality, bug tracking helps to collect experiences and feature requests. Browsing the source code repository is supported by the Mercurial web frontend. Contributions ranging from bug reports, cross-platform issues, patches and features are highly appreciated.

    MonetDB is actively used in our research and real life applications. Nightly builds and regression testing ensure its quality, bug tracking helps to collect experiences and feature requests. Browsing the source code repository is supported by the Mercurial web frontend. Contributions ranging from bug reports, cross-platform issues, patches and features are highly appreciated

    Status:
    UnderDevelopment
    Publisher:
    CWI
    Language:
    English
    Keywords:
    column-stored, XML, data management
    Representation:
    Java, XML, RDF, SQL, SPARQL
    License Document:
    MonetDB Public License

Global Sensor Networks

  • Description:

    GSN is a Java environment that runs on one or more computers composing the backbone of the acquisition network. A set of wrappers allow to feed live data into the system. Then, the data streams are processed according to XML specification files. The system is built upon a concept of sensors (real sensors or virtual sensors, that is a new data source created from live data) that are connected together in order to built the required processing path. For example, one can imagine an anemometer that would sent its data into GSN through a wrapper (various wrappers are already available and writing new ones is quick), then that data stream could be sent to an averaging mote, the output of this mote could then be splited and sent for one part to a database for recording and to a web site for displaying the average measured wind in real time. All of this example could be done by editing only a few XML files in order to connect the various motes together.

    Version:
    1.1.2
    Version Description:
    New sensorscope listener server fixing problem of missing data Added count-based checkpoints to the CSV wrapper, logging line counts instead of latest timestamp Added new sensor types to sensorscope listener (NO2, CO, CO2, snow height, dendrometer, new temperature and air pressure sensors, data logger voltage, and GPS coordinates) Added /dynamicgeodata servlet, which allows making basic spatial queries for moving sensors Added new REST API for accessing data (beta) Fixed leak in unused connections (StorageManager pool configuration)
    Version Properties:
    Support for PostGIS for geospatial queries
    Status:
    UnderDevelopment
    Creation Date:
    11/11/2011
    Last Modification:
    02/2013
    Publisher:
    EPFL
    Language:
    English
    Keywords:
    data stream, sensor network, distributed system
    Representation:
    Java XML XMLSchema RDF C++
    Contact Name:
    Sofiane Sarni
    Contact Email:
    sofiane.sarni@epfl.ch
    License Document:
    GPL
    License Description:
    GSN is distributed under GPL for the general public, in case GPL doesn't suite your needs, please contact us.
    Distribution Format:
    jar