AIDA is a framework and online tool for entity detection and disambiguation. Given a natural-language text or a Web table, it maps mentions of ambiguous names onto canonical entities (e.g., individual people or places) registered in the YAGO2 knowledge base.
The official home of UK legislation, revised and as enacted 1267-present. This website is managed by The National Archives on behalf of HM Government. Publishing all UK legislation is a core part of the remit of Her Majesty’s Stationery Office (HMSO), part of The National Archives, and the Office of the Queen's Printer for Scotland.
Attempto Controlled English (ACE) is a controlled natural language, i.e. a rich subset of standard English designed to serve as knowledge representation language. ACE allows users to express texts precisely, and in the terms of their respective application domain.
SUBDUE is a graph-based knowledge discovery system that finds structural, relational patterns in data representing entities and relationships. SUBDUE represents data using a labeled, directed graph in which entities are represented by labeled vertices or subgraphs, and relationships are represented by labeled edges between the entities. SUBDUE uses the minimum description length (MDL) principle to identify patterns that minimize the number of bits needed to describe the input graph after being compressed by the pattern. SUBDUE can perform several learning tasks, including unsupervised learning, supervised learning, clustering and graph grammar learning.
Clerezza is a service platform based on OSGi (Open Services Gateway initiative) which provides a set of functionality for management of semantically linked data accessible through RESTful Web Services and in a secured way. Furthermore, Clerezza allows to easily develop semantic web applications by providing tools to manipulate RDF data, create RESTful Web Services and Renderlets using ScalaServerPages.
‘320 and Up’ starts with a tiny screen stylesheet that contains only reset, colour and typography styles. Media Queries then load assets and layout styles progressively and only as they’re needed. Think of this as responsible responsive design.
A workshop of the 10th International Semantic Web Conference (ISWC 2011), 23-27 October 2011, Bonn, Germany.
The goal of the Semantic Sensor Networks workshop is to develop an understanding of the ways semantic web technologies, including ontologies, agent architectures and semantic web services, can contribute to the growth, application and deployment of large-scale sensor networks and their applications. The workshop provides an inter-disciplinary forum to explore and promote these concepts.
The consumption of Linked Data is a task of increasing complexity on the Web of Data. This complexity is due to several factors, including the sheer size of the Web of Data, the diversity of vocabularies used to describe this data and the lack of schema information in knowledge bases. ALOE provide a semi-automatic solution to address this problem.
A reconciliation service is a web service that, given some text which is a name or label for something, and optionally some additional details, returns a ranked list of potential entities matching the criteria. The candidate text does not have to match each entity's official name perfectly, and that's the whole point of reconciliation--to get from ambiguous text name to precisely identified entities. For instance, given the text "apple", a reconciliation service probably should return the Apple Inc. company, the apple fruit, and New York city (also known as the Big Apple).
Zemanta is a tool that looks over your shoulder while you blog and gives you tips and advice, suggests related content and pictures and makes sure your posts get promoted as they deserve to be. We at Zemanta are thinking hard to help make blogging easier for you. We're engineering better creative tools to help you get the most out of your blogging time.
The Calais Web Service ingests unstructured text and returns Resource Description Framework formatted results identifying entities, facts and events within the text.[4] The service appears to be based on technology acquired when Reuters purchased ClearForest in 2007.
DBpedia Spotlight is a tool for annotating mentions of DBpedia resources in text, providing a solution for linking unstructured information sources to the Linked Open Data cloud through DBpedia. DBpedia Spotlight performs named entity extraction, including entity detection and Name Resolution (a.k.a. disambiguation). It can also be used for building your solution for Named Entity Recognition, amongst other information extraction tasks.
bigdata(R) is a scale-out storage and computing fabric supporting optional transactions, very high concurrency, and very high aggregate IO rates.
Features statement-level provenance, free-text search, and incremental load and retraction, inference etc.
FOX is a framework that integrates the Linked Data Cloud and makes uses of the diversity of NLP algorithms to extract RDF triples of high accuracy out of NL. In its current version, it integrates and merges the results of Named Entity Recognition, Keyword Extraction and Relation Extraction tools.
LIMES is a link discovery framework for the Web of Data. It implements time-efficient approaches for large-scale link discovery based on the characteristics of metric spaces. It is easily configurable via a web interface. It can also be downloaded as standalone tool for carrying out link discovery locally.
I often end up running a big SPARQL query (usually on a server), exporting the results a TSV, and post processing the results with some combination of vi, perl, awk, sort etc., then loading the processed data into a copy of Excel to get stats out of it, or produce a chart or whatever.
The other day I was wondering if you could pull results directly from a SPARQL endpoint into Excel. Well, it turns out that you can, via something called an Internet Query File.
Tupelo is a data and metadata management system based on semantic web technologies. Tupelo provides a variety of generic utilities for managing data and metadata using both best-of-breed semantic database implementations such as Jena and Sesame, as well as ordinary storage technologies such as flat files. Tupelo makes data and metadata portable across a variety of Contexts and deployment scenarios, including desktop applications, web-based applications, and more complex distributed architectures. Its use of global identification and explicit semantics means that metadata created and managed with Tupelo can be easily exported and used by a wide variety of RDF-aware tools and technologies.
Gephi is an open-source software for visualizing and analyzing large networks graphs. Gephi uses a 3D render engine to display graphs in real-time and speed up the exploration. Use Gephi to explore, analyse, spatialise, filter, cluterize, manipulate and export all types of graphs.