Mloss is a community effort at
producing reproducible research
via open source software, open
access to data and results, and
open standards for interchange.
The Car Options Ontology is designed to be used in combination with GoodRelations for the commercial aspects of offers for sale or rental, and the Vehicle Sales Ontology for car features.
Imagine you can see 160 years of history, all on one screen. You can zoom and pan, you can look at a particular day, you can even do a search. And when you do, the results come up not as a list, but as a heat map that shows where in history that topic appears, and how often.
Gephi is an open-source software for visualizing and analyzing large networks graphs. Gephi uses a 3D render engine to display graphs in real-time and speed up the exploration. Use Gephi to explore, analyse, spatialise, filter, cluterize, manipulate and export all types of graphs.
Tupelo is a data and metadata management system based on semantic web technologies. Tupelo provides a variety of generic utilities for managing data and metadata using both best-of-breed semantic database implementations such as Jena and Sesame, as well as ordinary storage technologies such as flat files. Tupelo makes data and metadata portable across a variety of Contexts and deployment scenarios, including desktop applications, web-based applications, and more complex distributed architectures. Its use of global identification and explicit semantics means that metadata created and managed with Tupelo can be easily exported and used by a wide variety of RDF-aware tools and technologies.
I often end up running a big SPARQL query (usually on a server), exporting the results a TSV, and post processing the results with some combination of vi, perl, awk, sort etc., then loading the processed data into a copy of Excel to get stats out of it, or produce a chart or whatever.
The other day I was wondering if you could pull results directly from a SPARQL endpoint into Excel. Well, it turns out that you can, via something called an Internet Query File.
LIMES is a link discovery framework for the Web of Data. It implements time-efficient approaches for large-scale link discovery based on the characteristics of metric spaces. It is easily configurable via a web interface. It can also be downloaded as standalone tool for carrying out link discovery locally.
FOX is a framework that integrates the Linked Data Cloud and makes uses of the diversity of NLP algorithms to extract RDF triples of high accuracy out of NL. In its current version, it integrates and merges the results of Named Entity Recognition, Keyword Extraction and Relation Extraction tools.
bigdata(R) is a scale-out storage and computing fabric supporting optional transactions, very high concurrency, and very high aggregate IO rates.
Features statement-level provenance, free-text search, and incremental load and retraction, inference etc.
DBpedia Spotlight is a tool for annotating mentions of DBpedia resources in text, providing a solution for linking unstructured information sources to the Linked Open Data cloud through DBpedia. DBpedia Spotlight performs named entity extraction, including entity detection and Name Resolution (a.k.a. disambiguation). It can also be used for building your solution for Named Entity Recognition, amongst other information extraction tasks.
The Calais Web Service ingests unstructured text and returns Resource Description Framework formatted results identifying entities, facts and events within the text.[4] The service appears to be based on technology acquired when Reuters purchased ClearForest in 2007.
Zemanta is a tool that looks over your shoulder while you blog and gives you tips and advice, suggests related content and pictures and makes sure your posts get promoted as they deserve to be. We at Zemanta are thinking hard to help make blogging easier for you. We're engineering better creative tools to help you get the most out of your blogging time.
A reconciliation service is a web service that, given some text which is a name or label for something, and optionally some additional details, returns a ranked list of potential entities matching the criteria. The candidate text does not have to match each entity's official name perfectly, and that's the whole point of reconciliation--to get from ambiguous text name to precisely identified entities. For instance, given the text "apple", a reconciliation service probably should return the Apple Inc. company, the apple fruit, and New York city (also known as the Big Apple).
The consumption of Linked Data is a task of increasing complexity on the Web of Data. This complexity is due to several factors, including the sheer size of the Web of Data, the diversity of vocabularies used to describe this data and the lack of schema information in knowledge bases. ALOE provide a semi-automatic solution to address this problem.
A workshop of the 10th International Semantic Web Conference (ISWC 2011), 23-27 October 2011, Bonn, Germany.
The goal of the Semantic Sensor Networks workshop is to develop an understanding of the ways semantic web technologies, including ontologies, agent architectures and semantic web services, can contribute to the growth, application and deployment of large-scale sensor networks and their applications. The workshop provides an inter-disciplinary forum to explore and promote these concepts.
‘320 and Up’ starts with a tiny screen stylesheet that contains only reset, colour and typography styles. Media Queries then load assets and layout styles progressively and only as they’re needed. Think of this as responsible responsive design.
Clerezza is a service platform based on OSGi (Open Services Gateway initiative) which provides a set of functionality for management of semantically linked data accessible through RESTful Web Services and in a secured way. Furthermore, Clerezza allows to easily develop semantic web applications by providing tools to manipulate RDF data, create RESTful Web Services and Renderlets using ScalaServerPages.
SUBDUE is a graph-based knowledge discovery system that finds structural, relational patterns in data representing entities and relationships. SUBDUE represents data using a labeled, directed graph in which entities are represented by labeled vertices or subgraphs, and relationships are represented by labeled edges between the entities. SUBDUE uses the minimum description length (MDL) principle to identify patterns that minimize the number of bits needed to describe the input graph after being compressed by the pattern. SUBDUE can perform several learning tasks, including unsupervised learning, supervised learning, clustering and graph grammar learning.
Attempto Controlled English (ACE) is a controlled natural language, i.e. a rich subset of standard English designed to serve as knowledge representation language. ACE allows users to express texts precisely, and in the terms of their respective application domain.
The official home of UK legislation, revised and as enacted 1267-present. This website is managed by The National Archives on behalf of HM Government. Publishing all UK legislation is a core part of the remit of Her Majesty’s Stationery Office (HMSO), part of The National Archives, and the Office of the Queen's Printer for Scotland.
AIDA is a framework and online tool for entity detection and disambiguation. Given a natural-language text or a Web table, it maps mentions of ambiguous names onto canonical entities (e.g., individual people or places) registered in the YAGO2 knowledge base.
The news ontology is comprised of several ontologies, which describe assets (text, images, video) and the events and entities (people, places, organisations, abstract concepts etc.) that appear in news content.
Callimachus (kəlĭm'əkəs) is a framework for data-driven applications based on Linked Data principles. Callimachus allows Web authors to quickly and easily create semantically-enabled Web applications.
HEigen is a spectral analysis tool which computes top k eigenvalues and corresponding eigenvectors of extremely large(~billions of nodes and edges) graphs. HEigen runs on top of Hadoop platform.
The workshop will focus on the ways semantic web and related technologies (ontologies, linked data, etc.) can be used across the food and agriculture supply chain to facilitate the exchange of information and enable the creation of additional services for farmers, transporters and processors of food, retailers and final consumers, in effect the whole supply chain.
ConceptNet is a semantic network containing lots of things computers should know about the world, especially when understanding text written by people.
It is built from nodes representing concepts, in the form of words or short phrases of natural language, and labeled relationships between them. These are the kinds of things computers need to know to search for information better, answer questions, and understand people's goals. If you wanted to build your own Watson, this should be a good place to start!
MapProxy is an open source proxy for geospatial data. It caches, accelerates and transforms data from existing map services and serves any desktop or web GIS client.
Mapbender is the back office software and client framework for spatial data infrastructures. The software is implemented in PHP, JavaScript and XML and dual licensed under GNU GPL and Simplified BSD license. It provides a data model and web based interfaces for displaying, navigating and querying OGC compliant map services.
M. Banko, and E. Brill. ACL '01: Proceedings of the 39th Annual Meeting on Association for Computational Linguistics, page 26--33. Morristown, NJ, USA, Association for Computational Linguistics, (2001)
D. Lin. Proceedings of the 17th international conference on Computational linguistics, page 768--774. Morristown, NJ, USA, Association for Computational Linguistics, (1998)
G. Forman, and E. Kirshenbaum. CIKM '08: Proceeding of the 17th ACM conference on Information and knowledge management, page 1221--1230. New York, NY, USA, ACM, (2008)
D. Shahaf, C. Guestrin, and E. Horvitz. Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining, page 1122--1130. New York, NY, USA, ACM, (2012)
B. Elliott, E. Cheng, C. Thomas-Ogbuji, and Z. Ozsoyoglu. Proceedings of the 2009 International Database Engineering & Applications Symposium, page 31--42. New York, NY, USA, ACM, (2009)