Research Interests Comparator (RIC) is our fourth electronic text mining project. The goal of the RIC system is to dramatically improve the ability of biomedical researchers to find information that is relevant to their areas of study, and to provide them
TRITE is a pre-computed set of Medline similarity hits on topics of interest to medical researchers. TRITE uses the eTBLAST engine, operating on an edited set of topics selected from the Encyclopidia of Molecular Biology, Blackwell Science, Limited. The q
eTBLAST is a unique search engine for searching biomedical literature. Our service is very different from PubMed. While PubMed searches for "keywords", our search engine lets you input an entire paragraph and returns MEDLINE abstracts that are similar to
This database was constructed in a purely automated manner, using a set of heuristics to identify acronyms/abbreviations and their definitions within MEDLINE records. Currently, it is the world's largest and most comprehensive catalog of biomedical acrony
provides a common platform for discussing extensions of the MediaWiki software that allow for simple, machine-based processing of Wiki content. This usually requires some form of "semantic annotation," a single solution for semantic annotation that fits t
IAwiki is a collaborative knowledge base for the topic of InformationArchitecture. Anyone can contribute, and there are no pre-registration hoops to jump thru ... just click the "Edit This Page" link at the bottom of any page.
...our attempt at integrating terminologies between different disciplines, which look at the problem differently, so some of the terms have different meanings in different disciplines.
...apply the following criteria: scope, technical quality, extensibility, quality of definitions of documentation...metamodeling is not for the faint of heart...if used correctly, its reduces development risk of complex projects by a larger factor than an
...good at abstracting from lower-level details of integration and interoperability, helping with partitioning problems into orthogonal sub-problems of conceptual data, physical data optimization, and control flow, & ideal for complex web services project
A metamodel is a precise definition of the constructs and rules needed for creating semantic models...an attempt at describing the world around us for a particular purpose.
"Whether you realize it or not, you're already familiar with controlled vocabularies. The Library of Congress subject headings and Yahoo's search criteria are a couple of examples. So, as you've probably guessed by now, controlled vocabularies are predete
In an attempt to summarize the relationship among various metadata formats and how they relate to building Internet systems I wrote a glossary. I then ordered and tied the terms together with a bit of narrative to explain the relationships among the terms
Semweb's core grammar is RDF, based on defining meaningful ontological statements as consisting of discrete “subjects,” “predicates,” and “objects,” and that each of those "parts of speech" can be given its own unique identity, designated with
Semantic Web Health Care and Life Sciences Interest Group is designed to improve collaboration, research and development, and innovation adoption in the health care and life science industries, & aiding decision-making in clinical research
Powerful Search Engine designed for Document Management, Competitive Intelligence, Press Analysis and Text Mining, Web Mining, Knowledge Discovery, Strategic Watch...Has Report Writer, Web Spider, Publisher, more...
The Suggested Upper Merged Ontology (SUMO) and its domain ontologies form the largest formal public ontology in existence today. They are being used for research and applications in search, linguistics and reasoning. SUMO is the only formal ontology that
At heart, a blog is just a database...But most blogs only divide up the information by time...for a reader, it's probably the least interesting way of reading. I've turned Ishbadiddle into a semantic web...categorized, coded, 1,385 keywords...
"Swoogle is a search engine for the Semantic Web on the Web. Swoogle crawl the World Wide Web for a special class of web documents called Semantic Web documents, which are written in RDF."
Information is stuck inside HTML pages, formatted in esoteric ways, difficult for machines to process. "Web 3.0", precursor to a refined semantic web, will change this. ‘Web 3.0′ will transform web sites into web services. Unstructured information bec
Discussion of flaws and potential solutions to using social bookmarking sites (from del.icio.us to digg) as folks monetize, abuse, trick, and tweak them: no uniform tagging conventions, flat tag structures (non-relational), overly-generalized tags (catego
* Linking Biomedical Information Through Text Mining * Semantic Webs for Life Sciences * Computational Approaches for Pharmacogenomics * Computational Proteomics
OntoSelect monitors the web to provide an access point for ontologies on any possible topic or domain that is automatically updated, organized in a meaningful way and with support for ontology search and selection. Selected ontologies may be used for ins
There's a bit of buzz today...around Freebase...tools like Freebase and Google Base bring it down to earth for the average user...but the semweb is in need of its own killer app.
...Web as a programmable data source as well as a platform for...[the] Web page. Early on, programmable access to Web data entailed a lot of screen scraping. Nowadays it often still does, but it's
Teachers and students from 10 European schools are working on tools regarding ELECTRONIC LEARNING ENVIRONMENTS like ejournals, webquests, weblogs, websites, video conferences, voip, photo imaging et cetera. It turns out that platforms like Dokeos, Lo-net,
Here, we present OntoSem resources that are used for deep meaning analysis of Web pages. These resources include concepts (language independent) and lexicons (in English) that are utilized by the hakia's OntoSem parser (called hOntoParser). The illustrati
With this Web page, we are opening some aspects of hakia R&D to the view of our users. We undertook highly specific research tasks solely dedicated to the advancement of the core-competency in Web search. The main challenge is to make science work in a co
Unlike vanilla XML, RDF vocabularies can be freely mixed together in data without prior agreement. So you often see ad-hoc combinations of Dublin Core, RSS1, MusicBrainz, RDF-calendar, FOAF, Wordnet, thesaurus, Geo-info etc etc frequently deployed togethe
You’d think that as a result of open-source development practices, blog architectures would be pretty close to perfection in areas like Web standards and maximum SEO impact. You’d be wrong. Unbelievably, nearly every WordPress, MovableType, or TypePad
For multistaged analysis of an Email Corpus. Social network analysis and text mining techniques are connected to enable an in depth view into the underlying information. Written in Java, published under the GNU GPL and hosted by Sourceforge, it runs under
The initiative will have four founding directors: Tim Berners-Lee, director of the World Wide Web Consortium, senior research scientist at MIT and professor at the University of Southampton; Wendy Hall, professor of computer science and head of the School
The Web Science Research Initiative brings together academics, scientists, sociologists, entrepreneurs and decision makers from around the world. These people will create the first multidisciplinary research body to examine the World Wide Web and offer th
Imagine a site that's a broker or mirror of metadata from other sites? You could go to this site, enter a URL and have the metadata from that page presented to you in clean, crisp XML. Even better if this was a Web service w/API free for anyone...quite a
Products for discovering and storing metadata, natural language processing, & more. Third link's to Geospatial Semantic Web Blog w/update on Metalink's ability to map its descriptions into RDF.
datasets like Wikipedia Data Dumps, 2000 Movie Reviews, & UPC Database are difficult to recreate, have high levels of accuracy, are valuable...as this becomes easier to access, the value of these datasets decreases over time.
Many visitors to blogs are turning to feed readers for consuming their favorite content. Are we looking at a change in how we should judge the traffic of a Website? We are already seeing advertisements appear in feeds which helps solve the issue of moneti
5 problems we'lll likely run into..each problem is a side-effect of advances in technology, rushes to fill new niches, or the previous two plus the desire to make a quick dollar.
This piece speaks about the different parts of the Semantic Web and how they fit together. For a high-level interview, take a look at Sandro Hawke's The Semantic Web (Put Simply). On the other hand, if you're a Web developer who's interested in building S
Taxonomy of Markup · I use a taxonomy of markup which I'm pretty sure was first advanced in the seminal November 1987 CACM article Markup systems and the future of scholarly text processing, by Coombs, Renear, and DeRose, which was the first place I ever
Machines, with their rigid information processing capabilities, need everything spelled out for them. To be able to do something useful with this title and byline, a machine would need to be able to parse it correctly. It would need to know that the numbe
Report from the 13th ACM Conference on Information and Knowledge Management, Washington DC, USA, 2004. Sponsors: ACM Special Interest Group on Information Retrieval, ACM Association for Computing Machinery
The Semantic Web has promised a new era of easier data integration. This article introduces core Semantic Web concepts and standards and explains how to expose an LDAP directory as a service that Semantic Web applications can consume using the open source
RxNorm, a standardized nomenclature for clinical drugs, is produced by the National Library of Medicine (NLM). In this context, a clinical drug is a pharmaceutical product given to (or taken by) a patient with a therapeutic or diagnostic intent. In RxNorm
xNav is a browser for RxNorm, the NLM repository of standard names for clinical drugs. RxNav displays links from clinical drugs, both branded and generic, to their active ingredients, drug components and related brand names. RxNorm is one of a suite of de
Divvio Inc., will turn on a service that automatically finds audio, video, and, eventually, text, on your favorite subjects. Then it weaves these clips together to create personalized multimedia channels that are updated each time you sign on.
Retrieves PubMed abstracts for your keywords, detects Medical Subject Headings (MeSH) in the abstracts, displays a subset of MeSH relevant to your search, and allows you to browse the ontology and display only papers containing specific MeSH terms.
With due respect, I think you, and even more aggresiously Clay Shirkey, have been misrepresenting what the Semantic Web is, and critiquing based on that misunderstanding, not on the reality. Folks like Danny Hillis and Nova Spivack who were listening got
The Metaweb technology that supports Freebase is indeed centralized, but only for computational speed. Distributing queries of real-world complexity over a high-latency network is a very difficult problem...[Freebase will] allow wiki-style editing of the
Metaweb is linking to massive content on the web, drawing relationships...[and making it possible for]"you to add high quality structured information to your websites, mashups, and applications without worrying about restrictive corporate licenses. All
Metaweb Technologies, Inc. is a company based in San Francisco that is developing Metaweb, a semantic data storage infrastructure for the web, and Freebase, an "open, shared database of the world's knowledge". The company was founded by Danny Hillis and o
“The semantic web will do for data what the web did for documents,” Spivack says. “It will make it universally searchable and sharable.” The standard way to organize and present data on the semantic web is described by the Resource Description Fr
A. Ankolekar, M. Krötzsch, T. Tran, and D. Vrandecic. WWW '07: Proceedings of the 16th international conference on World Wide Web, page 825--834. New York, NY, USA, ACM Press, (2007)