This document describes the efforts of the StratusLab project to better understand its target communities, to gauge their experience with cloud technologies, to validate the defined use cases, and to extract relevant requirements from the communities. In parallel, the exercise was used as a dissemination tool to inform people about existing software packages, to increase the awareness of StratusLab, and to expand the our contacts within our target communities. The project created, distributed, and analyzed two surveys to achieve these goals. They validate the defined use cases and provide detailed requirements. One identified, critical issue relates to system administrators’ reluctance to allow users to run their own virtual machines on the infrastructure. The project must define the criteria to trust such images and provide sufficient sand-boxing to avoid threats to other machines and services.
The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing, including:
* Hadoop Core, our flagship sub-project, provides a distributed filesystem (HDFS) and support for the MapReduce distributed computing framework.
* HBase builds on Hadoop Core to provide a scalable, distributed database.
* Pig is a high-level data-flow language and execution framework for parallel computation. It is built on top of Hadoop Core.
* ZooKeeper is a highly available and reliable coordination system. Distributed applications use ZooKeeper to store and mediate updates for critical shared state.
* Hive is a data warehouse infrastructure built on Hadoop Core that provides data summarization, adhoc querying and analysis of datasets.
The goal of the Condor® Project is to develop, implement, deploy, and evaluate mechanisms and policies that support High Throughput Computing (HTC) on large collections of distributively owned computing resources. Guided by both the technological and sociological challenges of such a computing environment, the Condor Team has been building software tools that enable scientists and engineers to increase their computing throughput
OGF is an open community committed to driving the rapid evolution and adoption of applied distributed computing. Applied Distributed Computing is critical to developing new, innovative and scalable applications and infrastructures that are essential to p
It looks like we have to say goodbye to our good, old grids of the past -- at least to all those beautiful features and capabilities envisioned 10 years ago, when grids were supposed to evolve toward coordinated resource sharing and problem solving in dyn
Apparently, when CERN isn't colliding particles (and ripping massive holes in the space-time continuum), it's busy working on a new "internet" which will be 10,000 times faster than our current version.