bookmarks  65

  •  

    SQLJet is an independent pure Java implementation of a popular SQLite database management system. SQLJet is a software library that provides API that enables Java application to read and modify SQLite databases. SQLJet does not support SQL queries; there is an API to work with the database on a lower level.
    8 years ago by @gresch
    (0)
     
     
  •  

    Drizzle is a community-driven open source project that is forked from the popular MySQL database. The Drizzle team has removed non-essential code, re-factored the remaining code and modernized the code base moving to C .
    8 years ago by @gresch
    (0)
     
     
  •  

    Squeryl Home Learn more Scaladoc Community Follow _squeryl on TwitterTwitter A Scala ORM and DSL for talking with Databases with minimum verbosity and maximum type safety Write compiler validated statements. Squeryl statements that pass compilation won’t fail at runtime. Refactor your schema as often as is required, the Scala compiler and your IDE will tell you exactly which lines of code are affected. Never repeat yourself The Composability of Squeryl statements allows you to define them once and reuse them as sub queries within other statements. Write declaratively Write as declaratively as SQL, only with less boilerplate. SQL’s declarativeness is preserved, not encapsulated in a lower level API that requires imperative and procedural code to get things done. Explicitly control retrieval granularity and laziness A significant part of optimizing a database abstraction layer is to choose for every situation the right balance between fine and large grained retrieval, and the optimal mix of laziness and eagerness. Data retrieval strategies are explicit in Squeryl rather than driven by configuration like current generation Java ORMs read more
    8 years ago by @gresch
    (0)
     
     
  •  

    DBpedia is a community effort to extract structured information from Wikipedia and to make this information available on the Web. DBpedia allows you to ask sophisticated queries against Wikipedia, and to link other data sets on the Web to Wikipedia data. We hope this will make it easier for the amazing amount of information in Wikipedia to be used in new and interesting ways, and that it might inspire new mechanisms for navigating, linking and improving the encyclopaedia itself.
    8 years ago by @gresch
    (0)
     
     
  •  

    Xml based metalanguage for a database schema definition. The following artifacts could be generated from this definition: DDL scripts for various database servers, HTML - documentation, ERD.
    8 years ago by @gresch
    (0)
     
     
  •  

    Sausalito is a product that enables you to develop, test, deploy, and host scalable web applications in the cloud. On this page, you find some reading material in case you're interested in * technical details about Sausalito, * user experience reports, or * presentations and demos that we gave.
    8 years ago by @gresch
    (0)
     
     
  •  

    FlockDB is a distributed graph database for storing adjancency lists, with goals of supporting: * a high rate of add/update/remove operations * potientially complex set arithmetic queries * paging through query result sets containing millions of entries * ability to "archive" and later restore archived edges * horizontal scaling including replication * online data migration Non-goals include: * multi-hop queries (or graph-walking queries) * automatic shard migrations FlockDB is much simpler than other graph databases such as neo4j because it tries to solve fewer problems. It scales horizontally and is designed for on-line, low-latency, high throughput environments such as web-sites. Twitter uses FlockDB to store social graphs (who follows whom, who blocks whom) and secondary indices. As of April 2010, the Twitter FlockDB cluster stores 13+ billion edges and sustains peak traffic of 20k writes/second and 100k reads/second.
    8 years ago by @gresch
    (0)
     
     
  •  

    SQLRunner is a simple JDBC Client which helps Java developers with various methods. You can build, test and run select, update and DML statements (also callable statements with out put parameters) and export or import data in/from csv-files or XML-fi
    8 years ago by @gresch
    (0)
     
     
  •  

    SYMPLiK EXZELLENZ is a Java program to manipulate table data in Oracle database by using Microsoft Excel file. Unlike Oracle WebADI, More4Apps or other Excel-driven solution, this program does not use Office VBA (Macro) to process the data. It is a pure Java solution incorporated with Apache POI and Oracle JDBC driver to download, upload, insert, update and delete data from Oracle database tables.
    8 years ago by @gresch
    (0)
     
     
  •  

    AutoPatch was born from the needs of using an agile development process while working on systems that have persistent storage. Without AutoPatch, developers usually can't afford the maintenance headache of their own database, and DBAs are required just to apply changes to all of the various environments a serious development effort requires. The very application of database changes becomes an inefficient, error-prone, expensive process, all conspiring to discourage any refactoring that touches the model, or being a bottleneck when model changes are made. AutoPatch solves this problem, completely. With AutoPatch, an agile development process that requires a database change looks like this: * Developer alters the model, which requires a change to the database * Developer possibly consults a DBA, and develops a SQL patch against their personal database that implements the alteration * Developer commits the patch to source control at the same time as they commit their dependent code * Other developers' and environments' databases are automatically updated by AutoPatch the next time the new source is run This represents streamlined environment maintenance, allowing developers to cheaply have their own databases and all databases to stay in synch with massively lower costs and no environment skew. That's what AutoPatch does. Clusters with one database? Multiple schemas? Logical migrations, instead of just DDL changes? Need to do something special/custom? Need to distribute your changes commercially? All without paying anything? No problem.
    9 years ago by @gresch
    (0)
     
     
  •  

    Neo4j is a graph database. It is an embedded, disk-based, fully transactional Java persistence engine that stores data structured in graphs rather than in tables. A graph (mathematical lingo for a network) is a flexible data structure that allows a more agile and rapid style of development. You can think of Neo4j as a high-performance graph engine with all the features of a mature and robust database. The programmer works with an object-oriented, flexible network structure rather than with strict and static tables — yet enjoys all the benefits of a fully transactional, enterprise-strength database. Neo4j is released under a dual free software/commercial license model (which basically means that it’s “open source” but if you’re interested in using it in commercially, then you must buy a commercial license). Neo4j has been in commercial development for 8 years and in production for over 5 years. It is a mature and robust graph database that provides: * an intuitive graph-oriented model for data representation. Instead of static and rigid tables, rows and columns, you work with a flexible graph network consisting of nodes, relationships and properties. * a disk-based, native storage manager completely optimized for storing graph structures for maximum performance and scalability. * massive scalability. Neo4j can handle graphs of several billion nodes/relationships/properties on a single machine and can be sharded to scale out across multiple machines. * a powerful traversal framework for high-speed traversals in the node space. * a small footprint. Neo4j is a single <500k jar with one dependency (the Java Transaction API). * a simple and convenient object-oriented API. * optional layers to expose Neo4j as an RDF store, i.e. easily inject / extract data as RDF, express meta model semantics using OWL and query the node space using SPARQL. When it comes to scalability numbers, remember that several triples are usually mapped to a single node. (currently being developed under the umbrella of the OpenMetadir project)
    9 years ago by @gresch
    (0)
     
     
  •  

    RESTORE DATABASE Example. MS SQL is a pain, pain, pain. Hint: Use e.g. RESTORE FILELISTONLY FROM DISK='C:\MyData.bak' to get the necessary information. Here the command from the cited blog entry: ALTER DATABASE MyData SET SINGLE_USER WITH ROLLBACK IMMEDIATE RESTORE DATABASE MyData FROM DISK = 'C\:BackupMyData.bak' WITH MOVE 'MyData_Data' TO 'C:\Data\datafile.mdf', MOVE 'MyData_Log' TO 'C:\Data\logfile.ldf', REPLACE
    10 years ago by @gresch
    (0)
     
     
  •  

    Very interesting approach! "Apache Empire-db is an Open Source relational data persistence component which allows database vendor independent dynamic query definition as well as safe and simple data retrieval and updating. Compared to most other solutions like e.g. Hibernate, TopLink, iBATIS or JPA implementations, Empire-db takes a considerably different approach, with a special focus on compile-time safety, reduced redundancies and improved developer productivity."
    10 years ago by @gresch
    (0)
     
     
  •  

    The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing, including: * Hadoop Core, our flagship sub-project, provides a distributed filesystem (HDFS) and support for the MapReduce distributed computing framework. * HBase builds on Hadoop Core to provide a scalable, distributed database. * Pig is a high-level data-flow language and execution framework for parallel computation. It is built on top of Hadoop Core. * ZooKeeper is a highly available and reliable coordination system. Distributed applications use ZooKeeper to store and mediate updates for critical shared state. * Hive is a data warehouse infrastructure built on Hadoop Core that provides data summarization, adhoc querying and analysis of datasets.
    10 years ago by @gresch
    (0)
     
     
  •  

    What CouchDB is * A document database server, accessible via a RESTful JSON API. * Ad-hoc and schema-free with a flat address space. * Distributed, featuring robust, incremental replication with bi-directional conflict detection and management. * Query-able and index-able, featuring a table oriented reporting engine that uses Javascript as a query language. What it is Not * A relational database. * A replacement for relational databases. * An object-oriented database. Or more specifically, meant to function as a seamless persistence layer for an OO programming language.
    10 years ago by @gresch
    (0)
     
     
  •  

    Relational databases define a strict structure and provide a rigid way to maintain data for a software application. Apache's open source CouchDB offers a new method of storing data, in what is referred to as a schema-free document-oriented database model. Instead of the highly structured data storage of a relational model, CouchDB stores data in a semi-structured fashion, using a JavaScript-based view model for generating structured aggregation and report results from these semi-structured documents. CouchDB has been developed from the ground up with Web applications as the primary focus and has its sights on becoming the de-facto database for Web application development.
    10 years ago by @gresch
    (0)
     
     
  •  

    A library for using Memcached as a second level distributed cache in Hibernate. * Based on the excellent spymemcached client * Includes support for the Whalin (danga) memcached client * Supports entity and query caching. * See the Configuration page * See how easy it is to configure in Grails
    10 years ago by @gresch
    (0)
     
     
  •  

    The project provides extensions to Hibernate to provide complete auditing capabilities. It uses a set of tables in the DB that do not depend on the object model
    10 years ago by @gresch
    (0)
     
     
  •  

    Das Problem … PHP-Scripte werden nach einer bestimmten Laufzeit abgebrochen (normalerweise nach 30 Sekunden), und so funktioniert ein Backup mit diversen Tools nur bis zu einer bestimmten Größe. Braucht das Script länger als die 30 Sekunden, so wird es vom Server einfach abgebrochen, und man kommt nicht mehr an sein Backup heran. Gleiches gilt für das Wiedereinspielen eines Backups. Wer einmal ein Dumpfile von Hand in viele kleinere Einzelabschnitte zerlegt hat, um eine Datenbank wieder herzustellen, der weiß genau, wovon wir reden. MySQLDumper ist die Lösung … MySQLDumper umgeht den Timeout-Error mit Hilfe eines kleinen Tricks: Er liest nur eine bestimmte Anzahl von Datensätzen aus der Tabelle aus, merkt sich, wie weit er gekommen ist, und ruft sich anschließend selbst auf. Dadurch erhält das Script bei jedem Aufruf wieder die vollen 30 Sekunden und umgeht so den Timeout-Error. Das gleiche Prinzip benutzt MySQLDumper auch beim Wiederherstellen der Daten. MySQLDumper kann die Daten beim Sichern sofort packen. Auch das Wiederherstellungsscript kann direkt aus dieser gepackten Datei lesen, ohne dass sie auf dem Server entpackt werden muss! Natürlich kann man die Datei auch ungepackt lassen, aber spätestens beim Hochladen eines Backups weiß man dies zu schätzen.
    10 years ago by @gresch
    (0)
     
     
  •  

    DbFit is a set of FIT fixtures which enables FIT/FitNesse tests to execute directly against a database. This enables developers to manipulate database objects in a relational tabular form, making database testing and management much easier then with xUnit-style tools. The library is free to use, released under GNU GPL.
    10 years ago by @gresch
    (0)
     
     

publications  

    No matching posts.
  • ⟨⟨
  • ⟩⟩