activemq/kahadb
Gary Tully 1463bec066 https://issues.apache.org/jira/browse/AMQ-3768: ClassCastException when running some Durable Consumer test cases. root cause of the classcast was the reuse of a freed node that was still referenced as the head page id of a listindex. The fix is to not modify the head page id of a listindex when removing and coalescing linked pages, the head page remains valid for the duration of a subscription. Eventually got a test case that could easlily reproduce
git-svn-id: https://svn.apache.org/repos/asf/activemq/trunk@1304020 13f79535-47bb-0310-9956-ffa450edef68
2012-03-22 20:01:40 +00:00
..
src https://issues.apache.org/jira/browse/AMQ-3768: ClassCastException when running some Durable Consumer test cases. root cause of the classcast was the reuse of a freed node that was still referenced as the head page id of a listindex. The fix is to not modify the head page id of a listindex when removing and coalescing linked pages, the head page remains valid for the duration of a subscription. Eventually got a test case that could easlily reproduce 2012-03-22 20:01:40 +00:00
README Moving the kahadb work from the sandbox into trunk now that 5.2 has been released. 2008-11-21 18:19:50 +00:00
pom.xml Fixes AMQ-3570: Improve ActiveMQ's XSD to allow elements to appear in any order 2011-10-27 13:19:49 +00:00

README

=======================================================================
 KahaDB : An Embedded Lightweight Non-Relational Database
=======================================================================

Goals:
  * Provide a simple collections based interface for storing persistent
    data.
  * Provide a lower level key value interface for storing data. Secondary
    keys should be supported.
  * Fast recovery

Status:
  * Moved the org.apache.activemq.kaha package into this module
    and renamed it as org.apache.kahadb
  * Moved a refactored version of the low level indexing logic
    used by Apache xindice to the org.apache.kahadb.xindice page.
  
TODO:
  * None of our current index support fast recovery.  They suffer
    from partial page write failures so on a crash, they have
    to be fully rebuilt from the data logs.
  * Indexes should know what the last transaction id that updated
    it, so that missing operations can be replayed. 
  * Investigate implementing data log replication for HA.