activemq/kahadb
Gary Tully d458a5d17f fix org.apache.kahadb.journal.JournalTest.testBatchWriteCompleteAfterClose() after changes for AMQ-2143, the file and offset in a location are set after an enqueue to a write batch, so doing the inflight add just overwrote the same entry, -1:-1. Which fixed the leak but broke inflights, proper fix is to do the inflight addition in enqueue, just after the location details are known and with a lock on appender enqueue
git-svn-id: https://svn.apache.org/repos/asf/activemq/trunk@748475 13f79535-47bb-0310-9956-ffa450edef68
2009-02-27 10:22:50 +00:00
..
src fix org.apache.kahadb.journal.JournalTest.testBatchWriteCompleteAfterClose() after changes for AMQ-2143, the file and offset in a location are set after an enqueue to a write batch, so doing the inflight add just overwrote the same entry, -1:-1. Which fixed the leak but broke inflights, proper fix is to do the inflight addition in enqueue, just after the location details are known and with a lock on appender enqueue 2009-02-27 10:22:50 +00:00
README Moving the kahadb work from the sandbox into trunk now that 5.2 has been released. 2008-11-21 18:19:50 +00:00
pom.xml applied patch from https://issues.apache.org/activemq/browse/AMQ-2068 with thanks 2009-01-16 11:47:02 +00:00

README

=======================================================================
 KahaDB : An Embedded Lightweight Non-Relational Database
=======================================================================

Goals:
  * Provide a simple collections based interface for storing persistent
    data.
  * Provide a lower level key value interface for storing data. Secondary
    keys should be supported.
  * Fast recovery

Status:
  * Moved the org.apache.activemq.kaha package into this module
    and renamed it as org.apache.kahadb
  * Moved a refactored version of the low level indexing logic
    used by Apache xindice to the org.apache.kahadb.xindice page.
  
TODO:
  * None of our current index support fast recovery.  They suffer
    from partial page write failures so on a crash, they have
    to be fully rebuilt from the data logs.
  * Indexes should know what the last transaction id that updated
    it, so that missing operations can be replayed. 
  * Investigate implementing data log replication for HA.