LUCENE-1007: allow flushing by ram, doc count, or both (whichever comes first) in IndexWriter

git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@581353 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael McCandless 2007-10-02 19:15:43 +00:00
parent b89d70d042
commit 54729a50b6
10 changed files with 172 additions and 60 deletions

View File

@ -179,6 +179,11 @@ Optimizations
implementations that use SegmentInfos to access an index and implementations that use SegmentInfos to access an index and
acquire a write lock for index modifications. (Michael Busch) acquire a write lock for index modifications. (Michael Busch)
11. LUCENE-1007: Allow flushing in IndexWriter to be triggered by
either RAM usage or document count or both (whichever comes
first), by adding symbolic constant DISABLE_AUTO_FLUSH to disable
one of the flush triggers. (Ning Li via Mike McCandless)
Documentation Documentation
Build Build

View File

@ -21,7 +21,6 @@ import org.apache.lucene.store.Directory;
import java.io.IOException; import java.io.IOException;
import java.util.List; import java.util.List;
import java.util.LinkedList;
import java.util.ArrayList; import java.util.ArrayList;
/** A {@link MergeScheduler} that runs each merge using a /** A {@link MergeScheduler} that runs each merge using a

View File

@ -126,7 +126,7 @@ final class DocumentsWriter {
private int pauseThreads; // Non-zero when we need all threads to private int pauseThreads; // Non-zero when we need all threads to
// pause (eg to flush) // pause (eg to flush)
private boolean flushPending; // True when a thread has decided to flush private boolean flushPending; // True when a thread has decided to flush
private boolean postingsIsFull; // True when it's time to write segment private boolean bufferIsFull; // True when it's time to write segment
private PrintStream infoStream; private PrintStream infoStream;
@ -148,6 +148,11 @@ final class DocumentsWriter {
// non-zero we will flush by RAM usage instead. // non-zero we will flush by RAM usage instead.
private int maxBufferedDocs = IndexWriter.DEFAULT_MAX_BUFFERED_DOCS; private int maxBufferedDocs = IndexWriter.DEFAULT_MAX_BUFFERED_DOCS;
// Coarse estimates used to measure RAM usage of buffered deletes
private static int OBJECT_HEADER_BYTES = 12;
private static int OBJECT_POINTER_BYTES = 4; // TODO: should be 8 on 64-bit platform
private static int BYTES_PER_CHAR = 2;
private BufferedNorms[] norms = new BufferedNorms[0]; // Holds norms until we flush private BufferedNorms[] norms = new BufferedNorms[0]; // Holds norms until we flush
DocumentsWriter(Directory directory, IndexWriter writer) throws IOException { DocumentsWriter(Directory directory, IndexWriter writer) throws IOException {
@ -165,18 +170,25 @@ final class DocumentsWriter {
/** Set how much RAM we can use before flushing. */ /** Set how much RAM we can use before flushing. */
void setRAMBufferSizeMB(double mb) { void setRAMBufferSizeMB(double mb) {
if (mb == IndexWriter.DISABLE_AUTO_FLUSH) {
ramBufferSize = IndexWriter.DISABLE_AUTO_FLUSH;
} else {
ramBufferSize = (long) (mb*1024*1024); ramBufferSize = (long) (mb*1024*1024);
} }
}
double getRAMBufferSizeMB() { double getRAMBufferSizeMB() {
if (ramBufferSize == IndexWriter.DISABLE_AUTO_FLUSH) {
return ramBufferSize;
} else {
return ramBufferSize/1024./1024.; return ramBufferSize/1024./1024.;
} }
}
/** Set max buffered docs, which means we will flush by /** Set max buffered docs, which means we will flush by
* doc count instead of by RAM usage. */ * doc count instead of by RAM usage. */
void setMaxBufferedDocs(int count) { void setMaxBufferedDocs(int count) {
maxBufferedDocs = count; maxBufferedDocs = count;
ramBufferSize = 0;
} }
int getMaxBufferedDocs() { int getMaxBufferedDocs() {
@ -361,7 +373,7 @@ final class DocumentsWriter {
threadBindings.clear(); threadBindings.clear();
numBytesUsed = 0; numBytesUsed = 0;
balanceRAM(); balanceRAM();
postingsIsFull = false; bufferIsFull = false;
flushPending = false; flushPending = false;
segment = null; segment = null;
numDocsInRAM = 0; numDocsInRAM = 0;
@ -582,7 +594,7 @@ final class DocumentsWriter {
} }
} }
if (postingsIsFull && !flushPending) { if (bufferIsFull && !flushPending) {
flushPending = true; flushPending = true;
doFlushAfter = true; doFlushAfter = true;
} }
@ -961,7 +973,8 @@ final class DocumentsWriter {
for(int i=0;i<numFields;i++) for(int i=0;i<numFields;i++)
fieldDataArray[i].processField(analyzer); fieldDataArray[i].processField(analyzer);
if (numBytesUsed > 0.95 * ramBufferSize) if (ramBufferSize != IndexWriter.DISABLE_AUTO_FLUSH
&& numBytesUsed > 0.95 * ramBufferSize)
balanceRAM(); balanceRAM();
} }
@ -2137,10 +2150,8 @@ final class DocumentsWriter {
// We must at this point commit to flushing to ensure we // We must at this point commit to flushing to ensure we
// always get N docs when we flush by doc count, even if // always get N docs when we flush by doc count, even if
// > 1 thread is adding documents: // > 1 thread is adding documents:
/* new merge policy if (!flushPending && maxBufferedDocs != IndexWriter.DISABLE_AUTO_FLUSH
if (!flushPending && maxBufferedDocs > 0 && numDocsInRAM >= maxBufferedDocs) { && numDocsInRAM >= maxBufferedDocs) {
*/
if (!flushPending && ramBufferSize == 0 && numDocsInRAM >= maxBufferedDocs) {
flushPending = true; flushPending = true;
state.doFlushAfter = true; state.doFlushAfter = true;
} else } else
@ -2163,8 +2174,12 @@ final class DocumentsWriter {
} }
} }
if (delTerm != null) if (delTerm != null) {
addDeleteTerm(delTerm, state.docID); addDeleteTerm(delTerm, state.docID);
if (!state.doFlushAfter) {
state.doFlushAfter = timeToFlushDeletes();
}
}
return state; return state;
} }
@ -2211,9 +2226,11 @@ final class DocumentsWriter {
} }
// Reset buffered deletes. // Reset buffered deletes.
synchronized void clearBufferedDeleteTerms() { synchronized void clearBufferedDeleteTerms() throws IOException {
bufferedDeleteTerms.clear(); bufferedDeleteTerms.clear();
numBufferedDeleteTerms = 0; numBufferedDeleteTerms = 0;
if (numBytesUsed > 0)
resetPostingsData();
} }
synchronized boolean bufferDeleteTerms(Term[] terms) throws IOException { synchronized boolean bufferDeleteTerms(Term[] terms) throws IOException {
@ -2236,12 +2253,13 @@ final class DocumentsWriter {
} }
synchronized private boolean timeToFlushDeletes() { synchronized private boolean timeToFlushDeletes() {
return numBufferedDeleteTerms >= maxBufferedDeleteTerms && setFlushPending(); return (bufferIsFull
|| (maxBufferedDeleteTerms != IndexWriter.DISABLE_AUTO_FLUSH
&& numBufferedDeleteTerms >= maxBufferedDeleteTerms))
&& setFlushPending();
} }
void setMaxBufferedDeleteTerms(int maxBufferedDeleteTerms) { void setMaxBufferedDeleteTerms(int maxBufferedDeleteTerms) {
if (maxBufferedDeleteTerms < 1)
throw new IllegalArgumentException("maxBufferedDeleteTerms must at least be 1");
this.maxBufferedDeleteTerms = maxBufferedDeleteTerms; this.maxBufferedDeleteTerms = maxBufferedDeleteTerms;
} }
@ -2285,6 +2303,13 @@ final class DocumentsWriter {
Num num = (Num) bufferedDeleteTerms.get(term); Num num = (Num) bufferedDeleteTerms.get(term);
if (num == null) { if (num == null) {
bufferedDeleteTerms.put(term, new Num(docCount)); bufferedDeleteTerms.put(term, new Num(docCount));
// This is coarse approximation of actual bytes used:
numBytesUsed += (term.field().length() + term.text().length()) * BYTES_PER_CHAR
+ Integer.SIZE/8 + 5 * OBJECT_HEADER_BYTES + 5 * OBJECT_POINTER_BYTES;
if (ramBufferSize != IndexWriter.DISABLE_AUTO_FLUSH
&& numBytesUsed > ramBufferSize) {
bufferIsFull = true;
}
} else { } else {
num.setNum(docCount); num.setNum(docCount);
} }
@ -2827,7 +2852,7 @@ final class DocumentsWriter {
* pools to match the current docs. */ * pools to match the current docs. */
private synchronized void balanceRAM() { private synchronized void balanceRAM() {
if (ramBufferSize == 0.0 || postingsIsFull) if (ramBufferSize == IndexWriter.DISABLE_AUTO_FLUSH || bufferIsFull)
return; return;
// We free our allocations if we've allocated 5% over // We free our allocations if we've allocated 5% over
@ -2864,9 +2889,9 @@ final class DocumentsWriter {
while(numBytesAlloc > freeLevel) { while(numBytesAlloc > freeLevel) {
if (0 == freeByteBlocks.size() && 0 == freeCharBlocks.size() && 0 == postingsFreeCount) { if (0 == freeByteBlocks.size() && 0 == freeCharBlocks.size() && 0 == postingsFreeCount) {
// Nothing else to free -- must flush now. // Nothing else to free -- must flush now.
postingsIsFull = true; bufferIsFull = true;
if (infoStream != null) if (infoStream != null)
infoStream.println(" nothing to free; now set postingsIsFull"); infoStream.println(" nothing to free; now set bufferIsFull");
break; break;
} }
@ -2909,7 +2934,7 @@ final class DocumentsWriter {
" allocMB=" + nf.format(numBytesAlloc/1024./1024.) + " allocMB=" + nf.format(numBytesAlloc/1024./1024.) +
" triggerMB=" + nf.format(flushTrigger/1024./1024.)); " triggerMB=" + nf.format(flushTrigger/1024./1024.));
postingsIsFull = true; bufferIsFull = true;
} }
} }
} }

View File

@ -208,7 +208,7 @@ public class IndexModifier {
indexWriter.setMergeScheduler(new SerialMergeScheduler()); indexWriter.setMergeScheduler(new SerialMergeScheduler());
indexWriter.setInfoStream(infoStream); indexWriter.setInfoStream(infoStream);
indexWriter.setUseCompoundFile(useCompoundFile); indexWriter.setUseCompoundFile(useCompoundFile);
if (maxBufferedDocs != 0) if (maxBufferedDocs != IndexWriter.DISABLE_AUTO_FLUSH)
indexWriter.setMaxBufferedDocs(maxBufferedDocs); indexWriter.setMaxBufferedDocs(maxBufferedDocs);
indexWriter.setMaxFieldLength(maxFieldLength); indexWriter.setMaxFieldLength(maxFieldLength);
indexWriter.setMergeFactor(mergeFactor); indexWriter.setMergeFactor(mergeFactor);

View File

@ -37,7 +37,6 @@ import java.util.Set;
import java.util.HashSet; import java.util.HashSet;
import java.util.LinkedList; import java.util.LinkedList;
import java.util.Iterator; import java.util.Iterator;
import java.util.ListIterator;
import java.util.Map.Entry; import java.util.Map.Entry;
/** /**
@ -205,10 +204,15 @@ public class IndexWriter {
public final static int DEFAULT_MERGE_FACTOR = LogMergePolicy.DEFAULT_MERGE_FACTOR; public final static int DEFAULT_MERGE_FACTOR = LogMergePolicy.DEFAULT_MERGE_FACTOR;
/** /**
* Default value is 0 (because IndexWriter flushes by RAM * Value to denote a flush trigger is disabled
* usage by default). Change using {@link #setMaxBufferedDocs(int)}.
*/ */
public final static int DEFAULT_MAX_BUFFERED_DOCS = 0; public final static int DISABLE_AUTO_FLUSH = -1;
/**
* Disabled by default (because IndexWriter flushes by RAM usage
* by default). Change using {@link #setMaxBufferedDocs(int)}.
*/
public final static int DEFAULT_MAX_BUFFERED_DOCS = DISABLE_AUTO_FLUSH;
/** /**
* Default value is 16 MB (which means flush when buffered * Default value is 16 MB (which means flush when buffered
@ -217,9 +221,10 @@ public class IndexWriter {
public final static double DEFAULT_RAM_BUFFER_SIZE_MB = 16.0; public final static double DEFAULT_RAM_BUFFER_SIZE_MB = 16.0;
/** /**
* Default value is 1000. Change using {@link #setMaxBufferedDeleteTerms(int)}. * Disabled by default (because IndexWriter flushes by RAM usage
* by default). Change using {@link #setMaxBufferedDeleteTerms(int)}.
*/ */
public final static int DEFAULT_MAX_BUFFERED_DELETE_TERMS = 1000; public final static int DEFAULT_MAX_BUFFERED_DELETE_TERMS = DISABLE_AUTO_FLUSH;
/** /**
* @deprecated * @deprecated
@ -817,20 +822,28 @@ public class IndexWriter {
* indexing. * indexing.
* *
* <p>When this is set, the writer will flush every * <p>When this is set, the writer will flush every
* maxBufferedDocs added documents and never flush by RAM * maxBufferedDocs added documents. Pass in {@link
* usage.</p> * #DISABLE_AUTO_FLUSH} to prevent triggering a flush due
* to number of buffered documents. Note that if flushing
* by RAM usage is also enabled, then the flush will be
* triggered by whichever comes first.</p>
* *
* <p> The default value is 0 (writer flushes by RAM * <p>Disabled by default (writer flushes by RAM usage).</p>
* usage).</p>
* *
* @throws IllegalArgumentException if maxBufferedDocs is * @throws IllegalArgumentException if maxBufferedDocs is
* smaller than 2 * enabled but smaller than 2, or it disables maxBufferedDocs
* when ramBufferSize is already disabled
* @see #setRAMBufferSizeMB * @see #setRAMBufferSizeMB
*/ */
public void setMaxBufferedDocs(int maxBufferedDocs) { public void setMaxBufferedDocs(int maxBufferedDocs) {
ensureOpen(); ensureOpen();
if (maxBufferedDocs < 2) if (maxBufferedDocs != DISABLE_AUTO_FLUSH && maxBufferedDocs < 2)
throw new IllegalArgumentException("maxBufferedDocs must at least be 2"); throw new IllegalArgumentException(
"maxBufferedDocs must at least be 2 when enabled");
if (maxBufferedDocs == DISABLE_AUTO_FLUSH
&& getRAMBufferSizeMB() == DISABLE_AUTO_FLUSH)
throw new IllegalArgumentException(
"at least one of ramBufferSize and maxBufferedDocs must be enabled");
docWriter.setMaxBufferedDocs(maxBufferedDocs); docWriter.setMaxBufferedDocs(maxBufferedDocs);
pushMaxBufferedDocs(); pushMaxBufferedDocs();
} }
@ -841,7 +854,7 @@ public class IndexWriter {
* as its minMergeDocs, to keep backwards compatibility. * as its minMergeDocs, to keep backwards compatibility.
*/ */
private void pushMaxBufferedDocs() { private void pushMaxBufferedDocs() {
if (docWriter.getRAMBufferSizeMB() == 0.0) { if (docWriter.getMaxBufferedDocs() != DISABLE_AUTO_FLUSH) {
final MergePolicy mp = mergePolicy; final MergePolicy mp = mergePolicy;
if (mp instanceof LogDocMergePolicy) { if (mp instanceof LogDocMergePolicy) {
LogDocMergePolicy lmp = (LogDocMergePolicy) mp; LogDocMergePolicy lmp = (LogDocMergePolicy) mp;
@ -856,9 +869,8 @@ public class IndexWriter {
} }
/** /**
* Returns 0 if this writer is flushing by RAM usage, else * Returns the number of buffered added documents that will
* returns the number of buffered added documents that will * trigger a flush if enabled.
* trigger a flush.
* @see #setMaxBufferedDocs * @see #setMaxBufferedDocs
*/ */
public int getMaxBufferedDocs() { public int getMaxBufferedDocs() {
@ -873,20 +885,30 @@ public class IndexWriter {
* count and use as large a RAM buffer as you can. * count and use as large a RAM buffer as you can.
* *
* <p>When this is set, the writer will flush whenever * <p>When this is set, the writer will flush whenever
* buffered documents use this much RAM.</p> * buffered documents use this much RAM. Pass in {@link
* #DISABLE_AUTO_FLUSH} to prevent triggering a flush due
* to RAM usage. Note that if flushing by document count
* is also enabled, then the flush will be triggered by
* whichever comes first.</p>
* *
* <p> The default value is {@link #DEFAULT_RAM_BUFFER_SIZE_MB}.</p> * <p> The default value is {@link #DEFAULT_RAM_BUFFER_SIZE_MB}.</p>
*
* @throws IllegalArgumentException if ramBufferSize is
* enabled but non-positive, or it disables ramBufferSize
* when maxBufferedDocs is already disabled
*/ */
public void setRAMBufferSizeMB(double mb) { public void setRAMBufferSizeMB(double mb) {
if (mb <= 0.0) if (mb != DISABLE_AUTO_FLUSH && mb <= 0.0)
throw new IllegalArgumentException("ramBufferSize should be > 0.0 MB"); throw new IllegalArgumentException(
"ramBufferSize should be > 0.0 MB when enabled");
if (mb == DISABLE_AUTO_FLUSH && getMaxBufferedDocs() == DISABLE_AUTO_FLUSH)
throw new IllegalArgumentException(
"at least one of ramBufferSize and maxBufferedDocs must be enabled");
docWriter.setRAMBufferSizeMB(mb); docWriter.setRAMBufferSizeMB(mb);
} }
/** /**
* Returns 0.0 if this writer is flushing by document * Returns the value set by {@link #setRAMBufferSizeMB} if enabled.
* count, else returns the value set by {@link
* #setRAMBufferSizeMB}.
*/ */
public double getRAMBufferSizeMB() { public double getRAMBufferSizeMB() {
return docWriter.getRAMBufferSizeMB(); return docWriter.getRAMBufferSizeMB();
@ -898,17 +920,24 @@ public class IndexWriter {
* buffered in memory at the time, they are merged and a new segment is * buffered in memory at the time, they are merged and a new segment is
* created.</p> * created.</p>
* <p>The default value is {@link #DEFAULT_MAX_BUFFERED_DELETE_TERMS}. * <p>Disabled by default (writer flushes by RAM usage).</p>
* @throws IllegalArgumentException if maxBufferedDeleteTerms is smaller than 1</p> *
* @throws IllegalArgumentException if maxBufferedDeleteTerms
* is enabled but smaller than 1
* @see #setRAMBufferSizeMB
*/ */
public void setMaxBufferedDeleteTerms(int maxBufferedDeleteTerms) { public void setMaxBufferedDeleteTerms(int maxBufferedDeleteTerms) {
ensureOpen(); ensureOpen();
if (maxBufferedDeleteTerms != DISABLE_AUTO_FLUSH
&& maxBufferedDeleteTerms < 1)
throw new IllegalArgumentException(
"maxBufferedDeleteTerms must at least be 1 when enabled");
docWriter.setMaxBufferedDeleteTerms(maxBufferedDeleteTerms); docWriter.setMaxBufferedDeleteTerms(maxBufferedDeleteTerms);
} }
/** /**
* Returns the number of buffered deleted terms that will * Returns the number of buffered deleted terms that will
* trigger a flush. * trigger a flush if enabled.
* @see #setMaxBufferedDeleteTerms * @see #setMaxBufferedDeleteTerms
*/ */
public int getMaxBufferedDeleteTerms() { public int getMaxBufferedDeleteTerms() {
@ -1479,13 +1508,7 @@ public class IndexWriter {
} }
} }
/** Determines amount of RAM usage by the buffered docs at
* which point we trigger a flush to the index.
*/
private double ramBufferSize = DEFAULT_RAM_BUFFER_SIZE_MB*1024F*1024F;
/** If non-null, information about merges will be printed to this. /** If non-null, information about merges will be printed to this.
*/ */
private PrintStream infoStream = null; private PrintStream infoStream = null;
private static PrintStream defaultInfoStream = null; private static PrintStream defaultInfoStream = null;

View File

@ -18,8 +18,6 @@ package org.apache.lucene.index;
*/ */
import java.io.IOException; import java.io.IOException;
import java.util.List;
import java.util.ArrayList;
import java.util.Set; import java.util.Set;
import org.apache.lucene.store.Directory; import org.apache.lucene.store.Directory;

View File

@ -18,7 +18,6 @@ package org.apache.lucene.index;
*/ */
import java.io.IOException; import java.io.IOException;
import java.util.LinkedList;
/** A {@link MergeScheduler} that simply does each merge /** A {@link MergeScheduler} that simply does each merge
* sequentially, using the current thread. */ * sequentially, using the current thread. */

View File

@ -22,7 +22,6 @@ import org.apache.lucene.store.IndexOutput;
import org.apache.lucene.util.StringHelper; import org.apache.lucene.util.StringHelper;
import java.io.IOException; import java.io.IOException;
import java.util.Vector;
final class TermVectorsWriter { final class TermVectorsWriter {

View File

@ -75,7 +75,7 @@ public class TestIndexModifier extends TestCase {
// Lucene defaults: // Lucene defaults:
assertNull(i.getInfoStream()); assertNull(i.getInfoStream());
assertTrue(i.getUseCompoundFile()); assertTrue(i.getUseCompoundFile());
assertEquals(0, i.getMaxBufferedDocs()); assertEquals(IndexWriter.DISABLE_AUTO_FLUSH, i.getMaxBufferedDocs());
assertEquals(10000, i.getMaxFieldLength()); assertEquals(10000, i.getMaxFieldLength());
assertEquals(10, i.getMergeFactor()); assertEquals(10, i.getMergeFactor());
// test setting properties: // test setting properties:

View File

@ -1151,8 +1151,8 @@ public class TestIndexWriter extends TestCase
RAMDirectory dir = new RAMDirectory(); RAMDirectory dir = new RAMDirectory();
IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true); IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true);
writer.setMaxBufferedDocs(10); writer.setMaxBufferedDocs(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
int lastNumFile = dir.list().length;
long lastGen = -1; long lastGen = -1;
for(int j=1;j<52;j++) { for(int j=1;j<52;j++) {
Document doc = new Document(); Document doc = new Document();
@ -1169,25 +1169,89 @@ public class TestIndexWriter extends TestCase
assertTrue(gen > lastGen); assertTrue(gen > lastGen);
lastGen = gen; lastGen = gen;
writer.setRAMBufferSizeMB(0.000001); writer.setRAMBufferSizeMB(0.000001);
writer.setMaxBufferedDocs(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (j < 20) { } else if (j < 20) {
assertTrue(gen > lastGen); assertTrue(gen > lastGen);
lastGen = gen; lastGen = gen;
} else if (20 == j) { } else if (20 == j) {
writer.setRAMBufferSizeMB(16); writer.setRAMBufferSizeMB(16);
writer.setMaxBufferedDocs(IndexWriter.DISABLE_AUTO_FLUSH);
lastGen = gen; lastGen = gen;
} else if (j < 30) { } else if (j < 30) {
assertEquals(gen, lastGen); assertEquals(gen, lastGen);
} else if (30 == j) { } else if (30 == j) {
writer.setRAMBufferSizeMB(0.000001); writer.setRAMBufferSizeMB(0.000001);
writer.setMaxBufferedDocs(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (j < 40) { } else if (j < 40) {
assertTrue(gen> lastGen); assertTrue(gen> lastGen);
lastGen = gen; lastGen = gen;
} else if (40 == j) { } else if (40 == j) {
writer.setMaxBufferedDocs(10); writer.setMaxBufferedDocs(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
lastGen = gen; lastGen = gen;
} else if (j < 50) { } else if (j < 50) {
assertEquals(gen, lastGen); assertEquals(gen, lastGen);
writer.setMaxBufferedDocs(10); writer.setMaxBufferedDocs(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (50 == j) {
assertTrue(gen > lastGen);
}
}
writer.close();
dir.close();
}
public void testChangingRAMBuffer2() throws IOException {
RAMDirectory dir = new RAMDirectory();
IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true);
writer.setMaxBufferedDocs(10);
writer.setMaxBufferedDeleteTerms(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
for(int j=1;j<52;j++) {
Document doc = new Document();
doc.add(new Field("field", "aaa" + j, Field.Store.YES, Field.Index.TOKENIZED));
writer.addDocument(doc);
}
long lastGen = -1;
for(int j=1;j<52;j++) {
writer.deleteDocuments(new Term("field", "aaa" + j));
_TestUtil.syncConcurrentMerges(writer);
long gen = SegmentInfos.generationFromSegmentsFileName(SegmentInfos.getCurrentSegmentFileName(dir.list()));
if (j == 1)
lastGen = gen;
else if (j < 10) {
// No new files should be created
assertEquals(gen, lastGen);
} else if (10 == j) {
assertTrue(gen > lastGen);
lastGen = gen;
writer.setRAMBufferSizeMB(0.000001);
writer.setMaxBufferedDeleteTerms(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (j < 20) {
assertTrue(gen > lastGen);
lastGen = gen;
} else if (20 == j) {
writer.setRAMBufferSizeMB(16);
writer.setMaxBufferedDeleteTerms(IndexWriter.DISABLE_AUTO_FLUSH);
lastGen = gen;
} else if (j < 30) {
assertEquals(gen, lastGen);
} else if (30 == j) {
writer.setRAMBufferSizeMB(0.000001);
writer.setMaxBufferedDeleteTerms(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (j < 40) {
assertTrue(gen> lastGen);
lastGen = gen;
} else if (40 == j) {
writer.setMaxBufferedDeleteTerms(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
lastGen = gen;
} else if (j < 50) {
assertEquals(gen, lastGen);
writer.setMaxBufferedDeleteTerms(10);
writer.setRAMBufferSizeMB(IndexWriter.DISABLE_AUTO_FLUSH);
} else if (50 == j) { } else if (50 == j) {
assertTrue(gen > lastGen); assertTrue(gen > lastGen);
} }