LUCENE-3735: Fix PayloadProcessorProvider to no longer use Directory for lookup, instead AtomicReader

git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1239052 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Uwe Schindler 2012-02-01 10:51:40 +00:00
parent 5f38919e81
commit f252c064f2
7 changed files with 32 additions and 30 deletions

View File

@ -134,8 +134,8 @@ public abstract class TermsConsumer {
// set PayloadProcessor // set PayloadProcessor
if (mergeState.payloadProcessorProvider != null) { if (mergeState.payloadProcessorProvider != null) {
for (int i = 0; i < mergeState.readers.size(); i++) { for (int i = 0; i < mergeState.readers.size(); i++) {
if (mergeState.dirPayloadProcessor[i] != null) { if (mergeState.readerPayloadProcessor[i] != null) {
mergeState.currentPayloadProcessor[i] = mergeState.dirPayloadProcessor[i].getProcessor(mergeState.fieldInfo.name, term); mergeState.currentPayloadProcessor[i] = mergeState.readerPayloadProcessor[i].getProcessor(mergeState.fieldInfo.name, term);
} }
} }
} }
@ -168,8 +168,8 @@ public abstract class TermsConsumer {
// set PayloadProcessor // set PayloadProcessor
if (mergeState.payloadProcessorProvider != null) { if (mergeState.payloadProcessorProvider != null) {
for (int i = 0; i < mergeState.readers.size(); i++) { for (int i = 0; i < mergeState.readers.size(); i++) {
if (mergeState.dirPayloadProcessor[i] != null) { if (mergeState.readerPayloadProcessor[i] != null) {
mergeState.currentPayloadProcessor[i] = mergeState.dirPayloadProcessor[i].getProcessor(mergeState.fieldInfo.name, term); mergeState.currentPayloadProcessor[i] = mergeState.readerPayloadProcessor[i].getProcessor(mergeState.fieldInfo.name, term);
} }
} }
} }

View File

@ -38,7 +38,6 @@ import org.apache.lucene.index.DocumentsWriterPerThread.FlushedSegment;
import org.apache.lucene.index.FieldInfos.FieldNumberBiMap; import org.apache.lucene.index.FieldInfos.FieldNumberBiMap;
import org.apache.lucene.index.IndexWriterConfig.OpenMode; import org.apache.lucene.index.IndexWriterConfig.OpenMode;
import org.apache.lucene.index.MergeState.CheckAbort; import org.apache.lucene.index.MergeState.CheckAbort;
import org.apache.lucene.index.PayloadProcessorProvider.DirPayloadProcessor;
import org.apache.lucene.search.Query; import org.apache.lucene.search.Query;
import org.apache.lucene.store.AlreadyClosedException; import org.apache.lucene.store.AlreadyClosedException;
import org.apache.lucene.store.CompoundFileDirectory; import org.apache.lucene.store.CompoundFileDirectory;

View File

@ -19,7 +19,7 @@ package org.apache.lucene.index;
import java.util.List; import java.util.List;
import org.apache.lucene.index.PayloadProcessorProvider.DirPayloadProcessor; import org.apache.lucene.index.PayloadProcessorProvider.ReaderPayloadProcessor;
import org.apache.lucene.index.PayloadProcessorProvider.PayloadProcessor; import org.apache.lucene.index.PayloadProcessorProvider.PayloadProcessor;
import org.apache.lucene.store.Directory; import org.apache.lucene.store.Directory;
import org.apache.lucene.util.Bits; import org.apache.lucene.util.Bits;
@ -55,7 +55,7 @@ public class MergeState {
// TODO: this is a FactoryFactory here basically // TODO: this is a FactoryFactory here basically
// and we could make a codec(wrapper) to do all of this privately so IW is uninvolved // and we could make a codec(wrapper) to do all of this privately so IW is uninvolved
public PayloadProcessorProvider payloadProcessorProvider; public PayloadProcessorProvider payloadProcessorProvider;
public DirPayloadProcessor[] dirPayloadProcessor; public ReaderPayloadProcessor[] readerPayloadProcessor;
public PayloadProcessor[] currentPayloadProcessor; public PayloadProcessor[] currentPayloadProcessor;
// TODO: get rid of this? it tells you which segments are 'aligned' (e.g. for bulk merging) // TODO: get rid of this? it tells you which segments are 'aligned' (e.g. for bulk merging)

View File

@ -52,7 +52,7 @@ public abstract class PayloadProcessorProvider {
* concurrency issues, then you shouldn't worry about any such issues when * concurrency issues, then you shouldn't worry about any such issues when
* {@link PayloadProcessor}s are requested for different terms. * {@link PayloadProcessor}s are requested for different terms.
*/ */
public static abstract class DirPayloadProcessor { public static abstract class ReaderPayloadProcessor {
/** Returns a {@link PayloadProcessor} for the given term. */ /** Returns a {@link PayloadProcessor} for the given term. */
public abstract PayloadProcessor getProcessor(String field, BytesRef text) throws IOException; public abstract PayloadProcessor getProcessor(String field, BytesRef text) throws IOException;
@ -76,6 +76,6 @@ public abstract class PayloadProcessorProvider {
* through which {@link PayloadProcessor}s can be obtained for each * through which {@link PayloadProcessor}s can be obtained for each
* {@link Term}, or <code>null</code> if none should be used. * {@link Term}, or <code>null</code> if none should be used.
*/ */
public abstract DirPayloadProcessor getDirProcessor(Directory dir) throws IOException; public abstract ReaderPayloadProcessor getReaderProcessor(AtomicReader reader) throws IOException;
} }

View File

@ -286,7 +286,7 @@ final class SegmentMerger {
// Remap docIDs // Remap docIDs
mergeState.docMaps = new int[numReaders][]; mergeState.docMaps = new int[numReaders][];
mergeState.docBase = new int[numReaders]; mergeState.docBase = new int[numReaders];
mergeState.dirPayloadProcessor = new PayloadProcessorProvider.DirPayloadProcessor[numReaders]; mergeState.readerPayloadProcessor = new PayloadProcessorProvider.ReaderPayloadProcessor[numReaders];
mergeState.currentPayloadProcessor = new PayloadProcessorProvider.PayloadProcessor[numReaders]; mergeState.currentPayloadProcessor = new PayloadProcessorProvider.PayloadProcessor[numReaders];
int docBase = 0; int docBase = 0;
@ -323,12 +323,7 @@ final class SegmentMerger {
docBase += docCount; docBase += docCount;
if (mergeState.payloadProcessorProvider != null) { if (mergeState.payloadProcessorProvider != null) {
// TODO: the PayloadProcessorProvider should take AtomicReader as parameter mergeState.readerPayloadProcessor[i] = mergeState.payloadProcessorProvider.getReaderProcessor(reader.reader);
// and find out by itself if it can provide a processor:
if (!(reader.reader instanceof SegmentReader))
throw new UnsupportedOperationException("Payload processing currently requires exclusively SegmentReaders to be merged.");
final Directory dir = ((SegmentReader) reader.reader).directory();
mergeState.dirPayloadProcessor[i] = mergeState.payloadProcessorProvider.getDirProcessor(dir);
} }
i++; i++;

View File

@ -30,7 +30,7 @@ import org.apache.lucene.analysis.tokenattributes.PayloadAttribute;
import org.apache.lucene.document.Document; import org.apache.lucene.document.Document;
import org.apache.lucene.document.FieldType; import org.apache.lucene.document.FieldType;
import org.apache.lucene.document.TextField; import org.apache.lucene.document.TextField;
import org.apache.lucene.index.PayloadProcessorProvider.DirPayloadProcessor; import org.apache.lucene.index.PayloadProcessorProvider.ReaderPayloadProcessor;
import org.apache.lucene.index.PayloadProcessorProvider.PayloadProcessor; import org.apache.lucene.index.PayloadProcessorProvider.PayloadProcessor;
import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.DocIdSetIterator;
import org.apache.lucene.store.Directory; import org.apache.lucene.store.Directory;
@ -42,20 +42,24 @@ public class TestPayloadProcessorProvider extends LuceneTestCase {
private static final class PerDirPayloadProcessor extends PayloadProcessorProvider { private static final class PerDirPayloadProcessor extends PayloadProcessorProvider {
private Map<Directory, DirPayloadProcessor> processors; private final Map<Directory, ReaderPayloadProcessor> processors;
public PerDirPayloadProcessor(Map<Directory, DirPayloadProcessor> processors) { public PerDirPayloadProcessor(Map<Directory, ReaderPayloadProcessor> processors) {
this.processors = processors; this.processors = processors;
} }
@Override @Override
public DirPayloadProcessor getDirProcessor(Directory dir) throws IOException { public ReaderPayloadProcessor getReaderProcessor(AtomicReader reader) throws IOException {
return processors.get(dir); if (reader instanceof SegmentReader) {
return processors.get(((SegmentReader) reader).directory());
} else {
throw new UnsupportedOperationException("This shouldnot happen in this test: Reader is no SegmentReader");
}
} }
} }
private static final class PerTermPayloadProcessor extends DirPayloadProcessor { private static final class PerTermPayloadProcessor extends ReaderPayloadProcessor {
@Override @Override
public PayloadProcessor getProcessor(String field, BytesRef text) throws IOException { public PayloadProcessor getProcessor(String field, BytesRef text) throws IOException {
@ -185,7 +189,7 @@ public class TestPayloadProcessorProvider extends LuceneTestCase {
// Add two source dirs. By not adding the dest dir, we ensure its payloads // Add two source dirs. By not adding the dest dir, we ensure its payloads
// won't get processed. // won't get processed.
Map<Directory, DirPayloadProcessor> processors = new HashMap<Directory, DirPayloadProcessor>(); Map<Directory, ReaderPayloadProcessor> processors = new HashMap<Directory, ReaderPayloadProcessor>();
for (Directory d : dirs) { for (Directory d : dirs) {
processors.put(d, new PerTermPayloadProcessor()); processors.put(d, new PerTermPayloadProcessor());
} }
@ -241,7 +245,7 @@ public class TestPayloadProcessorProvider extends LuceneTestCase {
// Add two source dirs. By not adding the dest dir, we ensure its payloads // Add two source dirs. By not adding the dest dir, we ensure its payloads
// won't get processed. // won't get processed.
Map<Directory, DirPayloadProcessor> processors = new HashMap<Directory, DirPayloadProcessor>(); Map<Directory, ReaderPayloadProcessor> processors = new HashMap<Directory, ReaderPayloadProcessor>();
processors.put(dir, new PerTermPayloadProcessor()); processors.put(dir, new PerTermPayloadProcessor());
IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false))); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)));
writer.setPayloadProcessorProvider(new PerDirPayloadProcessor(processors)); writer.setPayloadProcessorProvider(new PerDirPayloadProcessor(processors));

View File

@ -7,7 +7,9 @@ import java.io.InputStream;
import java.util.HashMap; import java.util.HashMap;
import java.util.Map; import java.util.Map;
import org.apache.lucene.index.AtomicReader;
import org.apache.lucene.index.PayloadProcessorProvider; import org.apache.lucene.index.PayloadProcessorProvider;
import org.apache.lucene.index.SegmentReader;
import org.apache.lucene.index.Term; import org.apache.lucene.index.Term;
import org.apache.lucene.store.Directory; import org.apache.lucene.store.Directory;
@ -94,7 +96,7 @@ public class FacetsPayloadProcessorProvider extends PayloadProcessorProvider {
private final Directory workDir; private final Directory workDir;
private final DirPayloadProcessor dirProcessor; private final ReaderPayloadProcessor dirProcessor;
/** /**
* Construct FacetsPayloadProcessorProvider with FacetIndexingParams * Construct FacetsPayloadProcessorProvider with FacetIndexingParams
@ -110,14 +112,16 @@ public class FacetsPayloadProcessorProvider extends PayloadProcessorProvider {
} }
@Override @Override
public DirPayloadProcessor getDirProcessor(Directory dir) throws IOException { public ReaderPayloadProcessor getReaderProcessor(AtomicReader reader) throws IOException {
if (workDir != dir) { if (reader instanceof SegmentReader) {
return null; if (workDir == ((SegmentReader) reader).directory()) {
}
return dirProcessor; return dirProcessor;
} }
}
return null;
}
public static class FacetsDirPayloadProcessor extends DirPayloadProcessor { public static class FacetsDirPayloadProcessor extends ReaderPayloadProcessor {
private final Map<Term, CategoryListParams> termMap = new HashMap<Term, CategoryListParams>(1); private final Map<Term, CategoryListParams> termMap = new HashMap<Term, CategoryListParams>(1);