diff --git a/lucene/core/src/java/org/apache/lucene/codecs/PostingsReaderBase.java b/lucene/core/src/java/org/apache/lucene/codecs/PostingsReaderBase.java index 39476ed3f65..f656450de4e 100644 --- a/lucene/core/src/java/org/apache/lucene/codecs/PostingsReaderBase.java +++ b/lucene/core/src/java/org/apache/lucene/codecs/PostingsReaderBase.java @@ -37,6 +37,8 @@ import org.apache.lucene.util.Bits; * time. * @lucene.experimental */ +// nocommit mv under blocktree? but ... it's used by others (e.g. block terms) + // TODO: find a better name; this defines the API that the // terms dict impls use to talk to a postings impl. // TermsDict + PostingsReader/WriterBase == PostingsConsumer/Producer diff --git a/lucene/core/src/java/org/apache/lucene/codecs/blocktree/BlockTreeTermsReader.java b/lucene/core/src/java/org/apache/lucene/codecs/blocktree/BlockTreeTermsReader.java index cb7e1aedcaf..2fcb38c25c5 100644 --- a/lucene/core/src/java/org/apache/lucene/codecs/blocktree/BlockTreeTermsReader.java +++ b/lucene/core/src/java/org/apache/lucene/codecs/blocktree/BlockTreeTermsReader.java @@ -89,13 +89,13 @@ import org.apache.lucene.util.fst.Util; public class BlockTreeTermsReader extends FieldsProducer { // Open input to the main terms dict file (_X.tib) - private final IndexInput in; + final IndexInput in; //private static final boolean DEBUG = BlockTreeTermsWriter.DEBUG; // Reads the terms dict entries, to gather state to // produce DocsEnum on demand - private final PostingsReaderBase postingsReader; + final PostingsReaderBase postingsReader; private final TreeMap fields = new TreeMap<>(); @@ -105,7 +105,7 @@ public class BlockTreeTermsReader extends FieldsProducer { /** File offset where the directory starts in the index file. */ private long indexDirOffset; - private String segment; + final String segment; private final int version; @@ -183,7 +183,7 @@ public class BlockTreeTermsReader extends FieldsProducer { } final long indexStartFP = indexIn.readVLong(); FieldReader previous = fields.put(fieldInfo.name, - new FieldReader(fieldInfo, numTerms, rootCode, sumTotalTermFreq, sumDocFreq, docCount, + new FieldReader(this, fieldInfo, numTerms, rootCode, sumTotalTermFreq, sumDocFreq, docCount, indexStartFP, longsSize, indexIn, minTerm, maxTerm)); if (previous != null) { throw new CorruptIndexException("duplicate field: " + fieldInfo.name + " (resource=" + in + ")"); @@ -294,2566 +294,6 @@ public class BlockTreeTermsReader extends FieldsProducer { final Outputs fstOutputs = ByteSequenceOutputs.getSingleton(); final BytesRef NO_OUTPUT = fstOutputs.getNoOutput(); - /** BlockTree's implementation of {@link Terms}. */ - public final class FieldReader extends Terms { - final long numTerms; - final FieldInfo fieldInfo; - final long sumTotalTermFreq; - final long sumDocFreq; - final int docCount; - final long indexStartFP; - final long rootBlockFP; - final BytesRef rootCode; - final BytesRef minTerm; - final BytesRef maxTerm; - final int longsSize; - - private final FST index; - //private boolean DEBUG; - - FieldReader(FieldInfo fieldInfo, long numTerms, BytesRef rootCode, long sumTotalTermFreq, long sumDocFreq, int docCount, - long indexStartFP, int longsSize, IndexInput indexIn, BytesRef minTerm, BytesRef maxTerm) throws IOException { - assert numTerms > 0; - this.fieldInfo = fieldInfo; - //DEBUG = BlockTreeTermsReader.DEBUG && fieldInfo.name.equals("id"); - this.numTerms = numTerms; - this.sumTotalTermFreq = sumTotalTermFreq; - this.sumDocFreq = sumDocFreq; - this.docCount = docCount; - this.indexStartFP = indexStartFP; - this.rootCode = rootCode; - this.longsSize = longsSize; - this.minTerm = minTerm; - this.maxTerm = maxTerm; - // if (DEBUG) { - // System.out.println("BTTR: seg=" + segment + " field=" + fieldInfo.name + " rootBlockCode=" + rootCode + " divisor=" + indexDivisor); - // } - - rootBlockFP = (new ByteArrayDataInput(rootCode.bytes, rootCode.offset, rootCode.length)).readVLong() >>> BlockTreeTermsWriter.OUTPUT_FLAGS_NUM_BITS; - - if (indexIn != null) { - final IndexInput clone = indexIn.clone(); - //System.out.println("start=" + indexStartFP + " field=" + fieldInfo.name); - clone.seek(indexStartFP); - index = new FST<>(clone, ByteSequenceOutputs.getSingleton()); - - /* - if (false) { - final String dotFileName = segment + "_" + fieldInfo.name + ".dot"; - Writer w = new OutputStreamWriter(new FileOutputStream(dotFileName)); - Util.toDot(index, w, false, false); - System.out.println("FST INDEX: SAVED to " + dotFileName); - w.close(); - } - */ - } else { - index = null; - } - } - - @Override - public BytesRef getMin() throws IOException { - if (minTerm == null) { - // Older index that didn't store min/maxTerm - return super.getMin(); - } else { - return minTerm; - } - } - - @Override - public BytesRef getMax() throws IOException { - if (maxTerm == null) { - // Older index that didn't store min/maxTerm - return super.getMax(); - } else { - return maxTerm; - } - } - - /** For debugging -- used by CheckIndex too*/ - // TODO: maybe push this into Terms? - public Stats computeStats() throws IOException { - return new SegmentTermsEnum().computeBlockStats(); - } - - @Override - public boolean hasFreqs() { - return fieldInfo.getIndexOptions().compareTo(IndexOptions.DOCS_AND_FREQS) >= 0; - } - - @Override - public boolean hasOffsets() { - return fieldInfo.getIndexOptions().compareTo(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS_AND_OFFSETS) >= 0; - } - - @Override - public boolean hasPositions() { - return fieldInfo.getIndexOptions().compareTo(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS) >= 0; - } - - @Override - public boolean hasPayloads() { - return fieldInfo.hasPayloads(); - } - - @Override - public TermsEnum iterator(TermsEnum reuse) throws IOException { - return new SegmentTermsEnum(); - } - - @Override - public long size() { - return numTerms; - } - - @Override - public long getSumTotalTermFreq() { - return sumTotalTermFreq; - } - - @Override - public long getSumDocFreq() { - return sumDocFreq; - } - - @Override - public int getDocCount() { - return docCount; - } - - @Override - public TermsEnum intersect(CompiledAutomaton compiled, BytesRef startTerm) throws IOException { - if (compiled.type != CompiledAutomaton.AUTOMATON_TYPE.NORMAL) { - throw new IllegalArgumentException("please use CompiledAutomaton.getTermsEnum instead"); - } - return new IntersectEnum(compiled, startTerm); - } - - /** Returns approximate RAM bytes used */ - public long ramBytesUsed() { - return ((index!=null)? index.sizeInBytes() : 0); - } - - // NOTE: cannot seek! - private final class IntersectEnum extends TermsEnum { - private final IndexInput in; - - private Frame[] stack; - - @SuppressWarnings({"rawtypes","unchecked"}) private FST.Arc[] arcs = new FST.Arc[5]; - - private final RunAutomaton runAutomaton; - private final CompiledAutomaton compiledAutomaton; - - private Frame currentFrame; - - private final BytesRef term = new BytesRef(); - - private final FST.BytesReader fstReader; - - // TODO: can we share this with the frame in STE? - private final class Frame { - final int ord; - long fp; - long fpOrig; - long fpEnd; - long lastSubFP; - - // State in automaton - int state; - - int metaDataUpto; - - byte[] suffixBytes = new byte[128]; - final ByteArrayDataInput suffixesReader = new ByteArrayDataInput(); - - byte[] statBytes = new byte[64]; - final ByteArrayDataInput statsReader = new ByteArrayDataInput(); - - byte[] floorData = new byte[32]; - final ByteArrayDataInput floorDataReader = new ByteArrayDataInput(); - - // Length of prefix shared by all terms in this block - int prefix; - - // Number of entries (term or sub-block) in this block - int entCount; - - // Which term we will next read - int nextEnt; - - // True if this block is either not a floor block, - // or, it's the last sub-block of a floor block - boolean isLastInFloor; - - // True if all entries are terms - boolean isLeafBlock; - - int numFollowFloorBlocks; - int nextFloorLabel; - - Transition[] transitions; - int curTransitionMax; - int transitionIndex; - - FST.Arc arc; - - final BlockTermState termState; - - // metadata buffer, holding monotonic values - public long[] longs; - // metadata buffer, holding general values - public byte[] bytes; - ByteArrayDataInput bytesReader; - - // Cumulative output so far - BytesRef outputPrefix; - - private int startBytePos; - private int suffix; - - public Frame(int ord) throws IOException { - this.ord = ord; - this.termState = postingsReader.newTermState(); - this.termState.totalTermFreq = -1; - this.longs = new long[longsSize]; - } - - void loadNextFloorBlock() throws IOException { - assert numFollowFloorBlocks > 0; - //if (DEBUG) System.out.println(" loadNextFoorBlock trans=" + transitions[transitionIndex]); - - do { - fp = fpOrig + (floorDataReader.readVLong() >>> 1); - numFollowFloorBlocks--; - // if (DEBUG) System.out.println(" skip floor block2! nextFloorLabel=" + (char) nextFloorLabel + " vs target=" + (char) transitions[transitionIndex].getMin() + " newFP=" + fp + " numFollowFloorBlocks=" + numFollowFloorBlocks); - if (numFollowFloorBlocks != 0) { - nextFloorLabel = floorDataReader.readByte() & 0xff; - } else { - nextFloorLabel = 256; - } - // if (DEBUG) System.out.println(" nextFloorLabel=" + (char) nextFloorLabel); - } while (numFollowFloorBlocks != 0 && nextFloorLabel <= transitions[transitionIndex].getMin()); - - load(null); - } - - public void setState(int state) { - this.state = state; - transitionIndex = 0; - transitions = compiledAutomaton.sortedTransitions[state]; - if (transitions.length != 0) { - curTransitionMax = transitions[0].getMax(); - } else { - curTransitionMax = -1; - } - } - - void load(BytesRef frameIndexData) throws IOException { - - // if (DEBUG) System.out.println(" load fp=" + fp + " fpOrig=" + fpOrig + " frameIndexData=" + frameIndexData + " trans=" + (transitions.length != 0 ? transitions[0] : "n/a" + " state=" + state)); - - if (frameIndexData != null && transitions.length != 0) { - // Floor frame - if (floorData.length < frameIndexData.length) { - this.floorData = new byte[ArrayUtil.oversize(frameIndexData.length, 1)]; - } - System.arraycopy(frameIndexData.bytes, frameIndexData.offset, floorData, 0, frameIndexData.length); - floorDataReader.reset(floorData, 0, frameIndexData.length); - // Skip first long -- has redundant fp, hasTerms - // flag, isFloor flag - final long code = floorDataReader.readVLong(); - if ((code & BlockTreeTermsWriter.OUTPUT_FLAG_IS_FLOOR) != 0) { - numFollowFloorBlocks = floorDataReader.readVInt(); - nextFloorLabel = floorDataReader.readByte() & 0xff; - // if (DEBUG) System.out.println(" numFollowFloorBlocks=" + numFollowFloorBlocks + " nextFloorLabel=" + nextFloorLabel); - - // If current state is accept, we must process - // first block in case it has empty suffix: - if (!runAutomaton.isAccept(state)) { - // Maybe skip floor blocks: - while (numFollowFloorBlocks != 0 && nextFloorLabel <= transitions[0].getMin()) { - fp = fpOrig + (floorDataReader.readVLong() >>> 1); - numFollowFloorBlocks--; - // if (DEBUG) System.out.println(" skip floor block! nextFloorLabel=" + (char) nextFloorLabel + " vs target=" + (char) transitions[0].getMin() + " newFP=" + fp + " numFollowFloorBlocks=" + numFollowFloorBlocks); - if (numFollowFloorBlocks != 0) { - nextFloorLabel = floorDataReader.readByte() & 0xff; - } else { - nextFloorLabel = 256; - } - } - } - } - } - - in.seek(fp); - int code = in.readVInt(); - entCount = code >>> 1; - assert entCount > 0; - isLastInFloor = (code & 1) != 0; - - // term suffixes: - code = in.readVInt(); - isLeafBlock = (code & 1) != 0; - int numBytes = code >>> 1; - // if (DEBUG) System.out.println(" entCount=" + entCount + " lastInFloor?=" + isLastInFloor + " leafBlock?=" + isLeafBlock + " numSuffixBytes=" + numBytes); - if (suffixBytes.length < numBytes) { - suffixBytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(suffixBytes, 0, numBytes); - suffixesReader.reset(suffixBytes, 0, numBytes); - - // stats - numBytes = in.readVInt(); - if (statBytes.length < numBytes) { - statBytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(statBytes, 0, numBytes); - statsReader.reset(statBytes, 0, numBytes); - metaDataUpto = 0; - - termState.termBlockOrd = 0; - nextEnt = 0; - - // metadata - numBytes = in.readVInt(); - if (bytes == null) { - bytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - bytesReader = new ByteArrayDataInput(); - } else if (bytes.length < numBytes) { - bytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(bytes, 0, numBytes); - bytesReader.reset(bytes, 0, numBytes); - - if (!isLastInFloor) { - // Sub-blocks of a single floor block are always - // written one after another -- tail recurse: - fpEnd = in.getFilePointer(); - } - } - - // TODO: maybe add scanToLabel; should give perf boost - - public boolean next() { - return isLeafBlock ? nextLeaf() : nextNonLeaf(); - } - - // Decodes next entry; returns true if it's a sub-block - public boolean nextLeaf() { - //if (DEBUG) System.out.println(" frame.next ord=" + ord + " nextEnt=" + nextEnt + " entCount=" + entCount); - assert nextEnt != -1 && nextEnt < entCount: "nextEnt=" + nextEnt + " entCount=" + entCount + " fp=" + fp; - nextEnt++; - suffix = suffixesReader.readVInt(); - startBytePos = suffixesReader.getPosition(); - suffixesReader.skipBytes(suffix); - return false; - } - - public boolean nextNonLeaf() { - //if (DEBUG) System.out.println(" frame.next ord=" + ord + " nextEnt=" + nextEnt + " entCount=" + entCount); - assert nextEnt != -1 && nextEnt < entCount: "nextEnt=" + nextEnt + " entCount=" + entCount + " fp=" + fp; - nextEnt++; - final int code = suffixesReader.readVInt(); - suffix = code >>> 1; - startBytePos = suffixesReader.getPosition(); - suffixesReader.skipBytes(suffix); - if ((code & 1) == 0) { - // A normal term - termState.termBlockOrd++; - return false; - } else { - // A sub-block; make sub-FP absolute: - lastSubFP = fp - suffixesReader.readVLong(); - return true; - } - } - - public int getTermBlockOrd() { - return isLeafBlock ? nextEnt : termState.termBlockOrd; - } - - public void decodeMetaData() throws IOException { - - // lazily catch up on metadata decode: - final int limit = getTermBlockOrd(); - boolean absolute = metaDataUpto == 0; - assert limit > 0; - - // TODO: better API would be "jump straight to term=N"??? - while (metaDataUpto < limit) { - - // TODO: we could make "tiers" of metadata, ie, - // decode docFreq/totalTF but don't decode postings - // metadata; this way caller could get - // docFreq/totalTF w/o paying decode cost for - // postings - - // TODO: if docFreq were bulk decoded we could - // just skipN here: - - // stats - termState.docFreq = statsReader.readVInt(); - //if (DEBUG) System.out.println(" dF=" + state.docFreq); - if (fieldInfo.getIndexOptions() != IndexOptions.DOCS_ONLY) { - termState.totalTermFreq = termState.docFreq + statsReader.readVLong(); - //if (DEBUG) System.out.println(" totTF=" + state.totalTermFreq); - } - // metadata - for (int i = 0; i < longsSize; i++) { - longs[i] = bytesReader.readVLong(); - } - postingsReader.decodeTerm(longs, bytesReader, fieldInfo, termState, absolute); - - metaDataUpto++; - absolute = false; - } - termState.termBlockOrd = metaDataUpto; - } - } - - private BytesRef savedStartTerm; - - // TODO: in some cases we can filter by length? eg - // regexp foo*bar must be at least length 6 bytes - public IntersectEnum(CompiledAutomaton compiled, BytesRef startTerm) throws IOException { - // if (DEBUG) { - // System.out.println("\nintEnum.init seg=" + segment + " commonSuffix=" + brToString(compiled.commonSuffixRef)); - // } - runAutomaton = compiled.runAutomaton; - compiledAutomaton = compiled; - in = BlockTreeTermsReader.this.in.clone(); - stack = new Frame[5]; - for(int idx=0;idx(); - } - - if (index == null) { - fstReader = null; - } else { - fstReader = index.getBytesReader(); - } - - // TODO: if the automaton is "smallish" we really - // should use the terms index to seek at least to - // the initial term and likely to subsequent terms - // (or, maybe just fallback to ATE for such cases). - // Else the seek cost of loading the frames will be - // too costly. - - final FST.Arc arc = index.getFirstArc(arcs[0]); - // Empty string prefix must have an output in the index! - assert arc.isFinal(); - - // Special pushFrame since it's the first one: - final Frame f = stack[0]; - f.fp = f.fpOrig = rootBlockFP; - f.prefix = 0; - f.setState(runAutomaton.getInitialState()); - f.arc = arc; - f.outputPrefix = arc.output; - f.load(rootCode); - - // for assert: - assert setSavedStartTerm(startTerm); - - currentFrame = f; - if (startTerm != null) { - seekToStartTerm(startTerm); - } - } - - // only for assert: - private boolean setSavedStartTerm(BytesRef startTerm) { - savedStartTerm = startTerm == null ? null : BytesRef.deepCopyOf(startTerm); - return true; - } - - @Override - public TermState termState() throws IOException { - currentFrame.decodeMetaData(); - return currentFrame.termState.clone(); - } - - private Frame getFrame(int ord) throws IOException { - if (ord >= stack.length) { - final Frame[] next = new Frame[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; - System.arraycopy(stack, 0, next, 0, stack.length); - for(int stackOrd=stack.length;stackOrd getArc(int ord) { - if (ord >= arcs.length) { - @SuppressWarnings({"rawtypes","unchecked"}) final FST.Arc[] next = - new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; - System.arraycopy(arcs, 0, next, 0, arcs.length); - for(int arcOrd=arcs.length;arcOrd(); - } - arcs = next; - } - return arcs[ord]; - } - - private Frame pushFrame(int state) throws IOException { - final Frame f = getFrame(currentFrame == null ? 0 : 1+currentFrame.ord); - - f.fp = f.fpOrig = currentFrame.lastSubFP; - f.prefix = currentFrame.prefix + currentFrame.suffix; - // if (DEBUG) System.out.println(" pushFrame state=" + state + " prefix=" + f.prefix); - f.setState(state); - - // Walk the arc through the index -- we only - // "bother" with this so we can get the floor data - // from the index and skip floor blocks when - // possible: - FST.Arc arc = currentFrame.arc; - int idx = currentFrame.prefix; - assert currentFrame.suffix > 0; - BytesRef output = currentFrame.outputPrefix; - while (idx < f.prefix) { - final int target = term.bytes[idx] & 0xff; - // TODO: we could be more efficient for the next() - // case by using current arc as starting point, - // passed to findTargetArc - arc = index.findTargetArc(target, arc, getArc(1+idx), fstReader); - assert arc != null; - output = fstOutputs.add(output, arc.output); - idx++; - } - - f.arc = arc; - f.outputPrefix = output; - assert arc.isFinal(); - f.load(fstOutputs.add(output, arc.nextFinalOutput)); - return f; - } - - @Override - public BytesRef term() { - return term; - } - - @Override - public int docFreq() throws IOException { - //if (DEBUG) System.out.println("BTIR.docFreq"); - currentFrame.decodeMetaData(); - //if (DEBUG) System.out.println(" return " + currentFrame.termState.docFreq); - return currentFrame.termState.docFreq; - } - - @Override - public long totalTermFreq() throws IOException { - currentFrame.decodeMetaData(); - return currentFrame.termState.totalTermFreq; - } - - @Override - public DocsEnum docs(Bits skipDocs, DocsEnum reuse, int flags) throws IOException { - currentFrame.decodeMetaData(); - return postingsReader.docs(fieldInfo, currentFrame.termState, skipDocs, reuse, flags); - } - - @Override - public DocsAndPositionsEnum docsAndPositions(Bits skipDocs, DocsAndPositionsEnum reuse, int flags) throws IOException { - if (fieldInfo.getIndexOptions().compareTo(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS) < 0) { - // Positions were not indexed: - return null; - } - - currentFrame.decodeMetaData(); - return postingsReader.docsAndPositions(fieldInfo, currentFrame.termState, skipDocs, reuse, flags); - } - - private int getState() { - int state = currentFrame.state; - for(int idx=0;idx 0) { - // A prefix of the common suffix overlaps with - // the suffix of the block prefix so we first - // test whether the prefix part matches: - final byte[] termBytes = term.bytes; - int termBytesPos = currentFrame.prefix - lenInPrefix; - assert termBytesPos >= 0; - final int termBytesPosEnd = currentFrame.prefix; - while (termBytesPos < termBytesPosEnd) { - if (termBytes[termBytesPos++] != commonSuffixBytes[commonSuffixBytesPos++]) { - // if (DEBUG) { - // System.out.println(" skip: common suffix mismatch (in prefix)"); - // } - continue nextTerm; - } - } - suffixBytesPos = currentFrame.startBytePos; - } else { - suffixBytesPos = currentFrame.startBytePos + currentFrame.suffix - compiledAutomaton.commonSuffixRef.length; - } - - // Test overlapping suffix part: - final int commonSuffixBytesPosEnd = compiledAutomaton.commonSuffixRef.length; - while (commonSuffixBytesPos < commonSuffixBytesPosEnd) { - if (suffixBytes[suffixBytesPos++] != commonSuffixBytes[commonSuffixBytesPos++]) { - // if (DEBUG) { - // System.out.println(" skip: common suffix mismatch"); - // } - continue nextTerm; - } - } - } - - // TODO: maybe we should do the same linear test - // that AutomatonTermsEnum does, so that if we - // reach a part of the automaton where .* is - // "temporarily" accepted, we just blindly .next() - // until the limit - - // See if the term prefix matches the automaton: - int state = currentFrame.state; - for (int idx=0;idx[] arcs = - new FST.Arc[1]; - - public SegmentTermsEnum() throws IOException { - //if (DEBUG) System.out.println("BTTR.init seg=" + segment); - stack = new Frame[0]; - - // Used to hold seek by TermState, or cached seek - staticFrame = new Frame(-1); - - if (index == null) { - fstReader = null; - } else { - fstReader = index.getBytesReader(); - } - - // Init w/ root block; don't use index since it may - // not (and need not) have been loaded - for(int arcIdx=0;arcIdx(); - } - - currentFrame = staticFrame; - final FST.Arc arc; - if (index != null) { - arc = index.getFirstArc(arcs[0]); - // Empty string prefix must have an output in the index! - assert arc.isFinal(); - } else { - arc = null; - } - currentFrame = staticFrame; - //currentFrame = pushFrame(arc, rootCode, 0); - //currentFrame.loadBlock(); - validIndexPrefix = 0; - // if (DEBUG) { - // System.out.println("init frame state " + currentFrame.ord); - // printSeekState(); - // } - - //System.out.println(); - // computeBlockStats().print(System.out); - } - - // Not private to avoid synthetic access$NNN methods - void initIndexInput() { - if (this.in == null) { - this.in = BlockTreeTermsReader.this.in.clone(); - } - } - - /** Runs next() through the entire terms dict, - * computing aggregate statistics. */ - public Stats computeBlockStats() throws IOException { - - Stats stats = new Stats(segment, fieldInfo.name); - if (index != null) { - stats.indexNodeCount = index.getNodeCount(); - stats.indexArcCount = index.getArcCount(); - stats.indexNumBytes = index.sizeInBytes(); - } - - currentFrame = staticFrame; - FST.Arc arc; - if (index != null) { - arc = index.getFirstArc(arcs[0]); - // Empty string prefix must have an output in the index! - assert arc.isFinal(); - } else { - arc = null; - } - - // Empty string prefix must have an output in the - // index! - currentFrame = pushFrame(arc, rootCode, 0); - currentFrame.fpOrig = currentFrame.fp; - currentFrame.loadBlock(); - validIndexPrefix = 0; - - stats.startBlock(currentFrame, !currentFrame.isLastInFloor); - - allTerms: - while (true) { - - // Pop finished blocks - while (currentFrame.nextEnt == currentFrame.entCount) { - stats.endBlock(currentFrame); - if (!currentFrame.isLastInFloor) { - currentFrame.loadNextFloorBlock(); - stats.startBlock(currentFrame, true); - } else { - if (currentFrame.ord == 0) { - break allTerms; - } - final long lastFP = currentFrame.fpOrig; - currentFrame = stack[currentFrame.ord-1]; - assert lastFP == currentFrame.lastSubFP; - // if (DEBUG) { - // System.out.println(" reset validIndexPrefix=" + validIndexPrefix); - // } - } - } - - while(true) { - if (currentFrame.next()) { - // Push to new block: - currentFrame = pushFrame(null, currentFrame.lastSubFP, term.length); - currentFrame.fpOrig = currentFrame.fp; - // This is a "next" frame -- even if it's - // floor'd we must pretend it isn't so we don't - // try to scan to the right floor frame: - currentFrame.isFloor = false; - //currentFrame.hasTerms = true; - currentFrame.loadBlock(); - stats.startBlock(currentFrame, !currentFrame.isLastInFloor); - } else { - stats.term(term); - break; - } - } - } - - stats.finish(); - - // Put root frame back: - currentFrame = staticFrame; - if (index != null) { - arc = index.getFirstArc(arcs[0]); - // Empty string prefix must have an output in the index! - assert arc.isFinal(); - } else { - arc = null; - } - currentFrame = pushFrame(arc, rootCode, 0); - currentFrame.rewind(); - currentFrame.loadBlock(); - validIndexPrefix = 0; - term.length = 0; - - return stats; - } - - private Frame getFrame(int ord) throws IOException { - if (ord >= stack.length) { - final Frame[] next = new Frame[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; - System.arraycopy(stack, 0, next, 0, stack.length); - for(int stackOrd=stack.length;stackOrd getArc(int ord) { - if (ord >= arcs.length) { - @SuppressWarnings({"rawtypes","unchecked"}) final FST.Arc[] next = - new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; - System.arraycopy(arcs, 0, next, 0, arcs.length); - for(int arcOrd=arcs.length;arcOrd(); - } - arcs = next; - } - return arcs[ord]; - } - - // Pushes a frame we seek'd to - Frame pushFrame(FST.Arc arc, BytesRef frameData, int length) throws IOException { - scratchReader.reset(frameData.bytes, frameData.offset, frameData.length); - final long code = scratchReader.readVLong(); - final long fpSeek = code >>> BlockTreeTermsWriter.OUTPUT_FLAGS_NUM_BITS; - final Frame f = getFrame(1+currentFrame.ord); - f.hasTerms = (code & BlockTreeTermsWriter.OUTPUT_FLAG_HAS_TERMS) != 0; - f.hasTermsOrig = f.hasTerms; - f.isFloor = (code & BlockTreeTermsWriter.OUTPUT_FLAG_IS_FLOOR) != 0; - if (f.isFloor) { - f.setFloorData(scratchReader, frameData); - } - pushFrame(arc, fpSeek, length); - - return f; - } - - // Pushes next'd frame or seek'd frame; we later - // lazy-load the frame only when needed - Frame pushFrame(FST.Arc arc, long fp, int length) throws IOException { - final Frame f = getFrame(1+currentFrame.ord); - f.arc = arc; - if (f.fpOrig == fp && f.nextEnt != -1) { - //if (DEBUG) System.out.println(" push reused frame ord=" + f.ord + " fp=" + f.fp + " isFloor?=" + f.isFloor + " hasTerms=" + f.hasTerms + " pref=" + term + " nextEnt=" + f.nextEnt + " targetBeforeCurrentLength=" + targetBeforeCurrentLength + " term.length=" + term.length + " vs prefix=" + f.prefix); - if (f.prefix > targetBeforeCurrentLength) { - f.rewind(); - } else { - // if (DEBUG) { - // System.out.println(" skip rewind!"); - // } - } - assert length == f.prefix; - } else { - f.nextEnt = -1; - f.prefix = length; - f.state.termBlockOrd = 0; - f.fpOrig = f.fp = fp; - f.lastSubFP = -1; - // if (DEBUG) { - // final int sav = term.length; - // term.length = length; - // System.out.println(" push new frame ord=" + f.ord + " fp=" + f.fp + " hasTerms=" + f.hasTerms + " isFloor=" + f.isFloor + " pref=" + brToString(term)); - // term.length = sav; - // } - } - - return f; - } - - // asserts only - private boolean clearEOF() { - eof = false; - return true; - } - - // asserts only - private boolean setEOF() { - eof = true; - return true; - } - - // nocommit we need a seekExact(BytesRef target, long minVersion) API? - - @Override - public boolean seekExact(final BytesRef target) throws IOException { - - if (index == null) { - throw new IllegalStateException("terms index was not loaded"); - } - - if (term.bytes.length <= target.length) { - term.bytes = ArrayUtil.grow(term.bytes, 1+target.length); - } - - assert clearEOF(); - - // if (DEBUG) { - // System.out.println("\nBTTR.seekExact seg=" + segment + " target=" + fieldInfo.name + ":" + brToString(target) + " current=" + brToString(term) + " (exists?=" + termExists + ") validIndexPrefix=" + validIndexPrefix); - // printSeekState(); - // } - - FST.Arc arc; - int targetUpto; - BytesRef output; - - targetBeforeCurrentLength = currentFrame.ord; - - if (currentFrame != staticFrame) { - - // We are already seek'd; find the common - // prefix of new seek term vs current term and - // re-use the corresponding seek state. For - // example, if app first seeks to foobar, then - // seeks to foobaz, we can re-use the seek state - // for the first 5 bytes. - - // if (DEBUG) { - // System.out.println(" re-use current seek state validIndexPrefix=" + validIndexPrefix); - // } - - arc = arcs[0]; - assert arc.isFinal(); - output = arc.output; - targetUpto = 0; - - Frame lastFrame = stack[0]; - assert validIndexPrefix <= term.length; - - final int targetLimit = Math.min(target.length, validIndexPrefix); - - int cmp = 0; - - // TODO: reverse vLong byte order for better FST - // prefix output sharing - - // First compare up to valid seek frames: - while (targetUpto < targetLimit) { - cmp = (term.bytes[targetUpto]&0xFF) - (target.bytes[target.offset + targetUpto]&0xFF); - // if (DEBUG) { - // System.out.println(" cycle targetUpto=" + targetUpto + " (vs limit=" + targetLimit + ") cmp=" + cmp + " (targetLabel=" + (char) (target.bytes[target.offset + targetUpto]) + " vs termLabel=" + (char) (term.bytes[targetUpto]) + ")" + " arc.output=" + arc.output + " output=" + output); - // } - if (cmp != 0) { - break; - } - arc = arcs[1+targetUpto]; - //if (arc.label != (target.bytes[target.offset + targetUpto] & 0xFF)) { - //System.out.println("FAIL: arc.label=" + (char) arc.label + " targetLabel=" + (char) (target.bytes[target.offset + targetUpto] & 0xFF)); - //} - assert arc.label == (target.bytes[target.offset + targetUpto] & 0xFF): "arc.label=" + (char) arc.label + " targetLabel=" + (char) (target.bytes[target.offset + targetUpto] & 0xFF); - if (arc.output != NO_OUTPUT) { - output = fstOutputs.add(output, arc.output); - } - if (arc.isFinal()) { - lastFrame = stack[1+lastFrame.ord]; - } - targetUpto++; - } - - if (cmp == 0) { - final int targetUptoMid = targetUpto; - - // Second compare the rest of the term, but - // don't save arc/output/frame; we only do this - // to find out if the target term is before, - // equal or after the current term - final int targetLimit2 = Math.min(target.length, term.length); - while (targetUpto < targetLimit2) { - cmp = (term.bytes[targetUpto]&0xFF) - (target.bytes[target.offset + targetUpto]&0xFF); - // if (DEBUG) { - // System.out.println(" cycle2 targetUpto=" + targetUpto + " (vs limit=" + targetLimit + ") cmp=" + cmp + " (targetLabel=" + (char) (target.bytes[target.offset + targetUpto]) + " vs termLabel=" + (char) (term.bytes[targetUpto]) + ")"); - // } - if (cmp != 0) { - break; - } - targetUpto++; - } - - if (cmp == 0) { - cmp = term.length - target.length; - } - targetUpto = targetUptoMid; - } - - if (cmp < 0) { - // Common case: target term is after current - // term, ie, app is seeking multiple terms - // in sorted order - // if (DEBUG) { - // System.out.println(" target is after current (shares prefixLen=" + targetUpto + "); frame.ord=" + lastFrame.ord); - // } - currentFrame = lastFrame; - - } else if (cmp > 0) { - // Uncommon case: target term - // is before current term; this means we can - // keep the currentFrame but we must rewind it - // (so we scan from the start) - targetBeforeCurrentLength = 0; - // if (DEBUG) { - // System.out.println(" target is before current (shares prefixLen=" + targetUpto + "); rewind frame ord=" + lastFrame.ord); - // } - currentFrame = lastFrame; - currentFrame.rewind(); - } else { - // Target is exactly the same as current term - assert term.length == target.length; - if (termExists) { - // if (DEBUG) { - // System.out.println(" target is same as current; return true"); - // } - return true; - } else { - // if (DEBUG) { - // System.out.println(" target is same as current but term doesn't exist"); - // } - } - //validIndexPrefix = currentFrame.depth; - //term.length = target.length; - //return termExists; - } - - } else { - - targetBeforeCurrentLength = -1; - arc = index.getFirstArc(arcs[0]); - - // Empty string prefix must have an output (block) in the index! - assert arc.isFinal(); - assert arc.output != null; - - // if (DEBUG) { - // System.out.println(" no seek state; push root frame"); - // } - - output = arc.output; - - currentFrame = staticFrame; - - //term.length = 0; - targetUpto = 0; - currentFrame = pushFrame(arc, fstOutputs.add(output, arc.nextFinalOutput), 0); - } - - // if (DEBUG) { - // System.out.println(" start index loop targetUpto=" + targetUpto + " output=" + output + " currentFrame.ord=" + currentFrame.ord + " targetBeforeCurrentLength=" + targetBeforeCurrentLength); - // } - - while (targetUpto < target.length) { - - final int targetLabel = target.bytes[target.offset + targetUpto] & 0xFF; - - final FST.Arc nextArc = index.findTargetArc(targetLabel, arc, getArc(1+targetUpto), fstReader); - - if (nextArc == null) { - - // Index is exhausted - // if (DEBUG) { - // System.out.println(" index: index exhausted label=" + ((char) targetLabel) + " " + toHex(targetLabel)); - // } - - validIndexPrefix = currentFrame.prefix; - //validIndexPrefix = targetUpto; - - currentFrame.scanToFloorFrame(target); - - if (!currentFrame.hasTerms) { - termExists = false; - term.bytes[targetUpto] = (byte) targetLabel; - term.length = 1+targetUpto; - // if (DEBUG) { - // System.out.println(" FAST NOT_FOUND term=" + brToString(term)); - // } - return false; - } - - currentFrame.loadBlock(); - - final SeekStatus result = currentFrame.scanToTerm(target, true); - if (result == SeekStatus.FOUND) { - // if (DEBUG) { - // System.out.println(" return FOUND term=" + term.utf8ToString() + " " + term); - // } - return true; - } else { - // if (DEBUG) { - // System.out.println(" got " + result + "; return NOT_FOUND term=" + brToString(term)); - // } - return false; - } - } else { - // Follow this arc - arc = nextArc; - term.bytes[targetUpto] = (byte) targetLabel; - // Aggregate output as we go: - assert arc.output != null; - if (arc.output != NO_OUTPUT) { - output = fstOutputs.add(output, arc.output); - } - - // if (DEBUG) { - // System.out.println(" index: follow label=" + toHex(target.bytes[target.offset + targetUpto]&0xff) + " arc.output=" + arc.output + " arc.nfo=" + arc.nextFinalOutput); - // } - targetUpto++; - - if (arc.isFinal()) { - //if (DEBUG) System.out.println(" arc is final!"); - currentFrame = pushFrame(arc, fstOutputs.add(output, arc.nextFinalOutput), targetUpto); - //if (DEBUG) System.out.println(" curFrame.ord=" + currentFrame.ord + " hasTerms=" + currentFrame.hasTerms); - } - } - } - - //validIndexPrefix = targetUpto; - validIndexPrefix = currentFrame.prefix; - - currentFrame.scanToFloorFrame(target); - - // Target term is entirely contained in the index: - if (!currentFrame.hasTerms) { - termExists = false; - term.length = targetUpto; - // if (DEBUG) { - // System.out.println(" FAST NOT_FOUND term=" + brToString(term)); - // } - return false; - } - - currentFrame.loadBlock(); - - final SeekStatus result = currentFrame.scanToTerm(target, true); - if (result == SeekStatus.FOUND) { - // if (DEBUG) { - // System.out.println(" return FOUND term=" + term.utf8ToString() + " " + term); - // } - return true; - } else { - // if (DEBUG) { - // System.out.println(" got result " + result + "; return NOT_FOUND term=" + term.utf8ToString()); - // } - - return false; - } - } - - @Override - public SeekStatus seekCeil(final BytesRef target) throws IOException { - if (index == null) { - throw new IllegalStateException("terms index was not loaded"); - } - - if (term.bytes.length <= target.length) { - term.bytes = ArrayUtil.grow(term.bytes, 1+target.length); - } - - assert clearEOF(); - - //if (DEBUG) { - //System.out.println("\nBTTR.seekCeil seg=" + segment + " target=" + fieldInfo.name + ":" + target.utf8ToString() + " " + target + " current=" + brToString(term) + " (exists?=" + termExists + ") validIndexPrefix= " + validIndexPrefix); - //printSeekState(); - //} - - FST.Arc arc; - int targetUpto; - BytesRef output; - - targetBeforeCurrentLength = currentFrame.ord; - - if (currentFrame != staticFrame) { - - // We are already seek'd; find the common - // prefix of new seek term vs current term and - // re-use the corresponding seek state. For - // example, if app first seeks to foobar, then - // seeks to foobaz, we can re-use the seek state - // for the first 5 bytes. - - //if (DEBUG) { - //System.out.println(" re-use current seek state validIndexPrefix=" + validIndexPrefix); - //} - - arc = arcs[0]; - assert arc.isFinal(); - output = arc.output; - targetUpto = 0; - - Frame lastFrame = stack[0]; - assert validIndexPrefix <= term.length; - - final int targetLimit = Math.min(target.length, validIndexPrefix); - - int cmp = 0; - - // TOOD: we should write our vLong backwards (MSB - // first) to get better sharing from the FST - - // First compare up to valid seek frames: - while (targetUpto < targetLimit) { - cmp = (term.bytes[targetUpto]&0xFF) - (target.bytes[target.offset + targetUpto]&0xFF); - //if (DEBUG) { - //System.out.println(" cycle targetUpto=" + targetUpto + " (vs limit=" + targetLimit + ") cmp=" + cmp + " (targetLabel=" + (char) (target.bytes[target.offset + targetUpto]) + " vs termLabel=" + (char) (term.bytes[targetUpto]) + ")" + " arc.output=" + arc.output + " output=" + output); - //} - if (cmp != 0) { - break; - } - arc = arcs[1+targetUpto]; - assert arc.label == (target.bytes[target.offset + targetUpto] & 0xFF): "arc.label=" + (char) arc.label + " targetLabel=" + (char) (target.bytes[target.offset + targetUpto] & 0xFF); - // TOOD: we could save the outputs in local - // byte[][] instead of making new objs ever - // seek; but, often the FST doesn't have any - // shared bytes (but this could change if we - // reverse vLong byte order) - if (arc.output != NO_OUTPUT) { - output = fstOutputs.add(output, arc.output); - } - if (arc.isFinal()) { - lastFrame = stack[1+lastFrame.ord]; - } - targetUpto++; - } - - - if (cmp == 0) { - final int targetUptoMid = targetUpto; - // Second compare the rest of the term, but - // don't save arc/output/frame: - final int targetLimit2 = Math.min(target.length, term.length); - while (targetUpto < targetLimit2) { - cmp = (term.bytes[targetUpto]&0xFF) - (target.bytes[target.offset + targetUpto]&0xFF); - //if (DEBUG) { - //System.out.println(" cycle2 targetUpto=" + targetUpto + " (vs limit=" + targetLimit + ") cmp=" + cmp + " (targetLabel=" + (char) (target.bytes[target.offset + targetUpto]) + " vs termLabel=" + (char) (term.bytes[targetUpto]) + ")"); - //} - if (cmp != 0) { - break; - } - targetUpto++; - } - - if (cmp == 0) { - cmp = term.length - target.length; - } - targetUpto = targetUptoMid; - } - - if (cmp < 0) { - // Common case: target term is after current - // term, ie, app is seeking multiple terms - // in sorted order - //if (DEBUG) { - //System.out.println(" target is after current (shares prefixLen=" + targetUpto + "); clear frame.scanned ord=" + lastFrame.ord); - //} - currentFrame = lastFrame; - - } else if (cmp > 0) { - // Uncommon case: target term - // is before current term; this means we can - // keep the currentFrame but we must rewind it - // (so we scan from the start) - targetBeforeCurrentLength = 0; - //if (DEBUG) { - //System.out.println(" target is before current (shares prefixLen=" + targetUpto + "); rewind frame ord=" + lastFrame.ord); - //} - currentFrame = lastFrame; - currentFrame.rewind(); - } else { - // Target is exactly the same as current term - assert term.length == target.length; - if (termExists) { - //if (DEBUG) { - //System.out.println(" target is same as current; return FOUND"); - //} - return SeekStatus.FOUND; - } else { - //if (DEBUG) { - //System.out.println(" target is same as current but term doesn't exist"); - //} - } - } - - } else { - - targetBeforeCurrentLength = -1; - arc = index.getFirstArc(arcs[0]); - - // Empty string prefix must have an output (block) in the index! - assert arc.isFinal(); - assert arc.output != null; - - //if (DEBUG) { - //System.out.println(" no seek state; push root frame"); - //} - - output = arc.output; - - currentFrame = staticFrame; - - //term.length = 0; - targetUpto = 0; - currentFrame = pushFrame(arc, fstOutputs.add(output, arc.nextFinalOutput), 0); - } - - //if (DEBUG) { - //System.out.println(" start index loop targetUpto=" + targetUpto + " output=" + output + " currentFrame.ord+1=" + currentFrame.ord + " targetBeforeCurrentLength=" + targetBeforeCurrentLength); - //} - - while (targetUpto < target.length) { - - final int targetLabel = target.bytes[target.offset + targetUpto] & 0xFF; - - final FST.Arc nextArc = index.findTargetArc(targetLabel, arc, getArc(1+targetUpto), fstReader); - - if (nextArc == null) { - - // Index is exhausted - // if (DEBUG) { - // System.out.println(" index: index exhausted label=" + ((char) targetLabel) + " " + toHex(targetLabel)); - // } - - validIndexPrefix = currentFrame.prefix; - //validIndexPrefix = targetUpto; - - currentFrame.scanToFloorFrame(target); - - currentFrame.loadBlock(); - - final SeekStatus result = currentFrame.scanToTerm(target, false); - if (result == SeekStatus.END) { - term.copyBytes(target); - termExists = false; - - if (next() != null) { - //if (DEBUG) { - //System.out.println(" return NOT_FOUND term=" + brToString(term) + " " + term); - //} - return SeekStatus.NOT_FOUND; - } else { - //if (DEBUG) { - //System.out.println(" return END"); - //} - return SeekStatus.END; - } - } else { - //if (DEBUG) { - //System.out.println(" return " + result + " term=" + brToString(term) + " " + term); - //} - return result; - } - } else { - // Follow this arc - term.bytes[targetUpto] = (byte) targetLabel; - arc = nextArc; - // Aggregate output as we go: - assert arc.output != null; - if (arc.output != NO_OUTPUT) { - output = fstOutputs.add(output, arc.output); - } - - //if (DEBUG) { - //System.out.println(" index: follow label=" + toHex(target.bytes[target.offset + targetUpto]&0xff) + " arc.output=" + arc.output + " arc.nfo=" + arc.nextFinalOutput); - //} - targetUpto++; - - if (arc.isFinal()) { - //if (DEBUG) System.out.println(" arc is final!"); - currentFrame = pushFrame(arc, fstOutputs.add(output, arc.nextFinalOutput), targetUpto); - //if (DEBUG) System.out.println(" curFrame.ord=" + currentFrame.ord + " hasTerms=" + currentFrame.hasTerms); - } - } - } - - //validIndexPrefix = targetUpto; - validIndexPrefix = currentFrame.prefix; - - currentFrame.scanToFloorFrame(target); - - currentFrame.loadBlock(); - - final SeekStatus result = currentFrame.scanToTerm(target, false); - - if (result == SeekStatus.END) { - term.copyBytes(target); - termExists = false; - if (next() != null) { - //if (DEBUG) { - //System.out.println(" return NOT_FOUND term=" + term.utf8ToString() + " " + term); - //} - return SeekStatus.NOT_FOUND; - } else { - //if (DEBUG) { - //System.out.println(" return END"); - //} - return SeekStatus.END; - } - } else { - return result; - } - } - - @SuppressWarnings("unused") - private void printSeekState(PrintStream out) throws IOException { - if (currentFrame == staticFrame) { - out.println(" no prior seek"); - } else { - out.println(" prior seek state:"); - int ord = 0; - boolean isSeekFrame = true; - while(true) { - Frame f = getFrame(ord); - assert f != null; - final BytesRef prefix = new BytesRef(term.bytes, 0, f.prefix); - if (f.nextEnt == -1) { - out.println(" frame " + (isSeekFrame ? "(seek)" : "(next)") + " ord=" + ord + " fp=" + f.fp + (f.isFloor ? (" (fpOrig=" + f.fpOrig + ")") : "") + " prefixLen=" + f.prefix + " prefix=" + prefix + (f.nextEnt == -1 ? "" : (" (of " + f.entCount + ")")) + " hasTerms=" + f.hasTerms + " isFloor=" + f.isFloor + " code=" + ((f.fp< arc; - if (index != null) { - arc = index.getFirstArc(arcs[0]); - // Empty string prefix must have an output in the index! - assert arc.isFinal(); - } else { - arc = null; - } - currentFrame = pushFrame(arc, rootCode, 0); - currentFrame.loadBlock(); - } - - targetBeforeCurrentLength = currentFrame.ord; - - assert !eof; - //if (DEBUG) { - //System.out.println("\nBTTR.next seg=" + segment + " term=" + brToString(term) + " termExists?=" + termExists + " field=" + fieldInfo.name + " termBlockOrd=" + currentFrame.state.termBlockOrd + " validIndexPrefix=" + validIndexPrefix); - //printSeekState(); - //} - - if (currentFrame == staticFrame) { - // If seek was previously called and the term was - // cached, or seek(TermState) was called, usually - // caller is just going to pull a D/&PEnum or get - // docFreq, etc. But, if they then call next(), - // this method catches up all internal state so next() - // works properly: - //if (DEBUG) System.out.println(" re-seek to pending term=" + term.utf8ToString() + " " + term); - final boolean result = seekExact(term); - assert result; - } - - // Pop finished blocks - while (currentFrame.nextEnt == currentFrame.entCount) { - if (!currentFrame.isLastInFloor) { - currentFrame.loadNextFloorBlock(); - } else { - //if (DEBUG) System.out.println(" pop frame"); - if (currentFrame.ord == 0) { - //if (DEBUG) System.out.println(" return null"); - assert setEOF(); - term.length = 0; - validIndexPrefix = 0; - currentFrame.rewind(); - termExists = false; - return null; - } - final long lastFP = currentFrame.fpOrig; - currentFrame = stack[currentFrame.ord-1]; - - if (currentFrame.nextEnt == -1 || currentFrame.lastSubFP != lastFP) { - // We popped into a frame that's not loaded - // yet or not scan'd to the right entry - currentFrame.scanToFloorFrame(term); - currentFrame.loadBlock(); - currentFrame.scanToSubBlock(lastFP); - } - - // Note that the seek state (last seek) has been - // invalidated beyond this depth - validIndexPrefix = Math.min(validIndexPrefix, currentFrame.prefix); - //if (DEBUG) { - //System.out.println(" reset validIndexPrefix=" + validIndexPrefix); - //} - } - } - - while(true) { - if (currentFrame.next()) { - // Push to new block: - //if (DEBUG) System.out.println(" push frame"); - currentFrame = pushFrame(null, currentFrame.lastSubFP, term.length); - // This is a "next" frame -- even if it's - // floor'd we must pretend it isn't so we don't - // try to scan to the right floor frame: - currentFrame.isFloor = false; - //currentFrame.hasTerms = true; - currentFrame.loadBlock(); - } else { - //if (DEBUG) System.out.println(" return term=" + term.utf8ToString() + " " + term + " currentFrame.ord=" + currentFrame.ord); - return term; - } - } - } - - @Override - public BytesRef term() { - assert !eof; - return term; - } - - @Override - public int docFreq() throws IOException { - assert !eof; - //if (DEBUG) System.out.println("BTR.docFreq"); - currentFrame.decodeMetaData(); - //if (DEBUG) System.out.println(" return " + currentFrame.state.docFreq); - return currentFrame.state.docFreq; - } - - @Override - public long totalTermFreq() throws IOException { - assert !eof; - currentFrame.decodeMetaData(); - return currentFrame.state.totalTermFreq; - } - - @Override - public DocsEnum docs(Bits skipDocs, DocsEnum reuse, int flags) throws IOException { - assert !eof; - //if (DEBUG) { - //System.out.println("BTTR.docs seg=" + segment); - //} - currentFrame.decodeMetaData(); - //if (DEBUG) { - //System.out.println(" state=" + currentFrame.state); - //} - return postingsReader.docs(fieldInfo, currentFrame.state, skipDocs, reuse, flags); - } - - @Override - public DocsAndPositionsEnum docsAndPositions(Bits skipDocs, DocsAndPositionsEnum reuse, int flags) throws IOException { - if (fieldInfo.getIndexOptions().compareTo(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS) < 0) { - // Positions were not indexed: - return null; - } - - assert !eof; - currentFrame.decodeMetaData(); - return postingsReader.docsAndPositions(fieldInfo, currentFrame.state, skipDocs, reuse, flags); - } - - @Override - public void seekExact(BytesRef target, TermState otherState) { - // if (DEBUG) { - // System.out.println("BTTR.seekExact termState seg=" + segment + " target=" + target.utf8ToString() + " " + target + " state=" + otherState); - // } - assert clearEOF(); - if (target.compareTo(term) != 0 || !termExists) { - assert otherState != null && otherState instanceof BlockTermState; - currentFrame = staticFrame; - currentFrame.state.copyFrom(otherState); - term.copyBytes(target); - currentFrame.metaDataUpto = currentFrame.getTermBlockOrd(); - assert currentFrame.metaDataUpto > 0; - validIndexPrefix = 0; - } else { - // if (DEBUG) { - // System.out.println(" skip seek: already on target state=" + currentFrame.state); - // } - } - } - - @Override - public TermState termState() throws IOException { - assert !eof; - currentFrame.decodeMetaData(); - TermState ts = currentFrame.state.clone(); - //if (DEBUG) System.out.println("BTTR.termState seg=" + segment + " state=" + ts); - return ts; - } - - @Override - public void seekExact(long ord) { - throw new UnsupportedOperationException(); - } - - @Override - public long ord() { - throw new UnsupportedOperationException(); - } - - // Not static -- references term, postingsReader, - // fieldInfo, in - final class Frame { - // Our index in stack[]: - final int ord; - - boolean hasTerms; - boolean hasTermsOrig; - boolean isFloor; - - FST.Arc arc; - - // File pointer where this block was loaded from - long fp; - long fpOrig; - long fpEnd; - - byte[] suffixBytes = new byte[128]; - final ByteArrayDataInput suffixesReader = new ByteArrayDataInput(); - - byte[] statBytes = new byte[64]; - final ByteArrayDataInput statsReader = new ByteArrayDataInput(); - - byte[] floorData = new byte[32]; - final ByteArrayDataInput floorDataReader = new ByteArrayDataInput(); - - // Length of prefix shared by all terms in this block - int prefix; - - // Number of entries (term or sub-block) in this block - int entCount; - - // Which term we will next read, or -1 if the block - // isn't loaded yet - int nextEnt; - - // True if this block is either not a floor block, - // or, it's the last sub-block of a floor block - boolean isLastInFloor; - - // True if all entries are terms - boolean isLeafBlock; - - long lastSubFP; - - int nextFloorLabel; - int numFollowFloorBlocks; - - // Next term to decode metaData; we decode metaData - // lazily so that scanning to find the matching term is - // fast and only if you find a match and app wants the - // stats or docs/positions enums, will we decode the - // metaData - int metaDataUpto; - - final BlockTermState state; - - // metadata buffer, holding monotonic values - public long[] longs; - // metadata buffer, holding general values - public byte[] bytes; - ByteArrayDataInput bytesReader; - - public Frame(int ord) throws IOException { - this.ord = ord; - this.state = postingsReader.newTermState(); - this.state.totalTermFreq = -1; - this.longs = new long[longsSize]; - } - - public void setFloorData(ByteArrayDataInput in, BytesRef source) { - final int numBytes = source.length - (in.getPosition() - source.offset); - if (numBytes > floorData.length) { - floorData = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - System.arraycopy(source.bytes, source.offset+in.getPosition(), floorData, 0, numBytes); - floorDataReader.reset(floorData, 0, numBytes); - numFollowFloorBlocks = floorDataReader.readVInt(); - nextFloorLabel = floorDataReader.readByte() & 0xff; - //if (DEBUG) { - //System.out.println(" setFloorData fpOrig=" + fpOrig + " bytes=" + new BytesRef(source.bytes, source.offset + in.getPosition(), numBytes) + " numFollowFloorBlocks=" + numFollowFloorBlocks + " nextFloorLabel=" + toHex(nextFloorLabel)); - //} - } - - public int getTermBlockOrd() { - return isLeafBlock ? nextEnt : state.termBlockOrd; - } - - void loadNextFloorBlock() throws IOException { - //if (DEBUG) { - //System.out.println(" loadNextFloorBlock fp=" + fp + " fpEnd=" + fpEnd); - //} - assert arc == null || isFloor: "arc=" + arc + " isFloor=" + isFloor; - fp = fpEnd; - nextEnt = -1; - loadBlock(); - } - - /* Does initial decode of next block of terms; this - doesn't actually decode the docFreq, totalTermFreq, - postings details (frq/prx offset, etc.) metadata; - it just loads them as byte[] blobs which are then - decoded on-demand if the metadata is ever requested - for any term in this block. This enables terms-only - intensive consumes (eg certain MTQs, respelling) to - not pay the price of decoding metadata they won't - use. */ - void loadBlock() throws IOException { - - // Clone the IndexInput lazily, so that consumers - // that just pull a TermsEnum to - // seekExact(TermState) don't pay this cost: - initIndexInput(); - - if (nextEnt != -1) { - // Already loaded - return; - } - //System.out.println("blc=" + blockLoadCount); - - in.seek(fp); - int code = in.readVInt(); - entCount = code >>> 1; - assert entCount > 0; - isLastInFloor = (code & 1) != 0; - assert arc == null || (isLastInFloor || isFloor); - - // TODO: if suffixes were stored in random-access - // array structure, then we could do binary search - // instead of linear scan to find target term; eg - // we could have simple array of offsets - - // term suffixes: - code = in.readVInt(); - isLeafBlock = (code & 1) != 0; - int numBytes = code >>> 1; - if (suffixBytes.length < numBytes) { - suffixBytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(suffixBytes, 0, numBytes); - suffixesReader.reset(suffixBytes, 0, numBytes); - - /*if (DEBUG) { - if (arc == null) { - System.out.println(" loadBlock (next) fp=" + fp + " entCount=" + entCount + " prefixLen=" + prefix + " isLastInFloor=" + isLastInFloor + " leaf?=" + isLeafBlock); - } else { - System.out.println(" loadBlock (seek) fp=" + fp + " entCount=" + entCount + " prefixLen=" + prefix + " hasTerms?=" + hasTerms + " isFloor?=" + isFloor + " isLastInFloor=" + isLastInFloor + " leaf?=" + isLeafBlock); - } - }*/ - - // stats - numBytes = in.readVInt(); - if (statBytes.length < numBytes) { - statBytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(statBytes, 0, numBytes); - statsReader.reset(statBytes, 0, numBytes); - metaDataUpto = 0; - - state.termBlockOrd = 0; - nextEnt = 0; - lastSubFP = -1; - - // TODO: we could skip this if !hasTerms; but - // that's rare so won't help much - // metadata - numBytes = in.readVInt(); - if (bytes == null) { - bytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - bytesReader = new ByteArrayDataInput(); - } else if (bytes.length < numBytes) { - bytes = new byte[ArrayUtil.oversize(numBytes, 1)]; - } - in.readBytes(bytes, 0, numBytes); - bytesReader.reset(bytes, 0, numBytes); - - - // Sub-blocks of a single floor block are always - // written one after another -- tail recurse: - fpEnd = in.getFilePointer(); - // if (DEBUG) { - // System.out.println(" fpEnd=" + fpEnd); - // } - } - - void rewind() { - - // Force reload: - fp = fpOrig; - nextEnt = -1; - hasTerms = hasTermsOrig; - if (isFloor) { - floorDataReader.rewind(); - numFollowFloorBlocks = floorDataReader.readVInt(); - nextFloorLabel = floorDataReader.readByte() & 0xff; - } - - /* - //System.out.println("rewind"); - // Keeps the block loaded, but rewinds its state: - if (nextEnt > 0 || fp != fpOrig) { - if (DEBUG) { - System.out.println(" rewind frame ord=" + ord + " fpOrig=" + fpOrig + " fp=" + fp + " hasTerms?=" + hasTerms + " isFloor?=" + isFloor + " nextEnt=" + nextEnt + " prefixLen=" + prefix); - } - if (fp != fpOrig) { - fp = fpOrig; - nextEnt = -1; - } else { - nextEnt = 0; - } - hasTerms = hasTermsOrig; - if (isFloor) { - floorDataReader.rewind(); - numFollowFloorBlocks = floorDataReader.readVInt(); - nextFloorLabel = floorDataReader.readByte() & 0xff; - } - assert suffixBytes != null; - suffixesReader.rewind(); - assert statBytes != null; - statsReader.rewind(); - metaDataUpto = 0; - state.termBlockOrd = 0; - // TODO: skip this if !hasTerms? Then postings - // impl wouldn't have to write useless 0 byte - postingsReader.resetTermsBlock(fieldInfo, state); - lastSubFP = -1; - } else if (DEBUG) { - System.out.println(" skip rewind fp=" + fp + " fpOrig=" + fpOrig + " nextEnt=" + nextEnt + " ord=" + ord); - } - */ - } - - public boolean next() { - return isLeafBlock ? nextLeaf() : nextNonLeaf(); - } - - // Decodes next entry; returns true if it's a sub-block - public boolean nextLeaf() { - //if (DEBUG) System.out.println(" frame.next ord=" + ord + " nextEnt=" + nextEnt + " entCount=" + entCount); - assert nextEnt != -1 && nextEnt < entCount: "nextEnt=" + nextEnt + " entCount=" + entCount + " fp=" + fp; - nextEnt++; - suffix = suffixesReader.readVInt(); - startBytePos = suffixesReader.getPosition(); - term.length = prefix + suffix; - if (term.bytes.length < term.length) { - term.grow(term.length); - } - suffixesReader.readBytes(term.bytes, prefix, suffix); - // A normal term - termExists = true; - return false; - } - - public boolean nextNonLeaf() { - //if (DEBUG) System.out.println(" frame.next ord=" + ord + " nextEnt=" + nextEnt + " entCount=" + entCount); - assert nextEnt != -1 && nextEnt < entCount: "nextEnt=" + nextEnt + " entCount=" + entCount + " fp=" + fp; - nextEnt++; - final int code = suffixesReader.readVInt(); - suffix = code >>> 1; - startBytePos = suffixesReader.getPosition(); - term.length = prefix + suffix; - if (term.bytes.length < term.length) { - term.grow(term.length); - } - suffixesReader.readBytes(term.bytes, prefix, suffix); - if ((code & 1) == 0) { - // A normal term - termExists = true; - subCode = 0; - state.termBlockOrd++; - return false; - } else { - // A sub-block; make sub-FP absolute: - termExists = false; - subCode = suffixesReader.readVLong(); - lastSubFP = fp - subCode; - //if (DEBUG) { - //System.out.println(" lastSubFP=" + lastSubFP); - //} - return true; - } - } - - // TODO: make this array'd so we can do bin search? - // likely not worth it? need to measure how many - // floor blocks we "typically" get - public void scanToFloorFrame(BytesRef target) { - - if (!isFloor || target.length <= prefix) { - // if (DEBUG) { - // System.out.println(" scanToFloorFrame skip: isFloor=" + isFloor + " target.length=" + target.length + " vs prefix=" + prefix); - // } - return; - } - - final int targetLabel = target.bytes[target.offset + prefix] & 0xFF; - - // if (DEBUG) { - // System.out.println(" scanToFloorFrame fpOrig=" + fpOrig + " targetLabel=" + toHex(targetLabel) + " vs nextFloorLabel=" + toHex(nextFloorLabel) + " numFollowFloorBlocks=" + numFollowFloorBlocks); - // } - - if (targetLabel < nextFloorLabel) { - // if (DEBUG) { - // System.out.println(" already on correct block"); - // } - return; - } - - assert numFollowFloorBlocks != 0; - - long newFP = fpOrig; - while (true) { - final long code = floorDataReader.readVLong(); - newFP = fpOrig + (code >>> 1); - hasTerms = (code & 1) != 0; - // if (DEBUG) { - // System.out.println(" label=" + toHex(nextFloorLabel) + " fp=" + newFP + " hasTerms?=" + hasTerms + " numFollowFloor=" + numFollowFloorBlocks); - // } - - isLastInFloor = numFollowFloorBlocks == 1; - numFollowFloorBlocks--; - - if (isLastInFloor) { - nextFloorLabel = 256; - // if (DEBUG) { - // System.out.println(" stop! last block nextFloorLabel=" + toHex(nextFloorLabel)); - // } - break; - } else { - nextFloorLabel = floorDataReader.readByte() & 0xff; - if (targetLabel < nextFloorLabel) { - // if (DEBUG) { - // System.out.println(" stop! nextFloorLabel=" + toHex(nextFloorLabel)); - // } - break; - } - } - } - - if (newFP != fp) { - // Force re-load of the block: - // if (DEBUG) { - // System.out.println(" force switch to fp=" + newFP + " oldFP=" + fp); - // } - nextEnt = -1; - fp = newFP; - } else { - // if (DEBUG) { - // System.out.println(" stay on same fp=" + newFP); - // } - } - } - - public void decodeMetaData() throws IOException { - - //if (DEBUG) System.out.println("\nBTTR.decodeMetadata seg=" + segment + " mdUpto=" + metaDataUpto + " vs termBlockOrd=" + state.termBlockOrd); - - // lazily catch up on metadata decode: - final int limit = getTermBlockOrd(); - boolean absolute = metaDataUpto == 0; - assert limit > 0; - - // TODO: better API would be "jump straight to term=N"??? - while (metaDataUpto < limit) { - - // TODO: we could make "tiers" of metadata, ie, - // decode docFreq/totalTF but don't decode postings - // metadata; this way caller could get - // docFreq/totalTF w/o paying decode cost for - // postings - - // TODO: if docFreq were bulk decoded we could - // just skipN here: - - // stats - state.docFreq = statsReader.readVInt(); - //if (DEBUG) System.out.println(" dF=" + state.docFreq); - if (fieldInfo.getIndexOptions() != IndexOptions.DOCS_ONLY) { - state.totalTermFreq = state.docFreq + statsReader.readVLong(); - //if (DEBUG) System.out.println(" totTF=" + state.totalTermFreq); - } - // metadata - for (int i = 0; i < longsSize; i++) { - longs[i] = bytesReader.readVLong(); - } - postingsReader.decodeTerm(longs, bytesReader, fieldInfo, state, absolute); - - metaDataUpto++; - absolute = false; - } - state.termBlockOrd = metaDataUpto; - } - - // Used only by assert - private boolean prefixMatches(BytesRef target) { - for(int bytePos=0;bytePos();