mirror of https://github.com/apache/lucene.git
LUCENE-5666: get solr compiling
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/branches/lucene5666@1593856 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
639f03475a
commit
9714faa314
|
@ -33,7 +33,6 @@ import org.apache.lucene.queries.function.valuesource.DoubleFieldSource;
|
|||
import org.apache.lucene.queries.function.valuesource.FloatFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.IntFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.LongFieldSource;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.solr.analytics.expression.ExpressionFactory;
|
||||
import org.apache.solr.analytics.request.ExpressionRequest;
|
||||
import org.apache.solr.analytics.util.AnalyticsParams;
|
||||
|
@ -357,7 +356,7 @@ public class StatsCollectorSupplierFactory {
|
|||
if (sourceType!=NUMBER_TYPE&&sourceType!=FIELD_TYPE) {
|
||||
return null;
|
||||
}
|
||||
return new IntFieldSource(expressionString, FieldCache.NUMERIC_UTILS_INT_PARSER) {
|
||||
return new IntFieldSource(expressionString) {
|
||||
public String description() {
|
||||
return field;
|
||||
}
|
||||
|
@ -366,7 +365,7 @@ public class StatsCollectorSupplierFactory {
|
|||
if (sourceType!=NUMBER_TYPE&&sourceType!=FIELD_TYPE) {
|
||||
return null;
|
||||
}
|
||||
return new LongFieldSource(expressionString, FieldCache.NUMERIC_UTILS_LONG_PARSER) {
|
||||
return new LongFieldSource(expressionString) {
|
||||
public String description() {
|
||||
return field;
|
||||
}
|
||||
|
@ -375,7 +374,7 @@ public class StatsCollectorSupplierFactory {
|
|||
if (sourceType!=NUMBER_TYPE&&sourceType!=FIELD_TYPE) {
|
||||
return null;
|
||||
}
|
||||
return new FloatFieldSource(expressionString, FieldCache.NUMERIC_UTILS_FLOAT_PARSER) {
|
||||
return new FloatFieldSource(expressionString) {
|
||||
public String description() {
|
||||
return field;
|
||||
}
|
||||
|
@ -384,7 +383,7 @@ public class StatsCollectorSupplierFactory {
|
|||
if (sourceType!=NUMBER_TYPE&&sourceType!=FIELD_TYPE) {
|
||||
return null;
|
||||
}
|
||||
return new DoubleFieldSource(expressionString, FieldCache.NUMERIC_UTILS_DOUBLE_PARSER) {
|
||||
return new DoubleFieldSource(expressionString) {
|
||||
public String description() {
|
||||
return field;
|
||||
}
|
||||
|
|
|
@ -18,13 +18,9 @@
|
|||
package org.apache.solr.analytics.util;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.text.ParseException;
|
||||
import java.util.Arrays;
|
||||
import java.util.Date;
|
||||
|
||||
import org.apache.lucene.index.Terms;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.schema.FieldType;
|
||||
|
|
|
@ -18,18 +18,18 @@
|
|||
package org.apache.solr.analytics.util.valuesource;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.text.ParseException;
|
||||
import java.util.Date;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.queries.function.FunctionValues;
|
||||
import org.apache.lucene.queries.function.docvalues.LongDocValues;
|
||||
import org.apache.lucene.queries.function.valuesource.LongFieldSource;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.Bits;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.lucene.util.mutable.MutableValue;
|
||||
import org.apache.lucene.util.mutable.MutableValueDate;
|
||||
import org.apache.solr.schema.TrieDateField;
|
||||
|
@ -41,15 +41,11 @@ import org.apache.solr.schema.TrieDateField;
|
|||
public class DateFieldSource extends LongFieldSource {
|
||||
|
||||
public DateFieldSource(String field) {
|
||||
super(field, FieldCache.NUMERIC_UTILS_LONG_PARSER);
|
||||
}
|
||||
|
||||
public DateFieldSource(String field, FieldCache.Parser parser) {
|
||||
super(field, parser);
|
||||
super(field);
|
||||
}
|
||||
|
||||
public long externalToLong(String extVal) {
|
||||
return parser.parseValue(new BytesRef(extVal));
|
||||
return NumericUtils.prefixCodedToLong(new BytesRef(extVal));
|
||||
}
|
||||
|
||||
public Object longToObject(long val) {
|
||||
|
@ -63,8 +59,8 @@ public class DateFieldSource extends LongFieldSource {
|
|||
|
||||
@Override
|
||||
public FunctionValues getValues(Map context, AtomicReaderContext readerContext) throws IOException {
|
||||
final NumericDocValues arr = cache.getNumerics(readerContext.reader(), field, parser, true);
|
||||
final Bits valid = cache.getDocsWithField(readerContext.reader(), field);
|
||||
final NumericDocValues arr = DocValues.getNumeric(readerContext.reader(), field);
|
||||
final Bits valid = DocValues.getDocsWithField(readerContext.reader(), field);
|
||||
return new LongDocValues(this) {
|
||||
@Override
|
||||
public long longVal(int doc) {
|
||||
|
@ -111,16 +107,12 @@ public class DateFieldSource extends LongFieldSource {
|
|||
public boolean equals(Object o) {
|
||||
if (o.getClass() != this.getClass()) return false;
|
||||
DateFieldSource other = (DateFieldSource) o;
|
||||
if (parser==null) {
|
||||
return field.equals(other.field);
|
||||
} else {
|
||||
return field.equals(other.field) && parser.equals(other.parser);
|
||||
}
|
||||
return field.equals(other.field);
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
int h = parser == null ? this.getClass().hashCode() : parser.getClass().hashCode();
|
||||
int h = this.getClass().hashCode();
|
||||
h += super.hashCode();
|
||||
return h;
|
||||
}
|
||||
|
|
|
@ -19,6 +19,7 @@ package org.apache.solr.handler.component;
|
|||
|
||||
import org.apache.lucene.index.AtomicReader;
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.search.LeafCollector;
|
||||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
|
@ -27,7 +28,6 @@ import org.apache.lucene.search.Scorer;
|
|||
import org.apache.lucene.search.SimpleCollector;
|
||||
import org.apache.lucene.search.Sort;
|
||||
import org.apache.lucene.search.TopDocs;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.Collector;
|
||||
import org.apache.lucene.search.TopDocsCollector;
|
||||
|
@ -188,7 +188,7 @@ public class ExpandComponent extends SearchComponent implements PluginInfoInitia
|
|||
|
||||
SolrIndexSearcher searcher = req.getSearcher();
|
||||
AtomicReader reader = searcher.getAtomicReader();
|
||||
SortedDocValues values = FieldCache.DEFAULT.getTermsIndex(reader, field);
|
||||
SortedDocValues values = DocValues.getSorted(reader, field);
|
||||
FixedBitSet groupBits = new FixedBitSet(values.getValueCount());
|
||||
DocList docList = rb.getResults().docList;
|
||||
IntOpenHashSet collapsedSet = new IntOpenHashSet(docList.size()*2);
|
||||
|
|
|
@ -25,10 +25,10 @@ import java.util.Map;
|
|||
|
||||
import org.apache.lucene.index.AtomicReader;
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.queries.function.FunctionValues;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.solr.schema.SchemaField;
|
||||
import org.apache.solr.search.SolrIndexSearcher;
|
||||
|
@ -100,7 +100,7 @@ public class FieldFacetStats {
|
|||
// Currently only used by UnInvertedField stats
|
||||
public boolean facetTermNum(int docID, int statsTermNum) throws IOException {
|
||||
if (topLevelSortedValues == null) {
|
||||
topLevelSortedValues = FieldCache.DEFAULT.getTermsIndex(topLevelReader, name);
|
||||
topLevelSortedValues = DocValues.getSorted(topLevelReader, name);
|
||||
}
|
||||
|
||||
int term = topLevelSortedValues.getOrd(docID);
|
||||
|
|
|
@ -30,13 +30,13 @@ import java.util.Set;
|
|||
|
||||
import org.apache.lucene.document.FieldType.NumericType;
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.ReaderUtil;
|
||||
import org.apache.lucene.index.Terms;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.queries.function.FunctionValues;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.Bits;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRef;
|
||||
|
@ -156,13 +156,13 @@ final class NumericFacets {
|
|||
assert doc >= ctx.docBase;
|
||||
switch (numericType) {
|
||||
case LONG:
|
||||
longs = FieldCache.DEFAULT.getNumerics(ctx.reader(), fieldName, FieldCache.NUMERIC_UTILS_LONG_PARSER, true);
|
||||
longs = DocValues.getNumeric(ctx.reader(), fieldName);
|
||||
break;
|
||||
case INT:
|
||||
longs = FieldCache.DEFAULT.getNumerics(ctx.reader(), fieldName, FieldCache.NUMERIC_UTILS_INT_PARSER, true);
|
||||
longs = DocValues.getNumeric(ctx.reader(), fieldName);
|
||||
break;
|
||||
case FLOAT:
|
||||
final NumericDocValues floats = FieldCache.DEFAULT.getNumerics(ctx.reader(), fieldName, FieldCache.NUMERIC_UTILS_FLOAT_PARSER, true);
|
||||
final NumericDocValues floats = DocValues.getNumeric(ctx.reader(), fieldName);
|
||||
// TODO: this bit flipping should probably be moved to tie-break in the PQ comparator
|
||||
longs = new NumericDocValues() {
|
||||
@Override
|
||||
|
@ -174,7 +174,7 @@ final class NumericFacets {
|
|||
};
|
||||
break;
|
||||
case DOUBLE:
|
||||
final NumericDocValues doubles = FieldCache.DEFAULT.getNumerics(ctx.reader(), fieldName, FieldCache.NUMERIC_UTILS_DOUBLE_PARSER, true);
|
||||
final NumericDocValues doubles = DocValues.getNumeric(ctx.reader(), fieldName);
|
||||
// TODO: this bit flipping should probably be moved to tie-break in the PQ comparator
|
||||
longs = new NumericDocValues() {
|
||||
@Override
|
||||
|
@ -188,7 +188,7 @@ final class NumericFacets {
|
|||
default:
|
||||
throw new AssertionError();
|
||||
}
|
||||
docsWithField = FieldCache.DEFAULT.getDocsWithField(ctx.reader(), fieldName);
|
||||
docsWithField = DocValues.getDocsWithField(ctx.reader(), fieldName);
|
||||
}
|
||||
long v = longs.get(doc - ctx.docBase);
|
||||
if (v != 0 || docsWithField.get(doc - ctx.docBase)) {
|
||||
|
|
|
@ -22,11 +22,11 @@ import java.util.*;
|
|||
import java.util.concurrent.*;
|
||||
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.search.DocIdSet;
|
||||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.Filter;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRef;
|
||||
|
@ -236,7 +236,7 @@ class PerSegmentSingleValuedFaceting {
|
|||
BytesRef tempBR = new BytesRef();
|
||||
|
||||
void countTerms() throws IOException {
|
||||
si = FieldCache.DEFAULT.getTermsIndex(context.reader(), fieldName);
|
||||
si = DocValues.getSorted(context.reader(), fieldName);
|
||||
// SolrCore.log.info("reader= " + reader + " FC=" + System.identityHashCode(si));
|
||||
|
||||
if (prefix!=null) {
|
||||
|
|
|
@ -38,6 +38,7 @@ import java.util.concurrent.ThreadPoolExecutor;
|
|||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
import org.apache.lucene.index.AtomicReader;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.DocsEnum;
|
||||
import org.apache.lucene.index.Fields;
|
||||
import org.apache.lucene.index.MultiDocsEnum;
|
||||
|
@ -46,7 +47,6 @@ import org.apache.lucene.index.Term;
|
|||
import org.apache.lucene.index.Terms;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.Filter;
|
||||
import org.apache.lucene.search.MatchAllDocsQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
|
@ -644,7 +644,7 @@ public class SimpleFacets {
|
|||
FieldType ft = searcher.getSchema().getFieldType(fieldName);
|
||||
NamedList<Integer> res = new NamedList<>();
|
||||
|
||||
SortedDocValues si = FieldCache.DEFAULT.getTermsIndex(searcher.getAtomicReader(), fieldName);
|
||||
SortedDocValues si = DocValues.getSorted(searcher.getAtomicReader(), fieldName);
|
||||
|
||||
final BytesRef br = new BytesRef();
|
||||
|
||||
|
|
|
@ -23,12 +23,12 @@ import java.util.Map;
|
|||
import java.util.concurrent.atomic.AtomicLong;
|
||||
|
||||
import org.apache.lucene.index.AtomicReader;
|
||||
import org.apache.lucene.index.DocTermOrds;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
import org.apache.lucene.uninverting.DocTermOrds;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRef;
|
||||
import org.apache.lucene.util.FixedBitSet;
|
||||
|
|
|
@ -25,6 +25,7 @@ import org.apache.lucene.analysis.Analyzer;
|
|||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.GeneralField;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.index.StorableField;
|
||||
|
@ -32,7 +33,6 @@ import org.apache.lucene.queries.function.FunctionValues;
|
|||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.docvalues.BoolDocValues;
|
||||
import org.apache.lucene.queries.function.valuesource.OrdFieldSource;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRef;
|
||||
|
@ -179,7 +179,7 @@ class BoolFieldSource extends ValueSource {
|
|||
|
||||
@Override
|
||||
public FunctionValues getValues(Map context, AtomicReaderContext readerContext) throws IOException {
|
||||
final SortedDocValues sindex = FieldCache.DEFAULT.getTermsIndex(readerContext.reader(), field);
|
||||
final SortedDocValues sindex = DocValues.getSorted(readerContext.reader(), field);
|
||||
|
||||
// figure out what ord maps to true
|
||||
int nord = sindex.getValueCount();
|
||||
|
|
|
@ -178,7 +178,7 @@ public class EnumField extends PrimitiveFieldType {
|
|||
public SortField getSortField(SchemaField field, boolean top) {
|
||||
field.checkSortability();
|
||||
final Object missingValue = Integer.MIN_VALUE;
|
||||
SortField sf = new SortField(field.getName(), SortField.Type.INT, top, FieldCache.NUMERIC_UTILS_INT_PARSER);
|
||||
SortField sf = new SortField(field.getName(), SortField.Type.INT, top);
|
||||
sf.setMissingValue(missingValue);
|
||||
return sf;
|
||||
}
|
||||
|
@ -189,7 +189,7 @@ public class EnumField extends PrimitiveFieldType {
|
|||
@Override
|
||||
public ValueSource getValueSource(SchemaField field, QParser qparser) {
|
||||
field.checkFieldCacheSource(qparser);
|
||||
return new EnumFieldSource(field.getName(), FieldCache.NUMERIC_UTILS_INT_PARSER, enumIntToStringMap, enumStringToIntMap);
|
||||
return new EnumFieldSource(field.getName(), enumIntToStringMap, enumStringToIntMap);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -39,7 +39,6 @@ import org.apache.lucene.queries.function.valuesource.FloatFieldSource;
|
|||
import org.apache.lucene.queries.function.valuesource.IntFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.LongFieldSource;
|
||||
import org.apache.lucene.search.ConstantScoreQuery;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.FieldCacheRangeFilter;
|
||||
import org.apache.lucene.search.NumericRangeQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
|
@ -153,7 +152,7 @@ public class TrieField extends PrimitiveFieldType {
|
|||
else if( sortMissingFirst ) {
|
||||
missingValue = top ? Integer.MAX_VALUE : Integer.MIN_VALUE;
|
||||
}
|
||||
sf = new SortField( field.getName(), SortField.Type.INT, top, FieldCache.NUMERIC_UTILS_INT_PARSER);
|
||||
sf = new SortField( field.getName(), SortField.Type.INT, top);
|
||||
sf.setMissingValue(missingValue);
|
||||
return sf;
|
||||
|
||||
|
@ -164,7 +163,7 @@ public class TrieField extends PrimitiveFieldType {
|
|||
else if( sortMissingFirst ) {
|
||||
missingValue = top ? Float.POSITIVE_INFINITY : Float.NEGATIVE_INFINITY;
|
||||
}
|
||||
sf = new SortField( field.getName(), SortField.Type.FLOAT, top, FieldCache.NUMERIC_UTILS_FLOAT_PARSER);
|
||||
sf = new SortField( field.getName(), SortField.Type.FLOAT, top);
|
||||
sf.setMissingValue(missingValue);
|
||||
return sf;
|
||||
|
||||
|
@ -176,7 +175,7 @@ public class TrieField extends PrimitiveFieldType {
|
|||
else if( sortMissingFirst ) {
|
||||
missingValue = top ? Long.MAX_VALUE : Long.MIN_VALUE;
|
||||
}
|
||||
sf = new SortField( field.getName(), SortField.Type.LONG, top, FieldCache.NUMERIC_UTILS_LONG_PARSER);
|
||||
sf = new SortField( field.getName(), SortField.Type.LONG, top);
|
||||
sf.setMissingValue(missingValue);
|
||||
return sf;
|
||||
|
||||
|
@ -187,7 +186,7 @@ public class TrieField extends PrimitiveFieldType {
|
|||
else if( sortMissingFirst ) {
|
||||
missingValue = top ? Double.POSITIVE_INFINITY : Double.NEGATIVE_INFINITY;
|
||||
}
|
||||
sf = new SortField( field.getName(), SortField.Type.DOUBLE, top, FieldCache.NUMERIC_UTILS_DOUBLE_PARSER);
|
||||
sf = new SortField( field.getName(), SortField.Type.DOUBLE, top);
|
||||
sf.setMissingValue(missingValue);
|
||||
return sf;
|
||||
|
||||
|
@ -201,15 +200,15 @@ public class TrieField extends PrimitiveFieldType {
|
|||
field.checkFieldCacheSource(qparser);
|
||||
switch (type) {
|
||||
case INTEGER:
|
||||
return new IntFieldSource( field.getName(), FieldCache.NUMERIC_UTILS_INT_PARSER );
|
||||
return new IntFieldSource( field.getName());
|
||||
case FLOAT:
|
||||
return new FloatFieldSource( field.getName(), FieldCache.NUMERIC_UTILS_FLOAT_PARSER );
|
||||
return new FloatFieldSource( field.getName());
|
||||
case DATE:
|
||||
return new TrieDateFieldSource( field.getName(), FieldCache.NUMERIC_UTILS_LONG_PARSER );
|
||||
return new TrieDateFieldSource( field.getName());
|
||||
case LONG:
|
||||
return new LongFieldSource( field.getName(), FieldCache.NUMERIC_UTILS_LONG_PARSER );
|
||||
return new LongFieldSource( field.getName());
|
||||
case DOUBLE:
|
||||
return new DoubleFieldSource( field.getName(), FieldCache.NUMERIC_UTILS_DOUBLE_PARSER );
|
||||
return new DoubleFieldSource( field.getName());
|
||||
default:
|
||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Unknown type for trie field: " + field.name);
|
||||
}
|
||||
|
@ -706,8 +705,8 @@ public class TrieField extends PrimitiveFieldType {
|
|||
|
||||
class TrieDateFieldSource extends LongFieldSource {
|
||||
|
||||
public TrieDateFieldSource(String field, FieldCache.Parser parser) {
|
||||
super(field, parser);
|
||||
public TrieDateFieldSource(String field) {
|
||||
super(field);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
|
|
@ -27,6 +27,7 @@ import java.util.Set;
|
|||
|
||||
import org.apache.lucene.index.AtomicReader;
|
||||
import org.apache.lucene.index.AtomicReaderContext;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.DocsEnum;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
|
@ -38,7 +39,6 @@ import org.apache.lucene.queries.function.ValueSource;
|
|||
import org.apache.lucene.search.LeafCollector;
|
||||
import org.apache.lucene.search.Collector;
|
||||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.FilterCollector;
|
||||
import org.apache.lucene.search.IndexSearcher;
|
||||
import org.apache.lucene.search.Query;
|
||||
|
@ -289,11 +289,7 @@ public class CollapsingQParserPlugin extends QParserPlugin {
|
|||
|
||||
SortedDocValues docValues = null;
|
||||
FunctionQuery funcQuery = null;
|
||||
if(schemaField.hasDocValues()) {
|
||||
docValues = searcher.getAtomicReader().getSortedDocValues(this.field);
|
||||
} else {
|
||||
docValues = FieldCache.DEFAULT.getTermsIndex(searcher.getAtomicReader(), this.field);
|
||||
}
|
||||
docValues = DocValues.getSorted(searcher.getAtomicReader(), this.field);
|
||||
|
||||
FieldType fieldType = null;
|
||||
|
||||
|
@ -830,7 +826,7 @@ public class CollapsingQParserPlugin extends QParserPlugin {
|
|||
}
|
||||
|
||||
public void setNextReader(AtomicReaderContext context) throws IOException {
|
||||
this.vals = FieldCache.DEFAULT.getNumerics(context.reader(), this.field, FieldCache.NUMERIC_UTILS_INT_PARSER, false);
|
||||
this.vals = DocValues.getNumeric(context.reader(), this.field);
|
||||
}
|
||||
|
||||
public void collapse(int ord, int contextDoc, int globalDoc) throws IOException {
|
||||
|
@ -898,7 +894,7 @@ public class CollapsingQParserPlugin extends QParserPlugin {
|
|||
}
|
||||
|
||||
public void setNextReader(AtomicReaderContext context) throws IOException {
|
||||
this.vals = FieldCache.DEFAULT.getNumerics(context.reader(), this.field, FieldCache.NUMERIC_UTILS_LONG_PARSER, false);
|
||||
this.vals = DocValues.getNumeric(context.reader(), this.field);
|
||||
}
|
||||
|
||||
public void collapse(int ord, int contextDoc, int globalDoc) throws IOException {
|
||||
|
@ -967,7 +963,7 @@ public class CollapsingQParserPlugin extends QParserPlugin {
|
|||
}
|
||||
|
||||
public void setNextReader(AtomicReaderContext context) throws IOException {
|
||||
this.vals = FieldCache.DEFAULT.getNumerics(context.reader(), this.field, FieldCache.NUMERIC_UTILS_FLOAT_PARSER, false);
|
||||
this.vals = DocValues.getNumeric(context.reader(), this.field);
|
||||
}
|
||||
|
||||
public void collapse(int ord, int contextDoc, int globalDoc) throws IOException {
|
||||
|
|
|
@ -25,18 +25,21 @@ import org.apache.solr.common.util.SimpleOrderedMap;
|
|||
import org.apache.solr.core.SolrCore;
|
||||
import org.apache.solr.core.SolrInfoMBean;
|
||||
|
||||
/*
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.FieldCache.CacheEntry;
|
||||
import org.apache.lucene.util.FieldCacheSanityChecker;
|
||||
import org.apache.lucene.util.FieldCacheSanityChecker.Insanity;
|
||||
*/
|
||||
// nocommit: maybe provide something useful here instead.
|
||||
|
||||
/**
|
||||
* A SolrInfoMBean that provides introspection of the Lucene FieldCache, this is <b>NOT</b> a cache that is managed by Solr.
|
||||
* A SolrInfoMBean that provides introspection of the Solr FieldCache
|
||||
*
|
||||
*/
|
||||
public class SolrFieldCacheMBean implements SolrInfoMBean {
|
||||
|
||||
protected FieldCacheSanityChecker checker = new FieldCacheSanityChecker();
|
||||
//protected FieldCacheSanityChecker checker = new FieldCacheSanityChecker();
|
||||
|
||||
@Override
|
||||
public String getName() { return this.getClass().getName(); }
|
||||
|
@ -44,8 +47,7 @@ public class SolrFieldCacheMBean implements SolrInfoMBean {
|
|||
public String getVersion() { return SolrCore.version; }
|
||||
@Override
|
||||
public String getDescription() {
|
||||
return "Provides introspection of the Lucene FieldCache, "
|
||||
+ "this is **NOT** a cache that is managed by Solr.";
|
||||
return "Provides introspection of the Solr FieldCache ";
|
||||
}
|
||||
@Override
|
||||
public Category getCategory() { return Category.CACHE; }
|
||||
|
@ -60,6 +62,7 @@ public class SolrFieldCacheMBean implements SolrInfoMBean {
|
|||
@Override
|
||||
public NamedList getStatistics() {
|
||||
NamedList stats = new SimpleOrderedMap();
|
||||
/*
|
||||
CacheEntry[] entries = FieldCache.DEFAULT.getCacheEntries();
|
||||
stats.add("entries_count", entries.length);
|
||||
for (int i = 0; i < entries.length; i++) {
|
||||
|
@ -71,17 +74,9 @@ public class SolrFieldCacheMBean implements SolrInfoMBean {
|
|||
|
||||
stats.add("insanity_count", insanity.length);
|
||||
for (int i = 0; i < insanity.length; i++) {
|
||||
|
||||
/** RAM estimation is both CPU and memory intensive... we don't want to do it unless asked.
|
||||
// we only estimate the size of insane entries
|
||||
for (CacheEntry e : insanity[i].getCacheEntries()) {
|
||||
// don't re-estimate if we've already done it.
|
||||
if (null == e.getEstimatedSize()) e.estimateSize();
|
||||
}
|
||||
**/
|
||||
|
||||
stats.add("insanity#" + i, insanity[i].toString());
|
||||
}
|
||||
}*/
|
||||
return stats;
|
||||
}
|
||||
|
||||
|
|
|
@ -23,7 +23,6 @@ import java.util.List;
|
|||
import java.util.Map;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.LuceneTestCase.Slow;
|
||||
import org.apache.solr.client.solrj.SolrServer;
|
||||
import org.apache.solr.client.solrj.SolrServerException;
|
||||
|
@ -422,7 +421,8 @@ public class TestDistributedSearch extends BaseDistributedSearchTestCase {
|
|||
|
||||
// Thread.sleep(10000000000L);
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
|
||||
del("*:*"); // delete all docs and test stats request
|
||||
commit();
|
||||
|
|
|
@ -17,7 +17,6 @@
|
|||
|
||||
package org.apache.solr;
|
||||
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.index.LogDocMergePolicy;
|
||||
import org.noggit.JSONUtil;
|
||||
import org.noggit.ObjectBuilder;
|
||||
|
@ -518,7 +517,8 @@ public class TestGroupingSearch extends SolrTestCaseJ4 {
|
|||
,"/grouped/"+f+"/matches==10"
|
||||
,"/facet_counts/facet_fields/"+f+"==['1',3, '2',3, '3',2, '4',1, '5',1]"
|
||||
);
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
|
||||
// test that grouping works with highlighting
|
||||
assertJQ(req("fq",filt, "q","{!func}"+f2, "group","true", "group.field",f, "fl","id"
|
||||
|
|
|
@ -23,7 +23,6 @@ import java.util.List;
|
|||
import java.util.Map;
|
||||
import java.util.Random;
|
||||
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.LuceneTestCase.Slow;
|
||||
import org.apache.lucene.util.LuceneTestCase.SuppressCodecs;
|
||||
|
@ -134,7 +133,8 @@ public class TestRandomDVFaceting extends SolrTestCaseJ4 {
|
|||
}
|
||||
}
|
||||
} finally {
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -17,7 +17,6 @@
|
|||
|
||||
package org.apache.solr;
|
||||
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.LuceneTestCase.Slow;
|
||||
import org.apache.solr.common.params.ModifiableSolrParams;
|
||||
|
@ -133,7 +132,8 @@ public class TestRandomFaceting extends SolrTestCaseJ4 {
|
|||
}
|
||||
}
|
||||
} finally {
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -22,9 +22,9 @@ import java.util.List;
|
|||
import java.util.Locale;
|
||||
import java.util.Random;
|
||||
|
||||
import org.apache.lucene.index.DocTermOrds;
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.uninverting.DocTermOrds;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.solr.SolrTestCaseJ4;
|
||||
import org.apache.solr.common.params.FacetParams;
|
||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.solr.search.function;
|
|||
|
||||
import org.apache.lucene.codecs.Codec;
|
||||
import org.apache.lucene.index.FieldInvertState;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.search.similarities.DefaultSimilarity;
|
||||
import org.apache.lucene.search.similarities.TFIDFSimilarity;
|
||||
import org.apache.solr.SolrTestCaseJ4;
|
||||
|
@ -203,7 +202,8 @@ public class TestFunctionQuery extends SolrTestCaseJ4 {
|
|||
Arrays.asList("v1","\0:[* TO *]"), 88,12
|
||||
);
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -285,7 +285,8 @@ public class TestFunctionQuery extends SolrTestCaseJ4 {
|
|||
// System.out.println("Done test "+i);
|
||||
}
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -424,7 +425,8 @@ public class TestFunctionQuery extends SolrTestCaseJ4 {
|
|||
);
|
||||
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -642,7 +644,8 @@ public class TestFunctionQuery extends SolrTestCaseJ4 {
|
|||
singleTest(fieldAsFunc, "sqrt(\0)");
|
||||
assertTrue(orig != FileFloatSource.onlyForTesting);
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -669,7 +672,8 @@ public class TestFunctionQuery extends SolrTestCaseJ4 {
|
|||
100,10, 25,5, 0,0, 1,1);
|
||||
singleTest(fieldAsFunc, "log(\0)", 1,0);
|
||||
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
}
|
||||
|
||||
@Test
|
||||
|
|
|
@ -35,7 +35,6 @@ import java.util.concurrent.atomic.AtomicInteger;
|
|||
import junit.framework.Assert;
|
||||
|
||||
import org.apache.commons.io.FileUtils;
|
||||
import org.apache.lucene.search.FieldCache;
|
||||
import org.apache.lucene.util.Constants;
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
|
@ -284,7 +283,8 @@ public abstract class BaseDistributedSearchTestCase extends SolrTestCaseJ4 {
|
|||
@Override
|
||||
public void tearDown() throws Exception {
|
||||
destroyServers();
|
||||
FieldCache.DEFAULT.purgeAllCaches(); // avoid FC insanity
|
||||
// nocommit: split test if needed
|
||||
// FieldCache.DEFAULT.purgeAllCaches(); // hide FC insanity
|
||||
super.tearDown();
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue