Revert "Improve `min_score` implementation."
This reverts commit 56ea65b0f8
.
This commit is contained in:
parent
cb3122baab
commit
f69c350960
|
@ -0,0 +1,68 @@
|
||||||
|
/*
|
||||||
|
* Licensed to Elasticsearch under one or more contributor
|
||||||
|
* license agreements. See the NOTICE file distributed with
|
||||||
|
* this work for additional information regarding copyright
|
||||||
|
* ownership. Elasticsearch licenses this file to you under
|
||||||
|
* the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
* not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing,
|
||||||
|
* software distributed under the License is distributed on an
|
||||||
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
* KIND, either express or implied. See the License for the
|
||||||
|
* specific language governing permissions and limitations
|
||||||
|
* under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.elasticsearch.common.lucene;
|
||||||
|
|
||||||
|
import org.apache.lucene.index.LeafReaderContext;
|
||||||
|
import org.apache.lucene.search.*;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
public class MinimumScoreCollector extends SimpleCollector {
|
||||||
|
|
||||||
|
private final Collector collector;
|
||||||
|
private final float minimumScore;
|
||||||
|
|
||||||
|
private Scorer scorer;
|
||||||
|
private LeafCollector leafCollector;
|
||||||
|
|
||||||
|
public MinimumScoreCollector(Collector collector, float minimumScore) {
|
||||||
|
this.collector = collector;
|
||||||
|
this.minimumScore = minimumScore;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setScorer(Scorer scorer) throws IOException {
|
||||||
|
if (!(scorer instanceof ScoreCachingWrappingScorer)) {
|
||||||
|
scorer = new ScoreCachingWrappingScorer(scorer);
|
||||||
|
}
|
||||||
|
this.scorer = scorer;
|
||||||
|
leafCollector.setScorer(scorer);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void collect(int doc) throws IOException {
|
||||||
|
if (scorer.score() >= minimumScore) {
|
||||||
|
leafCollector.collect(doc);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void doSetNextReader(LeafReaderContext context) throws IOException {
|
||||||
|
leafCollector = collector.getLeafCollector(context);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean needsScores() {
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,285 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elasticsearch under one or more contributor
|
|
||||||
* license agreements. See the NOTICE file distributed with
|
|
||||||
* this work for additional information regarding copyright
|
|
||||||
* ownership. Elasticsearch licenses this file to you under
|
|
||||||
* the Apache License, Version 2.0 (the "License"); you may
|
|
||||||
* not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.lucene.search;
|
|
||||||
|
|
||||||
import org.apache.lucene.index.IndexReader;
|
|
||||||
import org.apache.lucene.index.LeafReaderContext;
|
|
||||||
import org.apache.lucene.index.Term;
|
|
||||||
import org.apache.lucene.search.BulkScorer;
|
|
||||||
import org.apache.lucene.search.DocIdSetIterator;
|
|
||||||
import org.apache.lucene.search.Explanation;
|
|
||||||
import org.apache.lucene.search.IndexSearcher;
|
|
||||||
import org.apache.lucene.search.LeafCollector;
|
|
||||||
import org.apache.lucene.search.Query;
|
|
||||||
import org.apache.lucene.search.ScoreCachingWrappingScorer;
|
|
||||||
import org.apache.lucene.search.Scorer;
|
|
||||||
import org.apache.lucene.search.TwoPhaseIterator;
|
|
||||||
import org.apache.lucene.search.Weight;
|
|
||||||
import org.apache.lucene.util.Bits;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.util.Objects;
|
|
||||||
import java.util.Set;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A {@link Query} wrapper that only emits as hits documents whose score is
|
|
||||||
* above a given threshold. This query only really makes sense for queries
|
|
||||||
* whose score is computed manually, like eg. function score queries.
|
|
||||||
*/
|
|
||||||
public final class MinScoreQuery extends Query {
|
|
||||||
|
|
||||||
private final Query query;
|
|
||||||
private final float minScore;
|
|
||||||
private final IndexSearcher searcher;
|
|
||||||
|
|
||||||
/** Sole constructor. */
|
|
||||||
public MinScoreQuery(Query query, float minScore) {
|
|
||||||
this(query, minScore, null);
|
|
||||||
}
|
|
||||||
|
|
||||||
MinScoreQuery(Query query, float minScore, IndexSearcher searcher) {
|
|
||||||
this.query = query;
|
|
||||||
this.minScore = minScore;
|
|
||||||
this.searcher = searcher;
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Return the wrapped query. */
|
|
||||||
public Query getQuery() {
|
|
||||||
return query;
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Return the minimum score. */
|
|
||||||
public float getMinScore() {
|
|
||||||
return minScore;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public String toString(String field) {
|
|
||||||
return getClass().getSimpleName() + "(" + query.toString(field) + ", minScore=" + minScore + ")";
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public boolean equals(Object obj) {
|
|
||||||
if (super.equals(obj) == false) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
MinScoreQuery that = (MinScoreQuery) obj;
|
|
||||||
return minScore == that.minScore
|
|
||||||
&& searcher == that.searcher
|
|
||||||
&& query.equals(that.query);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int hashCode() {
|
|
||||||
return 31 * super.hashCode() + Objects.hash(query, minScore, searcher);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Query rewrite(IndexReader reader) throws IOException {
|
|
||||||
if (getBoost() != 1f) {
|
|
||||||
return super.rewrite(reader);
|
|
||||||
}
|
|
||||||
Query rewritten = query.rewrite(reader);
|
|
||||||
if (rewritten != query) {
|
|
||||||
return new MinScoreQuery(rewritten, minScore);
|
|
||||||
}
|
|
||||||
return super.rewrite(reader);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Weight createWeight(IndexSearcher searcher, boolean needsScores) throws IOException {
|
|
||||||
final Weight weight = searcher.createWeight(query, true);
|
|
||||||
// We specialize the query for the provided index searcher because it
|
|
||||||
// can't really be cached as the documents that match depend on the
|
|
||||||
// Similarity implementation and the top-level reader.
|
|
||||||
final Query key = new MinScoreQuery(query, minScore, searcher);
|
|
||||||
return new Weight(key) {
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Scorer scorer(LeafReaderContext context) throws IOException {
|
|
||||||
Scorer scorer = weight.scorer(context);
|
|
||||||
if (scorer == null) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
return new MinScoreScorer(this, scorer, minScore);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public BulkScorer bulkScorer(LeafReaderContext context) throws IOException {
|
|
||||||
BulkScorer bulkScorer = weight.bulkScorer(context);
|
|
||||||
if (bulkScorer == null) {
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
return new MinScoreBulkScorer(bulkScorer, minScore);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void normalize(float norm, float boost) {
|
|
||||||
weight.normalize(norm, boost);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public float getValueForNormalization() throws IOException {
|
|
||||||
return weight.getValueForNormalization();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void extractTerms(Set<Term> terms) {
|
|
||||||
weight.extractTerms(terms);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Explanation explain(LeafReaderContext context, int doc) throws IOException {
|
|
||||||
Explanation expl = weight.explain(context, doc);
|
|
||||||
if (expl.isMatch() == false || expl.getValue() >= minScore) {
|
|
||||||
return expl;
|
|
||||||
} else {
|
|
||||||
return Explanation.noMatch("Min score is less than the configured min score=" + minScore, expl);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
private static class MinScoreScorer extends Scorer {
|
|
||||||
|
|
||||||
private final Scorer scorer;
|
|
||||||
private final float minScore;
|
|
||||||
private float score;
|
|
||||||
|
|
||||||
protected MinScoreScorer(Weight weight, Scorer scorer, float minScore) {
|
|
||||||
super(weight);
|
|
||||||
this.scorer = scorer;
|
|
||||||
this.minScore = minScore;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public float score() throws IOException {
|
|
||||||
return score;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int freq() throws IOException {
|
|
||||||
return scorer.freq();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int docID() {
|
|
||||||
return scorer.docID();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int nextDoc() throws IOException {
|
|
||||||
return doNext(scorer.nextDoc());
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int advance(int target) throws IOException {
|
|
||||||
return doNext(scorer.advance(target));
|
|
||||||
}
|
|
||||||
|
|
||||||
private int doNext(int doc) throws IOException {
|
|
||||||
for (; doc != NO_MORE_DOCS; doc = scorer.nextDoc()) {
|
|
||||||
final float score = scorer.score();
|
|
||||||
if (score >= minScore) {
|
|
||||||
this.score = score;
|
|
||||||
return doc;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return NO_MORE_DOCS;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public TwoPhaseIterator asTwoPhaseIterator() {
|
|
||||||
final TwoPhaseIterator twoPhase = scorer.asTwoPhaseIterator();
|
|
||||||
final DocIdSetIterator approximation = twoPhase == null
|
|
||||||
? scorer
|
|
||||||
: twoPhase.approximation();
|
|
||||||
return new TwoPhaseIterator(approximation) {
|
|
||||||
@Override
|
|
||||||
public boolean matches() throws IOException {
|
|
||||||
if (twoPhase != null && twoPhase.matches() == false) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
score = scorer.score();
|
|
||||||
return score >= minScore;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public long cost() {
|
|
||||||
return scorer.cost();
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
private static class MinScoreBulkScorer extends BulkScorer {
|
|
||||||
|
|
||||||
private final BulkScorer bulkScorer;
|
|
||||||
private final float minScore;
|
|
||||||
|
|
||||||
public MinScoreBulkScorer(BulkScorer bulkScorer, float minScore) {
|
|
||||||
this.bulkScorer = bulkScorer;
|
|
||||||
this.minScore = minScore;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public int score(LeafCollector collector, Bits acceptDocs, int min, int max) throws IOException {
|
|
||||||
return bulkScorer.score(new MinScoreLeafCollector(collector, minScore), acceptDocs, min, max);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public long cost() {
|
|
||||||
return bulkScorer.cost();
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
private static class MinScoreLeafCollector implements LeafCollector {
|
|
||||||
|
|
||||||
private final LeafCollector collector;
|
|
||||||
private final float minScore;
|
|
||||||
private Scorer scorer;
|
|
||||||
|
|
||||||
public MinScoreLeafCollector(LeafCollector collector, float minScore) {
|
|
||||||
this.collector = collector;
|
|
||||||
this.minScore = minScore;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void setScorer(Scorer scorer) throws IOException {
|
|
||||||
// we will need scores at least once, maybe more due to the wrapped
|
|
||||||
// collector, so we wrap with a ScoreCachingWrappingScorer
|
|
||||||
if (scorer instanceof ScoreCachingWrappingScorer == false) {
|
|
||||||
scorer = new ScoreCachingWrappingScorer(scorer);
|
|
||||||
}
|
|
||||||
this.scorer = scorer;
|
|
||||||
collector.setScorer(scorer);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void collect(int doc) throws IOException {
|
|
||||||
if (scorer.score() >= minScore) {
|
|
||||||
collector.collect(doc);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -30,12 +30,21 @@ abstract class CustomBoostFactorScorer extends Scorer {
|
||||||
final float maxBoost;
|
final float maxBoost;
|
||||||
final CombineFunction scoreCombiner;
|
final CombineFunction scoreCombiner;
|
||||||
|
|
||||||
CustomBoostFactorScorer(Weight w, Scorer scorer, float maxBoost, CombineFunction scoreCombiner)
|
Float minScore;
|
||||||
|
NextDoc nextDoc;
|
||||||
|
|
||||||
|
CustomBoostFactorScorer(Weight w, Scorer scorer, float maxBoost, CombineFunction scoreCombiner, Float minScore)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
super(w);
|
super(w);
|
||||||
|
if (minScore == null) {
|
||||||
|
nextDoc = new AnyNextDoc();
|
||||||
|
} else {
|
||||||
|
nextDoc = new MinScoreNextDoc();
|
||||||
|
}
|
||||||
this.scorer = scorer;
|
this.scorer = scorer;
|
||||||
this.maxBoost = maxBoost;
|
this.maxBoost = maxBoost;
|
||||||
this.scoreCombiner = scoreCombiner;
|
this.scoreCombiner = scoreCombiner;
|
||||||
|
this.minScore = minScore;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -45,16 +54,20 @@ abstract class CustomBoostFactorScorer extends Scorer {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public int advance(int target) throws IOException {
|
public int advance(int target) throws IOException {
|
||||||
return scorer.advance(target);
|
return nextDoc.advance(target);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public int nextDoc() throws IOException {
|
public int nextDoc() throws IOException {
|
||||||
return scorer.nextDoc();
|
return nextDoc.nextDoc();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public abstract float innerScore() throws IOException;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public abstract float score() throws IOException;
|
public float score() throws IOException {
|
||||||
|
return nextDoc.score();
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public int freq() throws IOException {
|
public int freq() throws IOException {
|
||||||
|
@ -66,4 +79,64 @@ abstract class CustomBoostFactorScorer extends Scorer {
|
||||||
return scorer.cost();
|
return scorer.cost();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public interface NextDoc {
|
||||||
|
public int advance(int target) throws IOException;
|
||||||
|
|
||||||
|
public int nextDoc() throws IOException;
|
||||||
|
|
||||||
|
public float score() throws IOException;
|
||||||
|
}
|
||||||
|
|
||||||
|
public class MinScoreNextDoc implements NextDoc {
|
||||||
|
float currentScore = Float.MAX_VALUE * -1.0f;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int nextDoc() throws IOException {
|
||||||
|
int doc;
|
||||||
|
do {
|
||||||
|
doc = scorer.nextDoc();
|
||||||
|
if (doc == NO_MORE_DOCS) {
|
||||||
|
return doc;
|
||||||
|
}
|
||||||
|
currentScore = innerScore();
|
||||||
|
} while (currentScore < minScore);
|
||||||
|
return doc;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public float score() throws IOException {
|
||||||
|
return currentScore;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int advance(int target) throws IOException {
|
||||||
|
int doc = scorer.advance(target);
|
||||||
|
if (doc == NO_MORE_DOCS) {
|
||||||
|
return doc;
|
||||||
|
}
|
||||||
|
currentScore = innerScore();
|
||||||
|
if (currentScore < minScore) {
|
||||||
|
return scorer.nextDoc();
|
||||||
|
}
|
||||||
|
return doc;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public class AnyNextDoc implements NextDoc {
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int nextDoc() throws IOException {
|
||||||
|
return scorer.nextDoc();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public float score() throws IOException {
|
||||||
|
return innerScore();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int advance(int target) throws IOException {
|
||||||
|
return scorer.advance(target);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -100,15 +100,17 @@ public class FiltersFunctionScoreQuery extends Query {
|
||||||
final FilterFunction[] filterFunctions;
|
final FilterFunction[] filterFunctions;
|
||||||
final ScoreMode scoreMode;
|
final ScoreMode scoreMode;
|
||||||
final float maxBoost;
|
final float maxBoost;
|
||||||
|
private final Float minScore;
|
||||||
|
|
||||||
final protected CombineFunction combineFunction;
|
final protected CombineFunction combineFunction;
|
||||||
|
|
||||||
public FiltersFunctionScoreQuery(Query subQuery, ScoreMode scoreMode, FilterFunction[] filterFunctions, float maxBoost, CombineFunction combineFunction) {
|
public FiltersFunctionScoreQuery(Query subQuery, ScoreMode scoreMode, FilterFunction[] filterFunctions, float maxBoost, Float minScore, CombineFunction combineFunction) {
|
||||||
this.subQuery = subQuery;
|
this.subQuery = subQuery;
|
||||||
this.scoreMode = scoreMode;
|
this.scoreMode = scoreMode;
|
||||||
this.filterFunctions = filterFunctions;
|
this.filterFunctions = filterFunctions;
|
||||||
this.maxBoost = maxBoost;
|
this.maxBoost = maxBoost;
|
||||||
this.combineFunction = combineFunction;
|
this.combineFunction = combineFunction;
|
||||||
|
this.minScore = minScore;
|
||||||
}
|
}
|
||||||
|
|
||||||
public Query getSubQuery() {
|
public Query getSubQuery() {
|
||||||
|
@ -193,7 +195,7 @@ public class FiltersFunctionScoreQuery extends Query {
|
||||||
Scorer filterScorer = filterWeights[i].scorer(context);
|
Scorer filterScorer = filterWeights[i].scorer(context);
|
||||||
docSets[i] = Lucene.asSequentialAccessBits(context.reader().maxDoc(), filterScorer);
|
docSets[i] = Lucene.asSequentialAccessBits(context.reader().maxDoc(), filterScorer);
|
||||||
}
|
}
|
||||||
return new FiltersFunctionFactorScorer(this, subQueryScorer, scoreMode, filterFunctions, maxBoost, functions, docSets, combineFunction, needsScores);
|
return new FiltersFunctionFactorScorer(this, subQueryScorer, scoreMode, filterFunctions, maxBoost, functions, docSets, combineFunction, minScore, needsScores);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -242,8 +244,8 @@ public class FiltersFunctionScoreQuery extends Query {
|
||||||
private final boolean needsScores;
|
private final boolean needsScores;
|
||||||
|
|
||||||
private FiltersFunctionFactorScorer(CustomBoostFactorWeight w, Scorer scorer, ScoreMode scoreMode, FilterFunction[] filterFunctions,
|
private FiltersFunctionFactorScorer(CustomBoostFactorWeight w, Scorer scorer, ScoreMode scoreMode, FilterFunction[] filterFunctions,
|
||||||
float maxBoost, LeafScoreFunction[] functions, Bits[] docSets, CombineFunction scoreCombiner, boolean needsScores) throws IOException {
|
float maxBoost, LeafScoreFunction[] functions, Bits[] docSets, CombineFunction scoreCombiner, Float minScore, boolean needsScores) throws IOException {
|
||||||
super(w, scorer, maxBoost, scoreCombiner);
|
super(w, scorer, maxBoost, scoreCombiner, minScore);
|
||||||
this.scoreMode = scoreMode;
|
this.scoreMode = scoreMode;
|
||||||
this.filterFunctions = filterFunctions;
|
this.filterFunctions = filterFunctions;
|
||||||
this.functions = functions;
|
this.functions = functions;
|
||||||
|
@ -252,7 +254,7 @@ public class FiltersFunctionScoreQuery extends Query {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public float score() throws IOException {
|
public float innerScore() throws IOException {
|
||||||
int docId = scorer.docID();
|
int docId = scorer.docID();
|
||||||
// Even if the weight is created with needsScores=false, it might
|
// Even if the weight is created with needsScores=false, it might
|
||||||
// be costly to call score(), so we explicitly check if scores
|
// be costly to call score(), so we explicitly check if scores
|
||||||
|
@ -349,12 +351,12 @@ public class FiltersFunctionScoreQuery extends Query {
|
||||||
}
|
}
|
||||||
FiltersFunctionScoreQuery other = (FiltersFunctionScoreQuery) o;
|
FiltersFunctionScoreQuery other = (FiltersFunctionScoreQuery) o;
|
||||||
return Objects.equals(this.subQuery, other.subQuery) && this.maxBoost == other.maxBoost &&
|
return Objects.equals(this.subQuery, other.subQuery) && this.maxBoost == other.maxBoost &&
|
||||||
Objects.equals(this.combineFunction, other.combineFunction) &&
|
Objects.equals(this.combineFunction, other.combineFunction) && Objects.equals(this.minScore, other.minScore) &&
|
||||||
Arrays.equals(this.filterFunctions, other.filterFunctions);
|
Arrays.equals(this.filterFunctions, other.filterFunctions);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public int hashCode() {
|
public int hashCode() {
|
||||||
return Objects.hash(super.hashCode(), subQuery, maxBoost, combineFunction, filterFunctions);
|
return Objects.hash(super.hashCode(), subQuery, maxBoost, combineFunction, minScore, filterFunctions);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -40,11 +40,13 @@ public class FunctionScoreQuery extends Query {
|
||||||
final ScoreFunction function;
|
final ScoreFunction function;
|
||||||
final float maxBoost;
|
final float maxBoost;
|
||||||
final CombineFunction combineFunction;
|
final CombineFunction combineFunction;
|
||||||
|
private Float minScore;
|
||||||
|
|
||||||
public FunctionScoreQuery(Query subQuery, ScoreFunction function, CombineFunction combineFunction, float maxBoost) {
|
public FunctionScoreQuery(Query subQuery, ScoreFunction function, Float minScore, CombineFunction combineFunction, float maxBoost) {
|
||||||
this.subQuery = subQuery;
|
this.subQuery = subQuery;
|
||||||
this.function = function;
|
this.function = function;
|
||||||
this.combineFunction = combineFunction;
|
this.combineFunction = combineFunction;
|
||||||
|
this.minScore = minScore;
|
||||||
this.maxBoost = maxBoost;
|
this.maxBoost = maxBoost;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -131,7 +133,7 @@ public class FunctionScoreQuery extends Query {
|
||||||
if (function != null) {
|
if (function != null) {
|
||||||
leafFunction = function.getLeafScoreFunction(context);
|
leafFunction = function.getLeafScoreFunction(context);
|
||||||
}
|
}
|
||||||
return new FunctionFactorScorer(this, subQueryScorer, leafFunction, maxBoost, combineFunction, needsScores);
|
return new FunctionFactorScorer(this, subQueryScorer, leafFunction, maxBoost, combineFunction, minScore, needsScores);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -154,15 +156,15 @@ public class FunctionScoreQuery extends Query {
|
||||||
private final LeafScoreFunction function;
|
private final LeafScoreFunction function;
|
||||||
private final boolean needsScores;
|
private final boolean needsScores;
|
||||||
|
|
||||||
private FunctionFactorScorer(CustomBoostFactorWeight w, Scorer scorer, LeafScoreFunction function, float maxBoost, CombineFunction scoreCombiner, boolean needsScores)
|
private FunctionFactorScorer(CustomBoostFactorWeight w, Scorer scorer, LeafScoreFunction function, float maxBoost, CombineFunction scoreCombiner, Float minScore, boolean needsScores)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
super(w, scorer, maxBoost, scoreCombiner);
|
super(w, scorer, maxBoost, scoreCombiner, minScore);
|
||||||
this.function = function;
|
this.function = function;
|
||||||
this.needsScores = needsScores;
|
this.needsScores = needsScores;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public float score() throws IOException {
|
public float innerScore() throws IOException {
|
||||||
// Even if the weight is created with needsScores=false, it might
|
// Even if the weight is created with needsScores=false, it might
|
||||||
// be costly to call score(), so we explicitly check if scores
|
// be costly to call score(), so we explicitly check if scores
|
||||||
// are needed
|
// are needed
|
||||||
|
@ -195,11 +197,11 @@ public class FunctionScoreQuery extends Query {
|
||||||
FunctionScoreQuery other = (FunctionScoreQuery) o;
|
FunctionScoreQuery other = (FunctionScoreQuery) o;
|
||||||
return Objects.equals(this.subQuery, other.subQuery) && Objects.equals(this.function, other.function)
|
return Objects.equals(this.subQuery, other.subQuery) && Objects.equals(this.function, other.function)
|
||||||
&& Objects.equals(this.combineFunction, other.combineFunction)
|
&& Objects.equals(this.combineFunction, other.combineFunction)
|
||||||
&& this.maxBoost == other.maxBoost;
|
&& Objects.equals(this.minScore, other.minScore) && this.maxBoost == other.maxBoost;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public int hashCode() {
|
public int hashCode() {
|
||||||
return Objects.hash(super.hashCode(), subQuery.hashCode(), function, combineFunction, maxBoost);
|
return Objects.hash(super.hashCode(), subQuery.hashCode(), function, combineFunction, minScore, maxBoost);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.search.Query;
|
||||||
import org.elasticsearch.common.io.stream.StreamInput;
|
import org.elasticsearch.common.io.stream.StreamInput;
|
||||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||||
import org.elasticsearch.common.io.stream.Writeable;
|
import org.elasticsearch.common.io.stream.Writeable;
|
||||||
import org.elasticsearch.common.lucene.search.MinScoreQuery;
|
|
||||||
import org.elasticsearch.common.lucene.search.function.CombineFunction;
|
import org.elasticsearch.common.lucene.search.function.CombineFunction;
|
||||||
import org.elasticsearch.common.lucene.search.function.FiltersFunctionScoreQuery;
|
import org.elasticsearch.common.lucene.search.function.FiltersFunctionScoreQuery;
|
||||||
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
||||||
|
@ -313,17 +312,10 @@ public class FunctionScoreQueryBuilder extends AbstractQueryBuilder<FunctionScor
|
||||||
combineFunction = DEFAULT_BOOST_MODE;
|
combineFunction = DEFAULT_BOOST_MODE;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
query = new FunctionScoreQuery(query, function, combineFunction, maxBoost);
|
return new FunctionScoreQuery(query, function, minScore, combineFunction, maxBoost);
|
||||||
} else {
|
|
||||||
// in all other cases we create a FiltersFunctionScoreQuery
|
|
||||||
query = new FiltersFunctionScoreQuery(query, scoreMode, filterFunctions, maxBoost, boostMode == null ? DEFAULT_BOOST_MODE : boostMode);
|
|
||||||
}
|
}
|
||||||
|
// in all other cases we create a FiltersFunctionScoreQuery
|
||||||
if (minScore != null) {
|
return new FiltersFunctionScoreQuery(query, scoreMode, filterFunctions, maxBoost, minScore, boostMode == null ? DEFAULT_BOOST_MODE : boostMode);
|
||||||
query = new MinScoreQuery(query, minScore);
|
|
||||||
}
|
|
||||||
|
|
||||||
return query;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -478,6 +478,11 @@ public class PercolateContext extends SearchContext {
|
||||||
throw new UnsupportedOperationException();
|
throw new UnsupportedOperationException();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Float minimumScore() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public SearchContext sort(Sort sort) {
|
public SearchContext sort(Sort sort) {
|
||||||
this.sort = sort;
|
this.sort = sort;
|
||||||
|
|
|
@ -27,7 +27,6 @@ import org.elasticsearch.cache.recycler.PageCacheRecycler;
|
||||||
import org.elasticsearch.common.Nullable;
|
import org.elasticsearch.common.Nullable;
|
||||||
import org.elasticsearch.common.ParseFieldMatcher;
|
import org.elasticsearch.common.ParseFieldMatcher;
|
||||||
import org.elasticsearch.common.lease.Releasables;
|
import org.elasticsearch.common.lease.Releasables;
|
||||||
import org.elasticsearch.common.lucene.search.MinScoreQuery;
|
|
||||||
import org.elasticsearch.common.lucene.search.Queries;
|
import org.elasticsearch.common.lucene.search.Queries;
|
||||||
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
||||||
import org.elasticsearch.common.lucene.search.function.WeightFactorFunction;
|
import org.elasticsearch.common.lucene.search.function.WeightFactorFunction;
|
||||||
|
@ -216,9 +215,6 @@ public class DefaultSearchContext extends SearchContext {
|
||||||
parsedQuery(new ParsedQuery(filtered, parsedQuery()));
|
parsedQuery(new ParsedQuery(filtered, parsedQuery()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (minimumScore != null) {
|
|
||||||
this.query = new MinScoreQuery(query, minimumScore);
|
|
||||||
}
|
|
||||||
try {
|
try {
|
||||||
this.query = searcher().rewrite(this.query);
|
this.query = searcher().rewrite(this.query);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
|
@ -495,6 +491,11 @@ public class DefaultSearchContext extends SearchContext {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Float minimumScore() {
|
||||||
|
return this.minimumScore;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public SearchContext sort(Sort sort) {
|
public SearchContext sort(Sort sort) {
|
||||||
this.sort = sort;
|
this.sort = sort;
|
||||||
|
|
|
@ -317,6 +317,11 @@ public abstract class FilteredSearchContext extends SearchContext {
|
||||||
return in.minimumScore(minimumScore);
|
return in.minimumScore(minimumScore);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Float minimumScore() {
|
||||||
|
return in.minimumScore();
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public SearchContext sort(Sort sort) {
|
public SearchContext sort(Sort sort) {
|
||||||
return in.sort(sort);
|
return in.sort(sort);
|
||||||
|
|
|
@ -230,6 +230,8 @@ public abstract class SearchContext extends DelegatingHasContextAndHeaders imple
|
||||||
|
|
||||||
public abstract SearchContext minimumScore(float minimumScore);
|
public abstract SearchContext minimumScore(float minimumScore);
|
||||||
|
|
||||||
|
public abstract Float minimumScore();
|
||||||
|
|
||||||
public abstract SearchContext sort(Sort sort);
|
public abstract SearchContext sort(Sort sort);
|
||||||
|
|
||||||
public abstract Sort sort();
|
public abstract Sort sort();
|
||||||
|
|
|
@ -44,6 +44,7 @@ import org.apache.lucene.search.Weight;
|
||||||
import org.elasticsearch.action.search.SearchType;
|
import org.elasticsearch.action.search.SearchType;
|
||||||
import org.elasticsearch.common.inject.Inject;
|
import org.elasticsearch.common.inject.Inject;
|
||||||
import org.elasticsearch.common.lucene.Lucene;
|
import org.elasticsearch.common.lucene.Lucene;
|
||||||
|
import org.elasticsearch.common.lucene.MinimumScoreCollector;
|
||||||
import org.elasticsearch.common.lucene.search.FilteredCollector;
|
import org.elasticsearch.common.lucene.search.FilteredCollector;
|
||||||
import org.elasticsearch.search.SearchParseElement;
|
import org.elasticsearch.search.SearchParseElement;
|
||||||
import org.elasticsearch.search.SearchPhase;
|
import org.elasticsearch.search.SearchPhase;
|
||||||
|
@ -276,6 +277,11 @@ public class QueryPhase implements SearchPhase {
|
||||||
allCollectors.addAll(searchContext.queryCollectors().values());
|
allCollectors.addAll(searchContext.queryCollectors().values());
|
||||||
collector = MultiCollector.wrap(allCollectors);
|
collector = MultiCollector.wrap(allCollectors);
|
||||||
|
|
||||||
|
// apply the minimum score after multi collector so we filter aggs as well
|
||||||
|
if (searchContext.minimumScore() != null) {
|
||||||
|
collector = new MinimumScoreCollector(collector, searchContext.minimumScore());
|
||||||
|
}
|
||||||
|
|
||||||
if (collector.getClass() == TotalHitCountCollector.class) {
|
if (collector.getClass() == TotalHitCountCollector.class) {
|
||||||
// Optimize counts in simple cases to return in constant time
|
// Optimize counts in simple cases to return in constant time
|
||||||
// instead of using a collector
|
// instead of using a collector
|
||||||
|
|
|
@ -1,195 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elasticsearch under one or more contributor
|
|
||||||
* license agreements. See the NOTICE file distributed with
|
|
||||||
* this work for additional information regarding copyright
|
|
||||||
* ownership. Elasticsearch licenses this file to you under
|
|
||||||
* the Apache License, Version 2.0 (the "License"); you may
|
|
||||||
* not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.lucene.search;
|
|
||||||
|
|
||||||
import org.apache.lucene.analysis.Analyzer;
|
|
||||||
import org.apache.lucene.analysis.MockAnalyzer;
|
|
||||||
import org.apache.lucene.analysis.MockTokenizer;
|
|
||||||
import org.apache.lucene.document.Document;
|
|
||||||
import org.apache.lucene.document.Field.Store;
|
|
||||||
import org.apache.lucene.document.TextField;
|
|
||||||
import org.apache.lucene.index.IndexReader;
|
|
||||||
import org.apache.lucene.index.MultiReader;
|
|
||||||
import org.apache.lucene.index.RandomIndexWriter;
|
|
||||||
import org.apache.lucene.index.Term;
|
|
||||||
import org.apache.lucene.search.BooleanClause.Occur;
|
|
||||||
import org.apache.lucene.search.*;
|
|
||||||
import org.apache.lucene.store.Directory;
|
|
||||||
import org.apache.lucene.util.IOUtils;
|
|
||||||
import org.apache.lucene.util.TestUtil;
|
|
||||||
import org.elasticsearch.test.ESTestCase;
|
|
||||||
import org.junit.AfterClass;
|
|
||||||
import org.junit.BeforeClass;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
|
|
||||||
public class MinScoreQueryTests extends ESTestCase {
|
|
||||||
|
|
||||||
private static String[] terms;
|
|
||||||
private static Directory dir;
|
|
||||||
private static IndexReader r;
|
|
||||||
private static IndexSearcher s;
|
|
||||||
|
|
||||||
@BeforeClass
|
|
||||||
public static void before() throws IOException {
|
|
||||||
dir = newDirectory();
|
|
||||||
terms = new String[TestUtil.nextInt(random(), 2, 15)];
|
|
||||||
for (int i = 0; i < terms.length; ++i) {
|
|
||||||
terms[i] = TestUtil.randomSimpleString(random());
|
|
||||||
}
|
|
||||||
final int numDocs = TestUtil.nextInt(random(), 1, 200);
|
|
||||||
Analyzer analyzer = new MockAnalyzer(random(), MockTokenizer.WHITESPACE, false);
|
|
||||||
RandomIndexWriter w = new RandomIndexWriter(random(), dir, newIndexWriterConfig(analyzer));
|
|
||||||
for (int i = 0; i < numDocs; ++i) {
|
|
||||||
StringBuilder value = new StringBuilder();
|
|
||||||
final int numTerms = random().nextInt(10);
|
|
||||||
for (int j = 0; j < numTerms; ++j) {
|
|
||||||
if (j > 0) {
|
|
||||||
value.append(' ');
|
|
||||||
}
|
|
||||||
// simulate zipf distribution
|
|
||||||
String term = terms[TestUtil.nextInt(random(), 0, TestUtil.nextInt(random(), 0, terms.length - 1))];
|
|
||||||
value.append(term);
|
|
||||||
}
|
|
||||||
Document doc = new Document();
|
|
||||||
doc.add(new TextField("field", value.toString(), Store.NO));
|
|
||||||
w.addDocument(doc);
|
|
||||||
}
|
|
||||||
r = w.getReader();
|
|
||||||
s = newSearcher(r);
|
|
||||||
w.close();
|
|
||||||
}
|
|
||||||
|
|
||||||
@AfterClass
|
|
||||||
public static void after() throws IOException {
|
|
||||||
IOUtils.close(r, dir);
|
|
||||||
terms = null;
|
|
||||||
r = null;
|
|
||||||
s = null;
|
|
||||||
dir = null;
|
|
||||||
}
|
|
||||||
|
|
||||||
private static Term randomTerm() {
|
|
||||||
return new Term("field", terms[random().nextInt(terms.length)]);
|
|
||||||
}
|
|
||||||
|
|
||||||
public void testEquals() throws IOException {
|
|
||||||
QueryUtils.checkEqual(new MinScoreQuery(new MatchAllDocsQuery(), 0.5f), new MinScoreQuery(new MatchAllDocsQuery(), 0.5f));
|
|
||||||
QueryUtils.checkUnequal(new MinScoreQuery(new MatchAllDocsQuery(), 0.5f), new MinScoreQuery(new MatchAllDocsQuery(), 0.3f));
|
|
||||||
QueryUtils.checkUnequal(new MinScoreQuery(new MatchAllDocsQuery(), 0.5f), new MinScoreQuery(new MatchNoDocsQuery(), 0.5f));
|
|
||||||
|
|
||||||
IndexSearcher s1 = new IndexSearcher(new MultiReader());
|
|
||||||
IndexSearcher s2 = new IndexSearcher(new MultiReader());
|
|
||||||
QueryUtils.checkEqual(new MinScoreQuery(new MatchAllDocsQuery(), 0.5f, s1), new MinScoreQuery(new MatchAllDocsQuery(), 0.5f, s1));
|
|
||||||
QueryUtils.checkUnequal(new MinScoreQuery(new MatchAllDocsQuery(), 0.5f, s1), new MinScoreQuery(new MatchAllDocsQuery(), 0.5f, s2));
|
|
||||||
}
|
|
||||||
|
|
||||||
/** pick a min score which is in the range of scores produced by the query */
|
|
||||||
private float randomMinScore(Query query) throws IOException {
|
|
||||||
TopDocs topDocs = s.search(query, 2);
|
|
||||||
float base = 0;
|
|
||||||
switch (topDocs.totalHits) {
|
|
||||||
case 0:
|
|
||||||
break;
|
|
||||||
case 1:
|
|
||||||
base = topDocs.scoreDocs[0].score;
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
base = (topDocs.scoreDocs[0].score + topDocs.scoreDocs[1].score) / 2;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
float delta = random().nextFloat() - 0.5f;
|
|
||||||
return base + delta;
|
|
||||||
}
|
|
||||||
|
|
||||||
private void assertMinScoreEquivalence(Query query, Query minScoreQuery, float minScore) throws IOException {
|
|
||||||
final TopDocs topDocs = s.search(query, s.getIndexReader().maxDoc());
|
|
||||||
final TopDocs minScoreTopDocs = s.search(minScoreQuery, s.getIndexReader().maxDoc());
|
|
||||||
|
|
||||||
int j = 0;
|
|
||||||
for (int i = 0; i < topDocs.totalHits; ++i) {
|
|
||||||
if (topDocs.scoreDocs[i].score >= minScore) {
|
|
||||||
assertEquals(topDocs.scoreDocs[i].doc, minScoreTopDocs.scoreDocs[j].doc);
|
|
||||||
assertEquals(topDocs.scoreDocs[i].score, minScoreTopDocs.scoreDocs[j].score, 1e-5f);
|
|
||||||
j++;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
assertEquals(minScoreTopDocs.totalHits, j);
|
|
||||||
}
|
|
||||||
|
|
||||||
public void testBasics() throws Exception {
|
|
||||||
final int iters = 5;
|
|
||||||
for (int iter = 0; iter < iters; ++iter) {
|
|
||||||
Term term = randomTerm();
|
|
||||||
Query query = new TermQuery(term);
|
|
||||||
float minScore = randomMinScore(query);
|
|
||||||
Query minScoreQuery = new MinScoreQuery(query, minScore);
|
|
||||||
assertMinScoreEquivalence(query, minScoreQuery, minScore);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public void testFilteredApproxQuery() throws Exception {
|
|
||||||
// same as testBasics but with a query that exposes approximations
|
|
||||||
final int iters = 5;
|
|
||||||
for (int iter = 0; iter < iters; ++iter) {
|
|
||||||
Term term = randomTerm();
|
|
||||||
Query query = new TermQuery(term);
|
|
||||||
float minScore = randomMinScore(query);
|
|
||||||
Query minScoreQuery = new MinScoreQuery(new RandomApproximationQuery(query, random()), minScore);
|
|
||||||
assertMinScoreEquivalence(query, minScoreQuery, minScore);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public void testNestedInConjunction() throws Exception {
|
|
||||||
// To test scorers as well, not only bulk scorers
|
|
||||||
Term t1 = randomTerm();
|
|
||||||
Term t2 = randomTerm();
|
|
||||||
Query tq1 = new TermQuery(t1);
|
|
||||||
Query tq2 = new TermQuery(t2);
|
|
||||||
float minScore = randomMinScore(tq1);
|
|
||||||
BooleanQuery bq1 = new BooleanQuery.Builder()
|
|
||||||
.add(new MinScoreQuery(tq1, minScore), Occur.MUST)
|
|
||||||
.add(tq2, Occur.FILTER)
|
|
||||||
.build();
|
|
||||||
BooleanQuery bq2 = new BooleanQuery.Builder()
|
|
||||||
.add(tq1, Occur.MUST)
|
|
||||||
.add(tq2, Occur.FILTER)
|
|
||||||
.build();
|
|
||||||
assertMinScoreEquivalence(bq2, bq1, minScore);
|
|
||||||
}
|
|
||||||
|
|
||||||
public void testNestedInConjunctionWithApprox() throws Exception {
|
|
||||||
// same, but with approximations
|
|
||||||
Term t1 = randomTerm();
|
|
||||||
Term t2 = randomTerm();
|
|
||||||
Query tq1 = new TermQuery(t1);
|
|
||||||
Query tq2 = new TermQuery(t2);
|
|
||||||
float minScore = randomMinScore(tq1);
|
|
||||||
BooleanQuery bq1 = new BooleanQuery.Builder()
|
|
||||||
.add(new MinScoreQuery(new RandomApproximationQuery(tq1, random()), minScore), Occur.MUST)
|
|
||||||
.add(tq2, Occur.FILTER)
|
|
||||||
.build();
|
|
||||||
BooleanQuery bq2 = new BooleanQuery.Builder()
|
|
||||||
.add(tq1, Occur.MUST)
|
|
||||||
.add(tq2, Occur.FILTER)
|
|
||||||
.build();
|
|
||||||
assertMinScoreEquivalence(bq2, bq1, minScore);
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -26,8 +26,11 @@ import org.apache.lucene.search.MatchAllDocsQuery;
|
||||||
import org.apache.lucene.search.Query;
|
import org.apache.lucene.search.Query;
|
||||||
import org.apache.lucene.search.TermQuery;
|
import org.apache.lucene.search.TermQuery;
|
||||||
import org.elasticsearch.common.ParsingException;
|
import org.elasticsearch.common.ParsingException;
|
||||||
import org.elasticsearch.common.lucene.search.MinScoreQuery;
|
import org.elasticsearch.common.lucene.search.function.CombineFunction;
|
||||||
import org.elasticsearch.common.lucene.search.function.*;
|
import org.elasticsearch.common.lucene.search.function.FieldValueFactorFunction;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.FiltersFunctionScoreQuery;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.WeightFactorFunction;
|
||||||
import org.elasticsearch.common.xcontent.XContentType;
|
import org.elasticsearch.common.xcontent.XContentType;
|
||||||
import org.elasticsearch.index.query.AbstractQueryBuilder;
|
import org.elasticsearch.index.query.AbstractQueryBuilder;
|
||||||
import org.elasticsearch.index.query.AbstractQueryTestCase;
|
import org.elasticsearch.index.query.AbstractQueryTestCase;
|
||||||
|
@ -179,12 +182,6 @@ public class FunctionScoreQueryBuilderTests extends AbstractQueryTestCase<Functi
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void doAssertLuceneQuery(FunctionScoreQueryBuilder queryBuilder, Query query, QueryShardContext context) throws IOException {
|
protected void doAssertLuceneQuery(FunctionScoreQueryBuilder queryBuilder, Query query, QueryShardContext context) throws IOException {
|
||||||
if (queryBuilder.getMinScore() != null) {
|
|
||||||
assertThat(query, instanceOf(MinScoreQuery.class));
|
|
||||||
MinScoreQuery msq = (MinScoreQuery) query;
|
|
||||||
assertEquals(queryBuilder.getMinScore(), msq.getMinScore(), 0f);
|
|
||||||
query = msq.getQuery();
|
|
||||||
}
|
|
||||||
assertThat(query, either(instanceOf(FunctionScoreQuery.class)).or(instanceOf(FiltersFunctionScoreQuery.class)));
|
assertThat(query, either(instanceOf(FunctionScoreQuery.class)).or(instanceOf(FiltersFunctionScoreQuery.class)));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -24,6 +24,7 @@ import org.apache.lucene.document.Document;
|
||||||
import org.apache.lucene.document.Field;
|
import org.apache.lucene.document.Field;
|
||||||
import org.apache.lucene.document.TextField;
|
import org.apache.lucene.document.TextField;
|
||||||
import org.apache.lucene.index.DirectoryReader;
|
import org.apache.lucene.index.DirectoryReader;
|
||||||
|
import org.apache.lucene.index.IndexReader;
|
||||||
import org.apache.lucene.index.IndexWriter;
|
import org.apache.lucene.index.IndexWriter;
|
||||||
import org.apache.lucene.index.LeafReaderContext;
|
import org.apache.lucene.index.LeafReaderContext;
|
||||||
import org.apache.lucene.index.SortedNumericDocValues;
|
import org.apache.lucene.index.SortedNumericDocValues;
|
||||||
|
@ -31,7 +32,6 @@ import org.apache.lucene.index.Term;
|
||||||
import org.apache.lucene.search.Explanation;
|
import org.apache.lucene.search.Explanation;
|
||||||
import org.apache.lucene.search.IndexSearcher;
|
import org.apache.lucene.search.IndexSearcher;
|
||||||
import org.apache.lucene.search.MatchAllDocsQuery;
|
import org.apache.lucene.search.MatchAllDocsQuery;
|
||||||
import org.apache.lucene.search.Query;
|
|
||||||
import org.apache.lucene.search.TermQuery;
|
import org.apache.lucene.search.TermQuery;
|
||||||
import org.apache.lucene.search.TopDocs;
|
import org.apache.lucene.search.TopDocs;
|
||||||
import org.apache.lucene.search.Weight;
|
import org.apache.lucene.search.Weight;
|
||||||
|
@ -39,8 +39,14 @@ import org.apache.lucene.store.Directory;
|
||||||
import org.apache.lucene.util.Accountable;
|
import org.apache.lucene.util.Accountable;
|
||||||
import org.apache.lucene.util.BytesRef;
|
import org.apache.lucene.util.BytesRef;
|
||||||
import org.elasticsearch.common.Nullable;
|
import org.elasticsearch.common.Nullable;
|
||||||
import org.elasticsearch.common.lucene.search.MinScoreQuery;
|
import org.elasticsearch.common.lucene.search.function.CombineFunction;
|
||||||
import org.elasticsearch.common.lucene.search.function.*;
|
import org.elasticsearch.common.lucene.search.function.FieldValueFactorFunction;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.FiltersFunctionScoreQuery;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.LeafScoreFunction;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.RandomScoreFunction;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.ScoreFunction;
|
||||||
|
import org.elasticsearch.common.lucene.search.function.WeightFactorFunction;
|
||||||
import org.elasticsearch.index.Index;
|
import org.elasticsearch.index.Index;
|
||||||
import org.elasticsearch.index.fielddata.AtomicFieldData;
|
import org.elasticsearch.index.fielddata.AtomicFieldData;
|
||||||
import org.elasticsearch.index.fielddata.AtomicNumericFieldData;
|
import org.elasticsearch.index.fielddata.AtomicNumericFieldData;
|
||||||
|
@ -312,7 +318,7 @@ public class FunctionScoreTests extends ESTestCase {
|
||||||
}
|
}
|
||||||
|
|
||||||
public Explanation getFunctionScoreExplanation(IndexSearcher searcher, ScoreFunction scoreFunction) throws IOException {
|
public Explanation getFunctionScoreExplanation(IndexSearcher searcher, ScoreFunction scoreFunction) throws IOException {
|
||||||
Query functionScoreQuery = new MinScoreQuery(new FunctionScoreQuery(new TermQuery(TERM), scoreFunction, CombineFunction.AVG, 100), 0f);
|
FunctionScoreQuery functionScoreQuery = new FunctionScoreQuery(new TermQuery(TERM), scoreFunction, 0.0f, CombineFunction.AVG, 100);
|
||||||
Weight weight = searcher.createNormalizedWeight(functionScoreQuery, true);
|
Weight weight = searcher.createNormalizedWeight(functionScoreQuery, true);
|
||||||
Explanation explanation = weight.explain(searcher.getIndexReader().leaves().get(0), 0);
|
Explanation explanation = weight.explain(searcher.getIndexReader().leaves().get(0), 0);
|
||||||
return explanation.getDetails()[1];
|
return explanation.getDetails()[1];
|
||||||
|
@ -376,22 +382,22 @@ public class FunctionScoreTests extends ESTestCase {
|
||||||
}
|
}
|
||||||
|
|
||||||
public Explanation getFiltersFunctionScoreExplanation(IndexSearcher searcher, ScoreFunction... scoreFunctions) throws IOException {
|
public Explanation getFiltersFunctionScoreExplanation(IndexSearcher searcher, ScoreFunction... scoreFunctions) throws IOException {
|
||||||
Query filtersFunctionScoreQuery = getFiltersFunctionScoreQuery(FiltersFunctionScoreQuery.ScoreMode.AVG, CombineFunction.AVG, scoreFunctions);
|
FiltersFunctionScoreQuery filtersFunctionScoreQuery = getFiltersFunctionScoreQuery(FiltersFunctionScoreQuery.ScoreMode.AVG, CombineFunction.AVG, scoreFunctions);
|
||||||
return getExplanation(searcher, filtersFunctionScoreQuery).getDetails()[1];
|
return getExplanation(searcher, filtersFunctionScoreQuery).getDetails()[1];
|
||||||
}
|
}
|
||||||
|
|
||||||
protected Explanation getExplanation(IndexSearcher searcher, Query filtersFunctionScoreQuery) throws IOException {
|
protected Explanation getExplanation(IndexSearcher searcher, FiltersFunctionScoreQuery filtersFunctionScoreQuery) throws IOException {
|
||||||
Weight weight = searcher.createNormalizedWeight(filtersFunctionScoreQuery, true);
|
Weight weight = searcher.createNormalizedWeight(filtersFunctionScoreQuery, true);
|
||||||
return weight.explain(searcher.getIndexReader().leaves().get(0), 0);
|
return weight.explain(searcher.getIndexReader().leaves().get(0), 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
public Query getFiltersFunctionScoreQuery(FiltersFunctionScoreQuery.ScoreMode scoreMode, CombineFunction combineFunction, ScoreFunction... scoreFunctions) {
|
public FiltersFunctionScoreQuery getFiltersFunctionScoreQuery(FiltersFunctionScoreQuery.ScoreMode scoreMode, CombineFunction combineFunction, ScoreFunction... scoreFunctions) {
|
||||||
FiltersFunctionScoreQuery.FilterFunction[] filterFunctions = new FiltersFunctionScoreQuery.FilterFunction[scoreFunctions.length];
|
FiltersFunctionScoreQuery.FilterFunction[] filterFunctions = new FiltersFunctionScoreQuery.FilterFunction[scoreFunctions.length];
|
||||||
for (int i = 0; i < scoreFunctions.length; i++) {
|
for (int i = 0; i < scoreFunctions.length; i++) {
|
||||||
filterFunctions[i] = new FiltersFunctionScoreQuery.FilterFunction(
|
filterFunctions[i] = new FiltersFunctionScoreQuery.FilterFunction(
|
||||||
new TermQuery(TERM), scoreFunctions[i]);
|
new TermQuery(TERM), scoreFunctions[i]);
|
||||||
}
|
}
|
||||||
return new MinScoreQuery(new FiltersFunctionScoreQuery(new TermQuery(TERM), scoreMode, filterFunctions, Float.MAX_VALUE, combineFunction), Float.MAX_VALUE * -1);
|
return new FiltersFunctionScoreQuery(new TermQuery(TERM), scoreMode, filterFunctions, Float.MAX_VALUE, Float.MAX_VALUE * -1, combineFunction);
|
||||||
}
|
}
|
||||||
|
|
||||||
public void checkFiltersFunctionScoreExplanation(Explanation randomExplanation, String functionExpl, int whichFunction) {
|
public void checkFiltersFunctionScoreExplanation(Explanation randomExplanation, String functionExpl, int whichFunction) {
|
||||||
|
@ -465,7 +471,7 @@ public class FunctionScoreTests extends ESTestCase {
|
||||||
weightFunctionStubs[i] = new WeightFactorFunction(weights[i], scoreFunctionStubs[i]);
|
weightFunctionStubs[i] = new WeightFactorFunction(weights[i], scoreFunctionStubs[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
Query filtersFunctionScoreQueryWithWeights = getFiltersFunctionScoreQuery(
|
FiltersFunctionScoreQuery filtersFunctionScoreQueryWithWeights = getFiltersFunctionScoreQuery(
|
||||||
FiltersFunctionScoreQuery.ScoreMode.MULTIPLY
|
FiltersFunctionScoreQuery.ScoreMode.MULTIPLY
|
||||||
, CombineFunction.REPLACE
|
, CombineFunction.REPLACE
|
||||||
, weightFunctionStubs
|
, weightFunctionStubs
|
||||||
|
@ -548,10 +554,10 @@ public class FunctionScoreTests extends ESTestCase {
|
||||||
assertThat(explainedScore / scoreWithWeight, is(1f));
|
assertThat(explainedScore / scoreWithWeight, is(1f));
|
||||||
}
|
}
|
||||||
|
|
||||||
public void checkWeightOnlyCreatesBoostFunction() throws IOException {
|
public void testWeightOnlyCreatesBoostFunction() throws IOException {
|
||||||
Query filtersFunctionScoreQueryWithWeights = new MinScoreQuery(new FunctionScoreQuery(new MatchAllDocsQuery(), new WeightFactorFunction(2), CombineFunction.MULTIPLY, 100), 0f);
|
FunctionScoreQuery filtersFunctionScoreQueryWithWeights = new FunctionScoreQuery(new MatchAllDocsQuery(), new WeightFactorFunction(2), 0.0f, CombineFunction.MULTIPLY, 100);
|
||||||
TopDocs topDocsWithWeights = searcher.search(filtersFunctionScoreQueryWithWeights, 1);
|
TopDocs topDocsWithWeights = searcher.search(filtersFunctionScoreQueryWithWeights, 1);
|
||||||
float score = topDocsWithWeights.scoreDocs[0].score;
|
float score = topDocsWithWeights.scoreDocs[0].score;
|
||||||
assertThat(score, equalTo(2.0f));
|
assertThat(score, equalTo(2.0f));
|
||||||
}
|
}
|
||||||
}
|
}
|
|
@ -143,6 +143,10 @@ public class QueryPhaseTests extends ESTestCase {
|
||||||
}
|
}
|
||||||
|
|
||||||
public void testMinScoreDisablesCountOptimization() throws Exception {
|
public void testMinScoreDisablesCountOptimization() throws Exception {
|
||||||
|
TestSearchContext context = new TestSearchContext();
|
||||||
|
context.parsedQuery(new ParsedQuery(new MatchAllDocsQuery()));
|
||||||
|
context.setSize(0);
|
||||||
|
|
||||||
final AtomicBoolean collected = new AtomicBoolean();
|
final AtomicBoolean collected = new AtomicBoolean();
|
||||||
IndexSearcher contextSearcher = new IndexSearcher(new MultiReader()) {
|
IndexSearcher contextSearcher = new IndexSearcher(new MultiReader()) {
|
||||||
protected void search(List<LeafReaderContext> leaves, Weight weight, Collector collector) throws IOException {
|
protected void search(List<LeafReaderContext> leaves, Weight weight, Collector collector) throws IOException {
|
||||||
|
@ -151,19 +155,11 @@ public class QueryPhaseTests extends ESTestCase {
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
TestSearchContext context = new TestSearchContext();
|
|
||||||
context.parsedQuery(new ParsedQuery(new MatchAllDocsQuery()));
|
|
||||||
context.setSize(0);
|
|
||||||
context.preProcess();
|
|
||||||
QueryPhase.execute(context, contextSearcher);
|
QueryPhase.execute(context, contextSearcher);
|
||||||
assertEquals(0, context.queryResult().topDocs().totalHits);
|
assertEquals(0, context.queryResult().topDocs().totalHits);
|
||||||
assertFalse(collected.get());
|
assertFalse(collected.get());
|
||||||
|
|
||||||
context = new TestSearchContext();
|
|
||||||
context.parsedQuery(new ParsedQuery(new MatchAllDocsQuery()));
|
|
||||||
context.setSize(0);
|
|
||||||
context.minimumScore(1);
|
context.minimumScore(1);
|
||||||
context.preProcess();
|
|
||||||
QueryPhase.execute(context, contextSearcher);
|
QueryPhase.execute(context, contextSearcher);
|
||||||
assertEquals(0, context.queryResult().topDocs().totalHits);
|
assertEquals(0, context.queryResult().topDocs().totalHits);
|
||||||
assertTrue(collected.get());
|
assertTrue(collected.get());
|
||||||
|
|
|
@ -31,7 +31,6 @@ import org.elasticsearch.common.HasContextAndHeaders;
|
||||||
import org.elasticsearch.common.HasHeaders;
|
import org.elasticsearch.common.HasHeaders;
|
||||||
import org.elasticsearch.common.ParseFieldMatcher;
|
import org.elasticsearch.common.ParseFieldMatcher;
|
||||||
import org.elasticsearch.common.collect.ImmutableOpenMap;
|
import org.elasticsearch.common.collect.ImmutableOpenMap;
|
||||||
import org.elasticsearch.common.lucene.search.MinScoreQuery;
|
|
||||||
import org.elasticsearch.common.util.BigArrays;
|
import org.elasticsearch.common.util.BigArrays;
|
||||||
import org.elasticsearch.index.IndexService;
|
import org.elasticsearch.index.IndexService;
|
||||||
import org.elasticsearch.index.analysis.AnalysisService;
|
import org.elasticsearch.index.analysis.AnalysisService;
|
||||||
|
@ -130,9 +129,6 @@ public class TestSearchContext extends SearchContext {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void preProcess() {
|
public void preProcess() {
|
||||||
if (minScore != null) {
|
|
||||||
this.query = new MinScoreQuery(query, minScore);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -379,6 +375,11 @@ public class TestSearchContext extends SearchContext {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Float minimumScore() {
|
||||||
|
return minScore;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public SearchContext sort(Sort sort) {
|
public SearchContext sort(Sort sort) {
|
||||||
return null;
|
return null;
|
||||||
|
|
Loading…
Reference in New Issue