mirror of https://github.com/apache/lucene.git
LUCENE-4234: Exception when FacetsCollector is used with ScoreFacetRequest
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1364576 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
a956609a85
commit
aa28aff77e
|
@ -93,6 +93,9 @@ Bug Fixes
|
||||||
all queries. Made Scorer.freq() abstract.
|
all queries. Made Scorer.freq() abstract.
|
||||||
(Koji Sekiguchi, Mike McCandless, Robert Muir)
|
(Koji Sekiguchi, Mike McCandless, Robert Muir)
|
||||||
|
|
||||||
|
* LUCENE-4234: Exception when FacetsCollector is used with ScoreFacetRequest,
|
||||||
|
and the number of matching documents is too large. (Gilad Barkai via Shai Erera)
|
||||||
|
|
||||||
Build
|
Build
|
||||||
|
|
||||||
* LUCENE-4094: Support overriding file.encoding on forked test JVMs
|
* LUCENE-4094: Support overriding file.encoding on forked test JVMs
|
||||||
|
|
|
@ -72,12 +72,14 @@ public class FacetsCollector extends Collector {
|
||||||
protected ScoredDocIdCollector initScoredDocCollector(
|
protected ScoredDocIdCollector initScoredDocCollector(
|
||||||
FacetSearchParams facetSearchParams, IndexReader indexReader,
|
FacetSearchParams facetSearchParams, IndexReader indexReader,
|
||||||
TaxonomyReader taxonomyReader) {
|
TaxonomyReader taxonomyReader) {
|
||||||
|
boolean scoresNeeded = false;
|
||||||
for (FacetRequest frq : facetSearchParams.getFacetRequests()) {
|
for (FacetRequest frq : facetSearchParams.getFacetRequests()) {
|
||||||
if (frq.requireDocumentScore()) {
|
if (frq.requireDocumentScore()) {
|
||||||
return ScoredDocIdCollector.create(1000, true);
|
scoresNeeded = true;
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return ScoredDocIdCollector.create(indexReader.maxDoc(), false);
|
return ScoredDocIdCollector.create(indexReader.maxDoc(), scoresNeeded);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -9,7 +9,7 @@ import org.apache.lucene.search.DocIdSet;
|
||||||
import org.apache.lucene.search.DocIdSetIterator;
|
import org.apache.lucene.search.DocIdSetIterator;
|
||||||
import org.apache.lucene.search.Scorer;
|
import org.apache.lucene.search.Scorer;
|
||||||
import org.apache.lucene.util.ArrayUtil;
|
import org.apache.lucene.util.ArrayUtil;
|
||||||
import org.apache.lucene.util.OpenBitSet;
|
import org.apache.lucene.util.FixedBitSet;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
@ -52,7 +52,7 @@ public abstract class ScoredDocIdCollector extends Collector {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void collect(int doc) {
|
public void collect(int doc) {
|
||||||
docIds.fastSet(docBase + doc);
|
docIds.set(docBase + doc);
|
||||||
++numDocIds;
|
++numDocIds;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -103,7 +103,9 @@ public abstract class ScoredDocIdCollector extends Collector {
|
||||||
@SuppressWarnings("synthetic-access")
|
@SuppressWarnings("synthetic-access")
|
||||||
public ScoringDocIdCollector(int maxDoc) {
|
public ScoringDocIdCollector(int maxDoc) {
|
||||||
super(maxDoc);
|
super(maxDoc);
|
||||||
scores = new float[maxDoc];
|
// only matching documents have an entry in the scores array. Therefore start with
|
||||||
|
// a small array and grow when needed.
|
||||||
|
scores = new float[64];
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -111,7 +113,7 @@ public abstract class ScoredDocIdCollector extends Collector {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void collect(int doc) throws IOException {
|
public void collect(int doc) throws IOException {
|
||||||
docIds.fastSet(docBase + doc);
|
docIds.set(docBase + doc);
|
||||||
|
|
||||||
float score = this.scorer.score();
|
float score = this.scorer.score();
|
||||||
if (numDocIds >= scores.length) {
|
if (numDocIds >= scores.length) {
|
||||||
|
@ -167,7 +169,7 @@ public abstract class ScoredDocIdCollector extends Collector {
|
||||||
|
|
||||||
protected int numDocIds;
|
protected int numDocIds;
|
||||||
protected int docBase;
|
protected int docBase;
|
||||||
protected final OpenBitSet docIds;
|
protected final FixedBitSet docIds;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Creates a new {@link ScoredDocIdCollector} with the given parameters.
|
* Creates a new {@link ScoredDocIdCollector} with the given parameters.
|
||||||
|
@ -187,7 +189,7 @@ public abstract class ScoredDocIdCollector extends Collector {
|
||||||
|
|
||||||
private ScoredDocIdCollector(int maxDoc) {
|
private ScoredDocIdCollector(int maxDoc) {
|
||||||
numDocIds = 0;
|
numDocIds = 0;
|
||||||
docIds = new OpenBitSet(maxDoc);
|
docIds = new FixedBitSet(maxDoc);
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Returns the default score used when scoring is disabled. */
|
/** Returns the default score used when scoring is disabled. */
|
||||||
|
|
|
@ -0,0 +1,88 @@
|
||||||
|
package org.apache.lucene.facet.search;
|
||||||
|
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import org.apache.lucene.analysis.core.KeywordAnalyzer;
|
||||||
|
import org.apache.lucene.document.Document;
|
||||||
|
import org.apache.lucene.document.Field.Store;
|
||||||
|
import org.apache.lucene.document.StringField;
|
||||||
|
import org.apache.lucene.facet.index.CategoryDocumentBuilder;
|
||||||
|
import org.apache.lucene.facet.search.params.FacetSearchParams;
|
||||||
|
import org.apache.lucene.facet.search.params.ScoreFacetRequest;
|
||||||
|
import org.apache.lucene.facet.search.results.FacetResult;
|
||||||
|
import org.apache.lucene.facet.taxonomy.CategoryPath;
|
||||||
|
import org.apache.lucene.facet.taxonomy.TaxonomyWriter;
|
||||||
|
import org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyReader;
|
||||||
|
import org.apache.lucene.facet.taxonomy.directory.DirectoryTaxonomyWriter;
|
||||||
|
import org.apache.lucene.index.DirectoryReader;
|
||||||
|
import org.apache.lucene.index.IndexWriter;
|
||||||
|
import org.apache.lucene.index.IndexWriterConfig;
|
||||||
|
import org.apache.lucene.search.IndexSearcher;
|
||||||
|
import org.apache.lucene.search.MatchAllDocsQuery;
|
||||||
|
import org.apache.lucene.search.MultiCollector;
|
||||||
|
import org.apache.lucene.search.TopScoreDocCollector;
|
||||||
|
import org.apache.lucene.store.Directory;
|
||||||
|
import org.apache.lucene.util.IOUtils;
|
||||||
|
import org.apache.lucene.util.LuceneTestCase;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
* contributor license agreements. See the NOTICE file distributed with
|
||||||
|
* this work for additional information regarding copyright ownership.
|
||||||
|
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
* (the "License"); you may not use this file except in compliance with
|
||||||
|
* the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
public class TestFacetsCollector extends LuceneTestCase {
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFacetsWithDocScore() throws Exception {
|
||||||
|
Directory indexDir = newDirectory();
|
||||||
|
Directory taxoDir = newDirectory();
|
||||||
|
|
||||||
|
TaxonomyWriter taxonomyWriter = new DirectoryTaxonomyWriter(taxoDir);
|
||||||
|
IndexWriter iw = new IndexWriter(indexDir, new IndexWriterConfig(
|
||||||
|
TEST_VERSION_CURRENT, new KeywordAnalyzer()));
|
||||||
|
|
||||||
|
CategoryDocumentBuilder cdb = new CategoryDocumentBuilder(taxonomyWriter);
|
||||||
|
Iterable<CategoryPath> cats = Arrays.asList(new CategoryPath("a"));
|
||||||
|
for(int i = atLeast(2000); i > 0; --i) {
|
||||||
|
Document doc = new Document();
|
||||||
|
doc.add(new StringField("f", "v", Store.NO));
|
||||||
|
cdb.setCategoryPaths(cats);
|
||||||
|
iw.addDocument(cdb.build(doc));
|
||||||
|
}
|
||||||
|
|
||||||
|
taxonomyWriter.close();
|
||||||
|
iw.close();
|
||||||
|
|
||||||
|
FacetSearchParams sParams = new FacetSearchParams();
|
||||||
|
sParams.addFacetRequest(new ScoreFacetRequest(new CategoryPath("a"), 10));
|
||||||
|
|
||||||
|
DirectoryReader r = DirectoryReader.open(indexDir);
|
||||||
|
DirectoryTaxonomyReader taxo = new DirectoryTaxonomyReader(taxoDir);
|
||||||
|
|
||||||
|
FacetsCollector fc = new FacetsCollector(sParams, r, taxo);
|
||||||
|
TopScoreDocCollector topDocs = TopScoreDocCollector.create(10, false);
|
||||||
|
new IndexSearcher(r).search(new MatchAllDocsQuery(), MultiCollector.wrap(fc, topDocs));
|
||||||
|
|
||||||
|
List<FacetResult> res = fc.getFacetResults();
|
||||||
|
double value = res.get(0).getFacetResultNode().getValue();
|
||||||
|
double expected = topDocs.topDocs().getMaxScore() * r.numDocs();
|
||||||
|
assertEquals(expected, value, 1E-10);
|
||||||
|
|
||||||
|
IOUtils.close(taxo, taxoDir, r, indexDir);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
Loading…
Reference in New Issue