lucene/lucene/benchmark
Adrien Grand 48026bbc25 Remove `supportsConcurrency` flag from top-docs collector managers. (#13977)
Our collector managers have a `supportsConcurrency` flag to optimize the case
when they are used in a single thread. This PR proposes to remove this flag now
that the optimization doesn't do much as a result of #13943.
2024-11-05 15:37:36 +01:00
..
conf Udpate ReadTask to not rely on search(Query, Collector) (#13602) 2024-08-29 10:53:53 +02:00
scripts LUCENE-9383: benchmark module: Gradle conversion (#1550) 2020-06-05 17:57:53 -04:00
src Remove `supportsConcurrency` flag from top-docs collector managers. (#13977) 2024-11-05 15:37:36 +01:00
.gitignore LUCENE-9651 Update benchmark module docs (#759) 2022-03-23 14:51:28 -05:00
README.enwiki LUCENE-7438: Renovate benchmark module's support for highlighting 2016-10-07 09:57:11 -04:00
build.gradle Gradle build: cleanup of dependency resolution and consolidation of dependency versions (#13484) 2024-06-17 09:49:21 +02:00

README.enwiki

Support exists for downloading, parsing, and loading the English
version of wikipedia (enwiki).

The build file can automatically try to download the most current
enwiki dataset (pages-articles.xml.bz2) from the "latest" directory,
http://download.wikimedia.org/enwiki/latest/. However, this file
doesn't always exist, depending on where wikipedia is in the dump
process and whether prior dumps have succeeded. If this file doesn't
exist, you can sometimes find an older or in progress version by
looking in the dated directories under
http://download.wikimedia.org/enwiki/. For example, as of this
writing, there is a page file in
http://download.wikimedia.org/enwiki/20070402/. You can download this
file manually and put it in temp. Note that the file you download will
probably have the date in the name, e.g.,
http://download.wikimedia.org/enwiki/20070402/enwiki-20070402-pages-articles.xml.bz2.

If you use the EnwikiContentSource then the data will be decompressed on the fly
during the benchmark.  If you want to benchmark indexing, you should probably decompress
it beforehand using the "enwiki" Ant target which will produce a work/enwiki.txt, after
which you can use LineDocSource in your benchmark.

After that, ant enwiki should process the data set and run a load
test. Ant target enwiki will download, decompress, and extract (to individual files
in work/enwiki) the dataset, respectively.