lucene/contrib/benchmark
Michael McCandless c7f865a4c7 LCUENE-1716: allow control over storage of norms (body norms), info stream and whether docs properties should be indexed as fields
git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@788777 13f79535-47bb-0310-9956-ffa450edef68
2009-06-26 17:26:54 +00:00
..
conf LUCENE-1595: Separate DocMaker into DocMaker and ContentSource. 2009-06-18 19:58:59 +00:00
lib LUCENE-1591: Update commons-compress to the released version 1.0 2009-05-22 10:38:21 +00:00
src LCUENE-1716: allow control over storage of norms (body norms), info stream and whether docs properties should be indexed as fields 2009-06-26 17:26:54 +00:00
.rsync-filter LUCENE-848. Add Wikipedia benchmarking support 2007-07-01 02:19:10 +00:00
CHANGES.txt LCUENE-1716: allow control over storage of norms (body norms), info stream and whether docs properties should be indexed as fields 2009-06-26 17:26:54 +00:00
README.enwiki LUCENE-1591: workaround bug in xerces so we can process Wikipedia's XML 2009-04-08 21:50:33 +00:00
build.xml LUCENE-979: remove a few more old benchmark things 2009-06-14 14:52:02 +00:00
pom.xml.template LUCENE-1525: add missing maven dependencies for contrib/{benchmark,highlighter}; fix svn eolstyle to native 2009-01-21 16:59:39 +00:00

README.enwiki

Support exists for downloading, parsing, and loading the English
version of wikipedia (enwiki).

The build file can automatically try to download the most current
enwiki dataset (pages-articles.xml.bz2) from the "latest" directory,
http://download.wikimedia.org/enwiki/latest/. However, this file
doesn't always exist, depending on where wikipedia is in the dump
process and whether prior dumps have succeeded. If this file doesn't
exist, you can sometimes find an older or in progress version by
looking in the dated directories under
http://download.wikimedia.org/enwiki/. For example, as of this
writing, there is a page file in
http://download.wikimedia.org/enwiki/20070402/. You can download this
file manually and put it in temp. Note that the file you download will
probably have the date in the name, e.g.,
http://download.wikimedia.org/enwiki/20070402/enwiki-20070402-pages-articles.xml.bz2. When
you put it in temp, rename it to enwiki-latest-pages-articles.xml.bz2.

After that, ant enwiki should process the data set and run a load
test. Ant targets get-enwiki, expand-enwiki, and extract-enwiki can
also be used to download, decompress, and extract (to individual files
in work/enwiki) the dataset, respectively.

NOTE: This bug in Xerces:

  https://issues.apache.org/jira/browse/XERCESJ-1257

which is still present as of 2.9.1, causes an exception like this when
processing Wikipedia's XML:

Caused by: org.apache.xerces.impl.io.MalformedByteSequenceException: Invalid byte 2 of 4-byte UTF-8 sequence.
	at org.apache.xerces.util.ErrorHandlerWrapper.createSAXParseException(Unknown Source)
	at org.apache.xerces.util.ErrorHandlerWrapper.fatalError(Unknown Source)
	at org.apache.xerces.impl.XMLErrorReporter.reportError(Unknown Source)
	at org.apache.xerces.impl.XMLErrorReporter.reportError(Unknown Source)
	at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source)
	at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source)
	at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
	at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
	at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
	at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
	at org.apache.lucene.benchmark.byTask.feeds.EnwikiDocMaker$Parser.run(EnwikiDocMaker.java:77)
	... 1 more

The original poster in the Xerces bug provided this patch:

--- UTF8Reader.java	2006-11-23 00:36:53.000000000 +0100
+++ /home/rainman/lucene/xerces-2_9_0/src/org/apache/xerces/impl/io/UTF8Reader.java	2008-04-04 00:40:58.000000000 +0200
@@ -534,6 +534,16 @@
                     invalidByte(4, 4, b2);
                 }
 
+                // check if output buffer is large enough to hold 2 surrogate chars
+                if( out + 1 >= offset + length ){
+                    fBuffer[0] = (byte)b0;
+                    fBuffer[1] = (byte)b1;
+                    fBuffer[2] = (byte)b2;
+                    fBuffer[3] = (byte)b3;
+                    fOffset = 4;
+                    return out - offset;
+		}
+
                 // decode bytes into surrogate characters
                 int uuuuu = ((b0 << 2) & 0x001C) | ((b1 >> 4) & 0x0003);
                 if (uuuuu > 0x10) {

which I've applied to Xerces 2.9.1 sources, and committed under
lib/xerces-2.9.1-patched-XERCESJ-1257.jar.  Once XERCESJ-1257 is fixed
we can upgrade to a standard Xerces release.