When pages get edited multiple times during indexing, duplicate entries end up getting inserted.
Lucene's inability to read and write on the same index makes this unnecessarily difficult to do right. Placing
updates on our own per-database queues, replacing duplicates during that time, and then applying
updates direct instead of through an in-memory directory might work reasonably well.
Version: unspecified
Severity: normal
URL: http://en.wikipedia.org/wiki/Special:Search?search=hurricane+dennis&fulltext=Search