TestParallelReaderpublic class TestParallelReader extends TestCase Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements. See the NOTICE file distributed with
this work for additional information regarding copyright ownership.
The ASF licenses this file to You under the Apache License, Version 2.0
(the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. |
Fields Summary |
---|
private Searcher | parallel | private Searcher | single |
Methods Summary |
---|
private org.apache.lucene.store.Directory | getDir1()
Directory dir1 = new MockRAMDirectory();
IndexWriter w1 = new IndexWriter(dir1, new StandardAnalyzer(), true);
Document d1 = new Document();
d1.add(new Field("f1", "v1", Field.Store.YES, Field.Index.TOKENIZED));
d1.add(new Field("f2", "v1", Field.Store.YES, Field.Index.TOKENIZED));
w1.addDocument(d1);
Document d2 = new Document();
d2.add(new Field("f1", "v2", Field.Store.YES, Field.Index.TOKENIZED));
d2.add(new Field("f2", "v2", Field.Store.YES, Field.Index.TOKENIZED));
w1.addDocument(d2);
w1.close();
return dir1;
| private org.apache.lucene.store.Directory | getDir2()
Directory dir2 = new RAMDirectory();
IndexWriter w2 = new IndexWriter(dir2, new StandardAnalyzer(), true);
Document d3 = new Document();
d3.add(new Field("f3", "v1", Field.Store.YES, Field.Index.TOKENIZED));
d3.add(new Field("f4", "v1", Field.Store.YES, Field.Index.TOKENIZED));
w2.addDocument(d3);
Document d4 = new Document();
d4.add(new Field("f3", "v2", Field.Store.YES, Field.Index.TOKENIZED));
d4.add(new Field("f4", "v2", Field.Store.YES, Field.Index.TOKENIZED));
w2.addDocument(d4);
w2.close();
return dir2;
| private org.apache.lucene.search.Searcher | parallel()
Directory dir1 = getDir1();
Directory dir2 = getDir2();
ParallelReader pr = new ParallelReader();
pr.add(IndexReader.open(dir1));
pr.add(IndexReader.open(dir2));
return new IndexSearcher(pr);
| private void | queryTest(org.apache.lucene.search.Query query)
Hits parallelHits = parallel.search(query);
Hits singleHits = single.search(query);
assertEquals(parallelHits.length(), singleHits.length());
for(int i = 0; i < parallelHits.length(); i++) {
assertEquals(parallelHits.score(i), singleHits.score(i), 0.001f);
Document docParallel = parallelHits.doc(i);
Document docSingle = singleHits.doc(i);
assertEquals(docParallel.get("f1"), docSingle.get("f1"));
assertEquals(docParallel.get("f2"), docSingle.get("f2"));
assertEquals(docParallel.get("f3"), docSingle.get("f3"));
assertEquals(docParallel.get("f4"), docSingle.get("f4"));
}
| protected void | setUp()
single = single();
parallel = parallel();
| private org.apache.lucene.search.Searcher | single()
Directory dir = new MockRAMDirectory();
IndexWriter w = new IndexWriter(dir, new StandardAnalyzer(), true);
Document d1 = new Document();
d1.add(new Field("f1", "v1", Field.Store.YES, Field.Index.TOKENIZED));
d1.add(new Field("f2", "v1", Field.Store.YES, Field.Index.TOKENIZED));
d1.add(new Field("f3", "v1", Field.Store.YES, Field.Index.TOKENIZED));
d1.add(new Field("f4", "v1", Field.Store.YES, Field.Index.TOKENIZED));
w.addDocument(d1);
Document d2 = new Document();
d2.add(new Field("f1", "v2", Field.Store.YES, Field.Index.TOKENIZED));
d2.add(new Field("f2", "v2", Field.Store.YES, Field.Index.TOKENIZED));
d2.add(new Field("f3", "v2", Field.Store.YES, Field.Index.TOKENIZED));
d2.add(new Field("f4", "v2", Field.Store.YES, Field.Index.TOKENIZED));
w.addDocument(d2);
w.close();
return new IndexSearcher(dir);
| public void | testDocument()
Directory dir1 = getDir1();
Directory dir2 = getDir2();
ParallelReader pr = new ParallelReader();
pr.add(IndexReader.open(dir1));
pr.add(IndexReader.open(dir2));
Document doc11 = pr.document(0, new MapFieldSelector(new String[] {"f1"}));
Document doc24 = pr.document(1, new MapFieldSelector(Arrays.asList(new String[] {"f4"})));
Document doc223 = pr.document(1, new MapFieldSelector(new String[] {"f2", "f3"}));
assertEquals(1, doc11.getFields().size());
assertEquals(1, doc24.getFields().size());
assertEquals(2, doc223.getFields().size());
assertEquals("v1", doc11.get("f1"));
assertEquals("v2", doc24.get("f4"));
assertEquals("v2", doc223.get("f2"));
assertEquals("v2", doc223.get("f3"));
| public void | testFieldNames()
Directory dir1 = getDir1();
Directory dir2 = getDir2();
ParallelReader pr = new ParallelReader();
pr.add(IndexReader.open(dir1));
pr.add(IndexReader.open(dir2));
Collection fieldNames = pr.getFieldNames(IndexReader.FieldOption.ALL);
assertEquals(4, fieldNames.size());
assertTrue(fieldNames.contains("f1"));
assertTrue(fieldNames.contains("f2"));
assertTrue(fieldNames.contains("f3"));
assertTrue(fieldNames.contains("f4"));
| public void | testIncompatibleIndexes()
// two documents:
Directory dir1 = getDir1();
// one document only:
Directory dir2 = new MockRAMDirectory();
IndexWriter w2 = new IndexWriter(dir2, new StandardAnalyzer(), true);
Document d3 = new Document();
d3.add(new Field("f3", "v1", Field.Store.YES, Field.Index.TOKENIZED));
w2.addDocument(d3);
w2.close();
ParallelReader pr = new ParallelReader();
pr.add(IndexReader.open(dir1));
try {
pr.add(IndexReader.open(dir2));
fail("didn't get exptected exception: indexes don't have same number of documents");
} catch (IllegalArgumentException e) {
// expected exception
}
| public void | testQueries()
queryTest(new TermQuery(new Term("f1", "v1")));
queryTest(new TermQuery(new Term("f1", "v2")));
queryTest(new TermQuery(new Term("f2", "v1")));
queryTest(new TermQuery(new Term("f2", "v2")));
queryTest(new TermQuery(new Term("f3", "v1")));
queryTest(new TermQuery(new Term("f3", "v2")));
queryTest(new TermQuery(new Term("f4", "v1")));
queryTest(new TermQuery(new Term("f4", "v2")));
BooleanQuery bq1 = new BooleanQuery();
bq1.add(new TermQuery(new Term("f1", "v1")), Occur.MUST);
bq1.add(new TermQuery(new Term("f4", "v1")), Occur.MUST);
queryTest(bq1);
|
|