Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Query for reranking KnnFloatVectorQuery with full-precision vectors #14009

Open
wants to merge 15 commits into
base: main
Choose a base branch
from
Original file line number Diff line number Diff line change
Expand Up @@ -99,7 +99,7 @@ public Query rewrite(IndexSearcher indexSearcher) throws IOException {
if (topK.scoreDocs.length == 0) {
return new MatchNoDocsQuery();
}
return createRewrittenQuery(reader, topK);
return createRewrittenQuery(reader, topK.scoreDocs);
}

private TopDocs searchLeaf(
Expand Down Expand Up @@ -255,18 +255,18 @@ protected TopDocs mergeLeafResults(TopDocs[] perLeafResults) {
return TopDocs.merge(k, perLeafResults);
}

private Query createRewrittenQuery(IndexReader reader, TopDocs topK) {
int len = topK.scoreDocs.length;
static Query createRewrittenQuery(IndexReader reader, ScoreDoc[] scoreDocs) {
int len = scoreDocs.length;

assert len > 0;
float maxScore = topK.scoreDocs[0].score;
float maxScore = scoreDocs[0].score;

Arrays.sort(topK.scoreDocs, Comparator.comparingInt(a -> a.doc));
Arrays.sort(scoreDocs, Comparator.comparingInt(a -> a.doc));
int[] docs = new int[len];
float[] scores = new float[len];
for (int i = 0; i < len; i++) {
docs[i] = topK.scoreDocs[i].doc;
scores[i] = topK.scoreDocs[i].score;
docs[i] = scoreDocs[i].doc;
scores[i] = scoreDocs[i].score;
}
int[] segmentStarts = findSegmentStarts(reader.leaves(), docs);
return new DocAndScoreQuery(docs, scores, maxScore, segmentStarts, reader.getContext().id());
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,117 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.lucene.search;

import static org.apache.lucene.search.AbstractKnnVectorQuery.createRewrittenQuery;

import java.io.IOException;
import java.util.Arrays;
import java.util.Objects;
import org.apache.lucene.index.FieldInfo;
import org.apache.lucene.index.FloatVectorValues;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.VectorSimilarityFunction;

/**
* A wrapper of KnnFloatVectorQuery which does full-precision reranking.
*
* @lucene.experimental
*/
public class RerankKnnFloatVectorQuery extends Query {

private final int k;
private final float[] target;
private final KnnFloatVectorQuery query;

/**
* Execute the KnnFloatVectorQuery and re-rank using full-precision vectors
*
* @param query the KNN query to execute as initial phase
* @param target the target of the search
* @param k the number of documents to find
* @throws IllegalArgumentException if <code>k</code> is less than 1
*/
public RerankKnnFloatVectorQuery(KnnFloatVectorQuery query, float[] target, int k) {
this.query = query;
this.target = target;
this.k = k;
}

@Override
public Query rewrite(IndexSearcher indexSearcher) throws IOException {
IndexReader reader = indexSearcher.getIndexReader();
Query rewritten = indexSearcher.rewrite(query);
// short-circuit: don't re-rank if we already got all possible results
if (query.getK() <= k) {
return rewritten;
}
Weight weight = indexSearcher.createWeight(rewritten, ScoreMode.COMPLETE_NO_SCORES, 1.0f);
HitQueue queue = new HitQueue(k, false);
for (var leaf : reader.leaves()) {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should this be switched to parallel execution similar to AbstractKnnVectorQuery?

Copy link
Contributor Author

@dungba88 dungba88 Nov 27, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Good question, I was using single-thread as a simple version and try to benchmark the latency first, since multi-thread could add some overhead as well. This class only does vector loading and similarity computation for a small set of vectors (k * oversample) so it's not as CPU-intensive as the AbstractKnnVectorQuery

I'll also try multi-thread and run the benchmark again. From the below benchmark, the re-ranking phase only adds a trivial amount of latency it might not help much. Also the benchmark code seems to force merge so there's only a single partition, we need to change so that there are multiple partitions.

Comment on lines +63 to +64
Copy link
Contributor

@shubhamvishu shubhamvishu Nov 28, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Here we have the access to IndexSearcher#getTaskExecutor and could use it to parallelize the work across segments(like we did earlier with some other query rewrites). But the HitQueue here isn't thread-safe. I don't know if using concurrency after making insertWithOverflow thread-safe would be really helpful since it looks like the added cost is cheap? or Maybe it will be?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

That's right. In order to apply parallelism we need to use a per-segment queue, then merge it like in AbstractKnnVectorQuery.mergeLeafResults. I think the added latency is already low, but still want to try if it helps.

Scorer scorer = weight.scorer(leaf);
if (scorer == null) {
continue;
}
FloatVectorValues floatVectorValues = leaf.reader().getFloatVectorValues(query.getField());
if (floatVectorValues == null) {
continue;
}
FieldInfo fi = leaf.reader().getFieldInfos().fieldInfo(query.getField());
if (fi == null) {
continue;
}
VectorSimilarityFunction comparer = fi.getVectorSimilarityFunction();
DocIdSetIterator iterator = scorer.iterator();
while (iterator.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) {
int docId = iterator.docID();
float[] vectorValue = floatVectorValues.vectorValue(docId);
float score = comparer.compare(vectorValue, target);
queue.insertWithOverflow(new ScoreDoc(leaf.docBase + docId, score));
}
}
int i = 0;
ScoreDoc[] scoreDocs = new ScoreDoc[queue.size()];
for (ScoreDoc topDoc : queue) {
scoreDocs[i++] = topDoc;
}
return createRewrittenQuery(reader, scoreDocs);
}

@Override
public int hashCode() {
int result = Arrays.hashCode(target);
result = 31 * result + Objects.hash(query, k);
return result;
}

@Override
public boolean equals(Object o) {
if (this == o) return true;
RerankKnnFloatVectorQuery that = (RerankKnnFloatVectorQuery) o;
return Objects.equals(query, that.query) && Arrays.equals(target, that.target) && k == that.k;
}

@Override
public void visit(QueryVisitor visitor) {
query.visit(visitor);
}

@Override
public String toString(String field) {
return getClass().getSimpleName() + ":" + query.toString(field) + "[" + k + "]";
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,124 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.lucene.search;

import java.util.HashMap;
import java.util.Map;
import java.util.Random;
import org.apache.lucene.codecs.lucene99.Lucene99HnswScalarQuantizedVectorsFormat;
import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
import org.apache.lucene.document.IntField;
import org.apache.lucene.document.KnnFloatVectorField;
import org.apache.lucene.index.DirectoryReader;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.IndexWriter;
import org.apache.lucene.index.IndexWriterConfig;
import org.apache.lucene.index.VectorSimilarityFunction;
import org.apache.lucene.store.ByteBuffersDirectory;
import org.apache.lucene.store.Directory;
import org.apache.lucene.tests.util.LuceneTestCase;
import org.apache.lucene.tests.util.TestUtil;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Test;

public class TestRerankKnnFloatVectorQuery extends LuceneTestCase {

private static final String FIELD = "vector";
private static final VectorSimilarityFunction VECTOR_SIMILARITY_FUNCTION =
VectorSimilarityFunction.COSINE;
private static final int NUM_VECTORS = 1000;
private static final int VECTOR_DIMENSION = 128;

private Directory directory;
private IndexWriterConfig config;

@Before
@Override
public void setUp() throws Exception {
super.setUp();
directory = new ByteBuffersDirectory();

// Set up the IndexWriterConfig to use quantized vector storage
config = new IndexWriterConfig();
config.setCodec(
TestUtil.alwaysKnnVectorsFormat(new Lucene99HnswScalarQuantizedVectorsFormat()));
}

@Test
public void testTwoPhaseKnnVectorQuery() throws Exception {
Map<Integer, float[]> vectors = new HashMap<>();

Random random = random();

int numVectors = atLeast(NUM_VECTORS);
int numSegments = random.nextInt(2, 10);

// Step 1: Index random vectors in quantized format
try (IndexWriter writer = new IndexWriter(directory, config)) {
for (int j = 0; j < numSegments; j++) {
for (int i = 0; i < numVectors; i++) {
float[] vector = randomFloatVector(VECTOR_DIMENSION, random);
Document doc = new Document();
int id = j * numVectors + i;
doc.add(new IntField("id", id, Field.Store.YES));
doc.add(new KnnFloatVectorField(FIELD, vector, VECTOR_SIMILARITY_FUNCTION));
writer.addDocument(doc);
vectors.put(id, vector);

writer.flush();
}
}
}

// Step 2: Run TwoPhaseKnnVectorQuery with a random target vector
try (IndexReader reader = DirectoryReader.open(directory)) {
IndexSearcher searcher = new IndexSearcher(reader);
float[] targetVector = randomFloatVector(VECTOR_DIMENSION, random);
int k = 10;
double oversample = random.nextFloat(1.5f, 3.0f);

KnnFloatVectorQuery knnQuery =
new KnnFloatVectorQuery(FIELD, targetVector, k + (int) (k * oversample));
RerankKnnFloatVectorQuery query = new RerankKnnFloatVectorQuery(knnQuery, targetVector, k);
TopDocs topDocs = searcher.search(query, k);

// Step 3: Verify that TopDocs scores match similarity with unquantized vectors
for (ScoreDoc scoreDoc : topDocs.scoreDocs) {
Document retrievedDoc = searcher.storedFields().document(scoreDoc.doc);
int id = retrievedDoc.getField("id").numericValue().intValue();
float[] docVector = vectors.get(id);
assert docVector != null : "Vector for id " + id + " not found";
float expectedScore = VECTOR_SIMILARITY_FUNCTION.compare(targetVector, docVector);
Assert.assertEquals(
"Score does not match expected similarity for doc ord: " + scoreDoc.doc + ", id: " + id,
expectedScore,
scoreDoc.score,
1e-5);
}
Comment on lines +107 to +113
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We can test that the results are sorted by exact distance.

Maybe we can also test that the result of the same query with oversample will be "at lease the same or better" than without oversample ? By "better" I mean we should have higher recall. But I'm not sure if it's deterministic

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thinking again, the docs should be sorted by ord, so my first point should be irrelevant.

}
}

private float[] randomFloatVector(int dimension, Random random) {
float[] vector = new float[dimension];
for (int i = 0; i < dimension; i++) {
vector[i] = random.nextFloat();
}
return vector;
}
}