Framework to customize the Relevance Feedback Naive Scoring Formula, introduced in "Relevance Feedback in Qdrant" article, to your dataset (Qdrant collection), retriever, and feedback model.
As a result, you will get a, b and c formula parameters, which you can plug into the Qdrant's Relevance Feedback Query interface & increase the relevance of retrieval in your Qdrant collection.
from qdrant_client import QdrantClient
from qdrant_relevance_feedback import RelevanceFeedback
from qdrant_relevance_feedback.feedback import FastembedFeedback
from qdrant_relevance_feedback.retriever import QdrantRetriever
if __name__ == "__main__":
RETRIEVER_VECTOR_NAME = None # your named vector handle in Qdrant's collection or None if it's a default vector
COLLECTION_NAME = "document_collection"
# LIMIT controls the cost-quality tradeoff: the feedback model scores LIMIT documents per query to generate ground truth labels.
LIMIT = 50 # responses per query
client = QdrantClient(
url="https://xyz-example.eu-central.aws.cloud.qdrant.io",
api_key="your-api-key",
cloud_inference=True
)
retriever = QdrantRetriever("mixedbread-ai/mxbai-embed-large-v1", modality="text", embed_options={"lazy_load": True}) #lazy_load is an example of propagating options, instead of loading a model into memory straightaway, it loads it on the 1st use
feedback = FastembedFeedback("colbert-ir/colbertv2.0", score_options={"lazy_load": True})
relevance_feedback = RelevanceFeedback(
retriever=retriever,
feedback=feedback,
client=client,
collection_name=COLLECTION_NAME,
vector_name=RETRIEVER_VECTOR_NAME,
payload_key="document" # should refer to the raw data, which, after being embedded, is used in Qdrant retrieval. So, set this to the payload field that contains your original data you're searching for.
)
formula_params = relevance_feedback.train(
queries=None, # if you have real queries for training, provide a list here
amount_of_queries=200, # otherwise, you can specify amount of "synthetic queries" - documents randomly sampled from your collection
limit=LIMIT,
context_limit=5, # top responses used for mining context pairs
)
print('formula params are: ', formula_params)When using
QdrantRetriever, both local (via Fastembed) and cloud inference are supported. Setcloud_inference=Trueto use cloud inference,cloud_inference=Falseor just empty otherwise.
Warning: If your use case doesn’t involve document-to-document semantic similarity search, training on sampled documents ("synthetic queries") alone may completely cancel the effect of relevance feedback scoring on real data. It’s far more effective to use real queries.
In RelevanceFeedback class above, you're expected to fill in a payload_key.
This key should refer to the raw data, which, after being embedded, is used in Qdrant retrieval (with RETRIEVER_VECTOR_NAME). This data, in its original form (before it is embedded for dense retrieval), is used for the feedback model ground truth scoring.
If you're storing original data externally to Qdrant's collection, you should override retrieve_payload method of RelevanceFeedback class.
In order to use a custom retriever, you can define your class inherited from Retriever and override embed_query method.
Here's an example with overriding Retriever.
from typing import Any
from qdrant_relevance_feedback.retriever import Retriever
try:
from openai import OpenAI
except ImportError:
OpenAI = None
class OpenAIRetriever(Retriever):
def __init__(self, model_name: str, api_key: str, embed_options: dict[str, Any] | None = None, **kwargs: Any) -> None:
assert OpenAI is not None, 'OpenAIRetriever requires `openai` package to be installed`'
self._client = OpenAI(api_key=api_key, **kwargs)
self._model_name = model_name
self._embed_options = embed_options or {}
def embed_query(self, query: str) -> list[float]:
return self._client.embeddings.create(
model=self._model_name,
input=query,
**self._embed_options
).data[0].embeddingNote: However, if you plan to use OpenAI embeddings for retrieval, we recommend using Qdrant Cloud Inference in
QdrantRetrieverto optimize latency.
QdrantRetriever(model_name="openai/text-embedding-3-small"), embed_options={"openai-api-key" : "sk-proj-..."})Same for the feedback model, but this time you'd need to inherit from Feedback class and override score method.
from typing import Any
from qdrant_relevance_feedback.feedback import Feedback
try:
import cohere
except ImportError:
cohere = None
class CohereFeedback(Feedback):
def __init__(self, model_name: str, api_key: str, score_options: dict[str, Any] | None = None, **kwargs: Any):
self._api_key = api_key
self._model_name = model_name # E.g. "rerank-v4.0-pro"
self._score_options = score_options or {}
assert cohere is not None, "CohereFeedback requires `cohere` package"
self._client = cohere.ClientV2(api_key=api_key, **kwargs)
def score(self, query: str, responses: list[str]) -> list[float]:
response = self._client.rerank(
model=self._model_name,
query=query,
documents=responses,
top_n=len(responses),
**self._score_options
).results
feedback_model_scores = [
item.relevance_score for item in sorted(response, key=lambda x: x.index)
]
return feedback_model_scoresOnce you have created your classes, use them the same way as the builtin ones: create objects and pass to RelevanceFeedback.
from qdrant_client import QdrantClient
from qdrant_relevance_feedback import RelevanceFeedback
if __name__ == "__main__":
RETRIEVER_VECTOR_NAME = None
COLLECTION_NAME = "document_collection"
LIMIT = 50
CONTEXT_LIMIT = 5
client = QdrantClient(...)
retriever = OpenAIRetriever("text-embedding-3-small", api_key="<your openai api key>")
feedback = CohereFeedback("rerank-v4.0-pro", api_key="<your cohere api key")
relevance_feedback = RelevanceFeedback(
retriever=retriever,
feedback=feedback,
client=client,
collection_name=COLLECTION_NAME,
vector_name=RETRIEVER_VECTOR_NAME,
payload_key="document"
)
formula_params = relevance_feedback.train(
queries=None,
limit=LIMIT,
context_limit=CONTEXT_LIMIT,
)
print('formula params are: ', formula_params)If you have a collection of images, represented by image_url and file_name payload fields, payload usage is more complicated, as qdrant_client and fastembed expect images to be paths to files on disk. Here is an example that downloads images to disk and returns their file path.
import os
import shutil
from pathlib import Path
import requests
from qdrant_client import QdrantClient, models
from qdrant_relevance_feedback import RelevanceFeedback
from qdrant_relevance_feedback.feedback import FastembedFeedback
from qdrant_relevance_feedback.retriever import QdrantRetriever
class RelevanceFeedbackImageCache(RelevanceFeedback):
def retrieve_payload(self, responses: list[models.ScoredPoint]):
cache_dir = (
Path(os.getenv("XDG_CACHE_HOME", "~/.cache")).expanduser()
/ "relevance_feedback"
)
cache_dir.mkdir(exist_ok=True, parents=True)
responses_content = []
for p in responses:
if not (cache_dir / p.payload["file_name"]).is_file():
with (cache_dir / p.payload["file_name"]).open("wb") as f:
shutil.copyfileobj(
requests.get(p.payload["image_url"], stream=True).raw, f
)
responses_content.append(str(cache_dir / p.payload["file_name"]))
return responses_content
if __name__ == "__main__":
RETRIEVER_VECTOR_NAME = None
COLLECTION_NAME = "image_collection"
LIMIT = 50 # responses per query
CONTEXT_LIMIT = 5 # top responses used for mining context pairs
client = QdrantClient(
url="https://xyz-example.eu-central.aws.cloud.qdrant.io",
api_key="your-api-key",
cloud_inference=True
)
retriever = QdrantRetriever("Qdrant/clip-ViT-B-32-vision", modality="image") # Note, that you'll also have to tell `QdrantRetriever` that you are dealing with images, not text.
feedback = FastembedFeedback("Qdrant/colpali-v1.3-fp16")
relevance_feedback = RelevanceFeedbackImageCache(
retriever=retriever,
feedback=feedback,
client=client,
collection_name=COLLECTION_NAME,
vector_name=RETRIEVER_VECTOR_NAME,
)
formula_params = relevance_feedback.train(
queries=None, # if you have specific queries for training, provide a list here
amount_of_queries=200, # otherwise, you can specify amount of synthetic queries - documents sampled from your collection
limit=LIMIT,
context_limit=CONTEXT_LIMIT,
)
print('formula params are: ', formula_params)Evaluates the trained naive formula on two metrics: relative gain based on abovethreshold@N and Discounted Cumulative Gain (DCG) Win Rate@N.
Detailed explanation of these metrics and their meaning can be found in the "Relevance Feedback in Qdrant" article.
from qdrant_relevance_feedback.evaluate import Evaluator
n = 10 # as in metric@n
EVAL_CONTEXT_LIMIT = 3 # top responses used for mining context pairs (what you'll use in production for you retrieval pipelines)
evaluator = Evaluator(relevance_feedback=relevance_feedback)
# Similar to `relevance_feedback.train`, you can provide your own set of predefined queries by passing `eval_queries=[<queries>]`,
# or use synthetic queries sampled from your collection.
results = evaluator.evaluate_queries(
at_n=n,
formula_params=formula_params,
eval_queries=None,
amount_of_eval_queries=100,
eval_context_limit=EVAL_CONTEXT_LIMIT
)