We are pleased to announce the official release of the pixel retrieval benchmarks PRoxford and PRparis. These benchmarks are introduced in our ICCV 23 paper, "Towards Content-based Pixel Retrieval in Revisited Oxford and Paris."
Challenge: Can you identify the left query object in the right ground truth image?
Traditional image-level retrieval often falls short in meeting user search requirements. For a more precise and user-centric experience, pixel-level retrieval is essential.
Pixel retrieval isn't merely theoretical—it has tangible benefits. Our research shows that incorporating pixel-level annotations during the retrieval process significantly improves user experience.
Get Involved: Dive into our user study and share your insights here.
-
Download Pixel-Level Labels: Download and extract our pixel-level labels from here.
-
Download Example Result: We provide an example result using the 'sift sp' method. Download and extract it from this link.
-
Run
miou.py
in theevaluation_code
directory to get mIoUs for each query. -
Run
[email protected]
in theevaluation_code
directory to get the final mAP@k.
To test your own method, named 'ABC':
- Save the image-level and pixel-level results in
results/image-level/ABC
andresults/pixel-level/ABC/matching_result
, respectively. - Modify the method name in
miou.py
and[email protected]
. Run them.
Your directory structure should be as follows:
-- pixel_retrieval-Segmented_instance_retrieval
|-- pixel_retrieval_benchmarks
|-- evaluation_code
|-- results
Note: We are still refining this evaluation code. If you encounter any issues while using it, please let us know. We also plan to release configurations and code to help readers replicate the results presented in our paper. Stay tuned for updates.
Pixel retrieval, or segmented instance retrieval, is a new task within computer vision. The primary objective is to identify pixels that represent a specific query object within a database. This involves the machine's ability to recognize, localize, and segment the query object in database images in real-time. Essentially, it's a pixel-level and instance-level recognition task, as illustrated in the figure below:
Pixel retrieval can be viewed as an advanced form of image retrieval, providing insights into which specific pixels correlate with the query object. When juxtaposed with semantic and instance segmentation, pixel retrieval demands a nuanced recognition capability for targets of varying granularity. For a comprehensive understanding, please refer to our research paper.
We're excited to present the first two pixel retrieval benchmarks: PROxford and PRParis. These are built upon the renowned image retrieval datasets, ROxford and RParis. Opting for these datasets offers several advantages:
- Relevance to Landmarks: Landmarks are of immense business value. Consider the number of photos in your collection that feature landmarks.
- Familiarity: Utilizing well-known datasets like ROxford and RParis makes it easier for the image retrieval community to transition to this new task.
- Complexity: The task is inherently challenging. Each query can have hundreds of positive images, making it ideal for testing instance-level recognition.
- Quality: The annotations are of high caliber, facilitating comparisons between human and machine recognition performance.
Our comprehensive experiments highlight the inherent challenges of the pixel retrieval task. It stands apart from existing challenges, indicating that there's ample scope for research to enhance content-based pixel retrieval and, by extension, the user search experience.
[Our insights or directions for future research or developments.]