Fast Local Spatial Verification for Feature-Agnostic Large-Scale Image Retrieval

Joel Brogan, Aparna Bharati, Daniel Moreira, Anderson Rocha, Kevin W. Bowyer, Patrick J. Flynn, Walter J. Scheirer

Research output: Contribution to journalArticlepeer-review

20 Scopus citations

Abstract

Images from social media can reflect diverse viewpoints, heated arguments, and expressions of creativity, adding new complexity to retrieval tasks. Researchers working on Content-Based Image Retrieval (CBIR) have traditionally tuned their algorithms to match filtered results with user search intent. However, we are now bombarded with composite images of unknown origin, authenticity, and even meaning. With such uncertainty, users may not have an initial idea of what the search query results should look like. For instance, hidden people, spliced objects, and subtly altered scenes can be difficult for a user to detect initially in a meme image, but may contribute significantly to its composition. It is pertinent to design systems that retrieve images with these nuanced relationships in addition to providing more traditional results, such as duplicates and near-duplicates - and to do so with enough efficiency at large scale. We propose a new approach for spatial verification that aims at modeling object-level regions using image keypoints retrieved from an image index, which is then used to accurately weight small contributing objects within the results, without the need for costly object detection steps. We call this method the Objects in Scene to Objects in Scene (OS2OS) score, and it is optimized for fast matrix operations, which can run quickly on either CPUs or GPUs. It performs comparably to state-of-the-art methods on classic CBIR problems (Oxford 5K, Paris 6K, and Google-Landmarks), and outperforms them in emerging retrieval tasks such as image composite matching in the NIST MFC2018 dataset and meme-style imagery from Reddit.

Original languageEnglish
Article number9492816
Pages (from-to)6892-6905
Number of pages14
JournalIEEE Transactions on Image Processing
Volume30
DOIs
StatePublished - 2021

Funding

Manuscript received August 4, 2020; revised April 28, 2021 and June 25, 2021; accepted July 2, 2021. Date of publication July 21, 2021; date of current version August 6, 2021. This work was supported in part by Defense Advanced Research Projects Agency (DARPA), in part by the Air Force Research Laboratory (AFRL) under Agreement FA8750-16-2-0173 and Agreement FA8750-20-2-1004, in part by Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP) through the DéjàVu Project under Grant 2017/12646-3, in part by Coordenação de Aperfeiçoamento de Pessoal de Nível (CAPES) through the DeepEyes Grant, and in part by National Council for Scientific and Technological Development (CNPq) under Grant 304472/2015-8. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Francesco G. B. De Natale. (Corresponding author: Joel Brogan.) Joel Brogan is with the Multimodal Sensor Analytics (MSA) Group, Oak Ridge National Laboratory, Oak Ridge, TN 37830 USA (e-mail: broganjr@ ornl.gov). This manuscript has been authored in part by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).

Keywords

  • Image forensics
  • content retrieval
  • forgery detection
  • image retrieval

Fingerprint

Dive into the research topics of 'Fast Local Spatial Verification for Feature-Agnostic Large-Scale Image Retrieval'. Together they form a unique fingerprint.

Cite this