Skip to content

A Tool for Querying Images Using Contextual Knowledge on Everyday Scenes

License

Notifications You must be signed in to change notification settings

MU-Data-Science/QIK

Repository files navigation

QIK

Querying Images Using Contextual Knowledge is a large-scale image retrieval system for complex everyday scenes. It combines advances in deep learning and natural language processing in a novel way to capture the relationships between objects in everyday scenes. It uses XML to represent tree representations of captions (e.g., parse tree) and stores and indexes them in an XML database. Given a query image, it constructs an XPath query based on the image's caption and identifies a set of candidate matches by querying the XML database. For ranking the candidates, it uses the tree edit distance between the parse tree (or dependency tree) of query caption and that of a candidate's caption. Recently, QIK+ was developed to extend the core concepts of QIK to video retrieval.

For more details, see the below publications.

Publications

  • Arun Zachariah and Praveen Rao - Video Retrieval for Everyday Scences With Common Objects. In the Annual ACM International Conference on Multimedia Retrieval (ICMR 2023), 6 pages, Thessaloniki, Greece. (to appear)

  • Arun Zachariah, Mohamed Gharibi, and Praveen Rao - A Large-Scale Image Retrieval System for Everyday Scenes. In the 2nd ACM International Conference on Multimedia in Asia (MM Asia 2020), 3 pages, Singapore. [PDF] [DOI]

  • Arun Zachariah, Mohamed Gharibi, and Praveen Rao - QIK: A System for Large-Scale Image Retrieval on Everyday Scenes With Common Objects. In the Annual ACM International Conference on Multimedia Retrieval (ICMR 2020), pages 126-135, Dublin, Ireland. [PDF] [DOI]

Errata

We made an inadvertent error in computing the mean average precision (MAP) values reported in the ACM ICMR 2020 paper. Our sincere apologies for this human error. The correct numbers are reported here.

System Setup

Execute scripts/deploy_scripts/init.sh

. init.sh [--home Home] [--qik QIK_Home] [-h | --help]

Eg:

cd scripts/deploy_scripts && . init.sh --home /mydata --qik /mydata/QIK

To setup the demo:

./demo_setup.sh [--home Home] [--qik QIK_Home] [--system 120k | 15k | unsplash] [-h | --help]

Eg:

cd scripts/deploy_scripts && ./demo_setup.sh --home /mydata --qik /mydata/QIK --system 120k

To start the web engine.

cd QIK_Web && python manage.py runserver <IP>:8000

The UI can be accessed at http://<IP>:8000/search/image_search

To construct the index.

List of images can be added to MetaDataGenerator/images.txt or the directory containing the images can be added to MetaDataGenerator/constants.py under IMAGE_DIR

To start the indexing process:

cd MetaDataGenerator && python process_images.py

Initial Evaluation Results

Initial evaluation results reported in the paper can be found here.

To reproduces the results, follow the steps in the README.

The scripts were tested on Ubuntu 16.04. If you would need assistance in setting up the system or if you have any other concerns, feel free to email Arun Zachariah and Praveen Rao. You can also report any issues here.

Video Retrieval

The implementation of QIK on videos is implemented here.

To cite using BibTeX

If you use QIK's code for your research, please cite our publications as below:

@inproceedings{10.1145/3372278.3390682,
  author = {Zachariah, Arun and Gharibi, Mohamed and Rao, Praveen},
  title = {QIK: A System for Large-Scale Image Retrieval on Everyday Scenes With Common Objects},
  year = {2020},
  publisher = {Association for Computing Machinery},
  address = {New York, NY, USA},
  doi = {10.1145/3372278.3390682},
  booktitle = {Proceedings of the 2020 International Conference on Multimedia Retrieval},
  pages = {126–135},
  numpages = {10},
  location = {Dublin, Ireland},
  series = {ICMR ’20}
}
@inproceedings{10.1145/3444685.3446253,
  author = {Zachariah, Arun and Gharibi, Mohamed and Rao, Praveen},
  title = {A Large-Scale Image Retrieval System for Everyday Scenes},
  year = {2021},
  publisher = {Association for Computing Machinery},
  address = {New York, NY, USA},
  doi = {10.1145/3444685.3446253},
  booktitle = {Proceedings of the 2nd ACM International Conference on Multimedia in Asia},
  articleno = {72},
  numpages = {3},
  location = {Virtual Event, Singapore},
  series = {MMAsia '20}
}

Contributors

Faculty: Praveen Rao (PI)

PhD Students: Arun Zachariah, Mohamed Gharibi

Acknowledgments

This work was supported by the National Science Foundation under Grant No. 1747751. (NSF IUCRC Center for Big Learning)

About

A Tool for Querying Images Using Contextual Knowledge on Everyday Scenes

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published