Informer
Newsletter of the BCS Information Retrieval Specialist Group
  • Informer Home
  • Issues
    • Autumn 2018
    • Winter 2019
    • Spring 2019
    • Summer 2019
    • Autumn 2019
    • Winter 2020
    • Spring 2020
    • Summer 2020
    • Autumn 2020
    • Winter 2021
    • Spring 2021
    • Summer 2021
    • Autumn 2021
    • Winter 2022
    • Spring 2022
    • Summer 2022
  • Articles by Topic
  • Authors
  • About Informer
Browse: Home / 2016 / January / My first attendance at Search Solutions

My first attendance at Search Solutions

By Tu Bui on 31st January 2016

As a PhD student just reaching the end of my first year at the Centre for Vision Speech and Signal Processing (CVSSP) at the University of Surrey, I was delighted to receive a postgraduate bursary from the British Computer Society IRSG to support my attendance at Search Solutions 2015.

Search Solutions was an ideal opportunity to gain a broad appreciation of information retrieval techniques, being directly aligned with my PhD research on robust, scalable search of image collections. Although many of the works showcased at Search Solutions focus on text based information retrieval, many contemporary visual search algorithms borrows heavily from techniques rooted in that field for their scalability e.g. Bag of Words, TF-IDF and Inverse Indexing – and of course deep learning is now being important for describing documents in both the image and text domains.

My work explores sketch-based image retrieval (SBIR); an emerging sub-field within visual search where a user provides a free-hand sketch as the query to search through potentially millions of database images. Although visual content is being generated at a staggering rate (Facebook 350M photos/day, Instagram 60M photos/day, Youtube 300 video hours/day), technology for the management of visual media has not kept pace with its generation. Most multimedia search engines are predominantly still using textual queries to search visual media, and whilst text efficiently conveys semantic concepts (e.g. find me a flower) it is neither intuitive nor concise to describe appearance in this manner (e.g. find me a flower that looks like this, or a video containing movement like this).

Sketching has been proved to be an efficient communication method since the age of caveman. Children can sketch without much effort before learning a language. Also, humans seem to have no difficulty interpreting sketch despite its sparsity and ambiguity. On top of that, the dramatic development of touch-screen devices such as tablet and smartphone makes sketching a trivial task. In fact, there are already several applications utilising search based sketch. Detexify helps Latex users find the code for any Maths symbol by simply drawing it. Google Android wear recently supports emoji recogniser where a user can sketch a clumsy emoji and the system will return the closest matchings. However, these applications are highly domain specific and their datasets are quite small. We have yet to see a practical SBIR system that robustly addresses generic image searching in a scalable manner.

Therefore the opportunity to hear about scalable search solutions in both an academic and enterprise context was of great relevance to me, and Search Solutions has broadened my horizons in this respect. I was impressed by how fast the search technologies have evolved. A few decades ago, a search engine could only handle Boolean expressions. Nowadays, it is able to communicate with us in a more “human” way. As demonstrated by Behshad Behzadi from Google, current search engines can deal with sophisticated questions like “How tall is the husband of Kim Kardashian?“. Not only does it have answers for most general questions about the world (by crawling a massive dataset of web pages) but is also smart enough to answer these questions within user context (e.g. “What is my frequent fly number?” by mining the owner’s data e.g. email stored within the cloud). Behzadi predicts the future of search engine to be an ultimate mobile assistant. I myself imagine that search engine would, one day, become a virtual friend which the user can seek advice for their daily life activities e.g. “What should I wear for the party tonight?“. If a search engine could analyse life-long data sources (email, browsing history, life-logged content), it should be able to figure out the style, habits, hobbies, interests and personality of its user. Also, a search engine with vision and hearing (e.g. camera, microphone, gyroscope, and other sensors integrated into a mobile phone) should understand more about the context of the conversation (e.g. “Is the owner happy at the moment?“) and give advice accordingly.

Our approach to the SBIR problem up to now is from Computer Vision (CV) perspective. We employ complex CV techniques to decode images into numerical feature vectors so that the distance between any two vectors in the feature space reflects the visual similarity between the two corresponding images. We attempt to address the partial deformation caused by human imperfection while sketching. Colour is also integrated into our framework as the second search modality. Our latest system can achieve interactive speeds on a dataset of 12M images (and was presented at ICCV 2015 this year). However, there are still many aspects to improve, including the need to develop a more precise “image2vec” encoding, a more scalable indexing and a capability to let a user refine their search results.

From the CV perspective, I wonder what the future of visual search will look like. Visual media online is growing rapidly with Cisco predicting that by 2019 over 80% of global consumer Internet traffic will comprise visual content. I feel that search over this media will not be deemed competent until CV systems can interpret images in the same way as human do. Particularly, once we solve the “semantic gap” between human powerful contextual language and low-level machine programming, a search engine with visual support can become a great human assistant.

 

About Tu Bui
Tu Bui

Tu Bui is a PhD student in the Centre for Vision, Speech and Signal Processing (CVSSP) at University of Surrey. His main research focuses on large scale Visual Search and Object Classification related to sketch. His work involves developing machine learning techniques to automatically recognise sketch drawn by amateur users and its applications in Content-based Image Retrieval. Tu gained his BEng degree in Electronics Engineering from University of Surrey in 2014.

« Previous Next »

Search

Recent comments

  • Tony Russell-Rose on And finally….from the Editor
  • New Informer – Winter 2021 | Information Interaction on ECIR 2022 and ECIR 2023 – locations confirmed
  • New Informer – Winter 2021 | Information Interaction on Strix Lecture 2020 event – 26 November 2020
  • New Informer – Winter 2021 | Information Interaction on Editorial
  • New Informer – Autumn 2020 | Information Interaction on ECIR 2021 – planning for a virtual conference

Categories

  • Autumn 2022
  • News and alerts
  • ECIR2023
  • Summer 2022
  • Spring 2022
  • IRSG management
  • Winter 2022
  • Autumn 2021
  • ECIR 2022
  • Summer 2021
  • ECIR 2021 Conference Supplement
  • Awards
  • Spring 2021
  • Winter 2021
  • Autumn 2020
  • Summer 2020
  • Spring 2020
  • Winter 2020
  • Autumn 2019
  • Summer 2019
  • Spring 2019
  • Winter 2019
  • Autumn 2018
  • Summer 2018
  • Spring 2018
  • Winter 2018
  • Autumn 2017
  • Summer 2017
  • Spring 2017
  • Winter 2017
  • Autumn 2016
  • Summer 2016
  • Spring 2016
  • Winter 2016
  • Autumn 2015
  • Promotion
  • Summer 2015
  • Spring 2015
  • Winter 2015
  • Autumn 2014
  • Summer 2014
  • Spring 2014
  • Winter 2014
  • Autumn 2013
  • Summer 2013
  • Org Overview
  • Spring 2013
  • Winter 2013
  • Conference Review
  • Feature Article
  • Editorial
  • Events
  • Book Review
  • Autumn 2012
  • Summer 2012
  • Spring 2012
  • Winter 2012
  • Uncategorized

Tags

awards BCS Boolean City University clustering conference conferences design ECIR editorial enterprise seach enterprise search events Faceted search facets HCIR information architecture information discovery Information Retrieval information seeking interaction design IR IR practice IRSG log analysis MSR multimedia retrieval navigation recruitment search Search Solutions search strategies sensemaking site search ss12 survey taxonomy text analytics tutorial user experience user study wayfinding web search weka workshop

Authors

  • Agnes Molnar (1)
  • Alberto Purpura (1)
  • Aldo Lipani (1)
  • Alejandra Gonzalez-Beltran (1)
  • Allan Hanbury (1)
  • Amit Kumar Jaiswal (1)
  • Andy Macfarlane (50)
  • Benjamin Kille (1)
  • Benno Stein (1)
  • Birger Larsen (1)
  • Carsten Eickhoff (1)
  • Cathal Gurrin (8)
  • Charlie Hull (2)
  • Chris Madge (1)
  • Thomas Mandl (1)
  • Claudia Hauff (1)
  • Colin Wilkie (1)
  • David Elsweiler (1)
  • David Haynes (1)
  • David Maxwell (1)
  • Deirdre Lungley (1)
  • Dennis Aumiller (2)
  • Djoerd Hiemstra (1)
  • Franco Maria Nardini (1)
  • Frank Hopfgartner (13)
  • Gabriel Tanase (1)
  • Gabriella Kazai (5)
  • Giorgio Maria Di Nunzio (1)
  • Haiming Liu (2)
  • Helen Clegg (1)
  • Helen Lippell (1)
  • Iadh Ounis (1)
  • Ingo Frommholz (9)
  • Joao Magalheis (4)
  • Jochen L. Leidner (3)
  • John Tait (7)
  • Jolanta Pietraszko (1)
  • Jon Chamberlain (4)
  • Jose Alberto Equivel (1)
  • Julie Glanville (1)
  • Kamran Abbasi (1)
  • Katherine Allen (3)
  • Kurt Kragh Sørensen (1)
  • Linda Achilles (1)
  • Luca Soldaini (1)
  • Marc Sloan (2)
  • Marco Palomino (2)
  • Marianne Sweeny (1)
  • Marina Santini (1)
  • Markus Schedl (1)
  • Martin White (144)
  • Mateusz Dubiel (1)
  • Michael Oakes (1)
  • Mike Salampasis (1)
  • Mohammad Aliannejadi (1)
  • Morgan Harvey (1)
  • Nandita Tripathi (1)
  • Natasha Chowdory (1)
  • Norbert Fuhr (1)
  • Olivia Foulds (1)
  • Parth Mehta (2)
  • Paul Cleverley (1)
  • Paul Matthews (2)
  • Pedro Ruas (1)
  • Philipp Mayr (1)
  • Roland Roller (1)
  • Roman Kern (1)
  • Ronan Cummins (2)
  • Sam Marshall (1)
  • Samuel Dodson (1)
  • Selina Meyer (1)
  • Silviu Paun (1)
  • Song Chen (1)
  • Stefan Rueger (1)
  • Stephane Goldstein (1)
  • Stephanie Segura Rodas (1)
  • Steven Zimmerman (8)
  • Thanh Vu (1)
  • Tony Russell-Rose (31)
  • Trung Huynh (1)
  • Tu Bui (1)
  • Tyler Tate (8)
  • Udo Kruschwitz (35)
  • Val Gillet (1)

Copyright © 2023 Informer.

Powered by WordPress and Hybrid.