Semantic Vector Encoding and Similarity Search Using Fulltext Search Engines

Investor logo

Warning

This publication doesn't include Institute of Computer Science. It includes Faculty of Informatics. Official publication website can be found on muni.cz.
Authors

RYGL Jan POMIKÁLEK Jan ŘEHŮŘEK Radim RŮŽIČKA Michal NOVOTNÝ Vít SOJKA Petr

Year of publication 2017
Type Article in Proceedings
Conference Proceedings of the 2nd Workshop on Representation Learning for NLP, RepL4NLP 2017 c/o ACL 2017
MU Faculty or unit

Faculty of Informatics

Citation
web
Doi http://dx.doi.org/10.18653/v1/W17-2611
Field Informatics
Keywords full-text search; similarity search; vector space; embeddings
Description Vector representations and vector space modeling (VSM) play a central role in modern machine learning. We propose a novel approach to ‘vector similarity searching’ over dense semantic representations of words and documents that can be deployed on top of traditional inverted-index-based fulltext engines, taking advantage of their robustness, stability, scalability and ubiquity. We show that this approach allows the indexing and querying of dense vectors in text domains. This opens up exciting avenues for major efficiency gains, along with simpler deployment, scaling and monitoring. The end result is a fast and scalable vector database with a tunable trade-off between vector search performance and quality, backed by a standard fulltext engine such as Elasticsearch. We empirically demonstrate its querying performance and quality by applying this solution to the task of semantic searching over a dense vector representation of the entire English Wikipedia.
Related projects:

You are running an old browser version. We recommend updating your browser to its latest version.

More info