@ncrn-cornell

In Defense of MinHash Over SimHash

, and . Proceedings of the 17th International Conference on Artificial Intelligence and Statistics (AISTATS), 33, Reykjavik, Iceland, (2014)

Abstract

MinHash and SimHash are the two widely adopted Locality Sensitive Hashing (LSH) algorithms for large-scale data processing applications. Deciding which LSH to use for a particular problem at hand is an important question, which has no clear answer in the existing literature. In this study, we provide a theoretical answer (validated by experiments) that MinHash virtually always outperforms SimHash when the data are binary, as common in practice such as search. The collision probability of MinHash is a function of resemblance similarity (R), while the collision probability of SimHash is a function of cosine similarity (S). To provide a common basis for comparison, we evaluate retrieval results in terms of S for both MinHash and SimHash. This evaluation is valid as we can prove that MinHash is a valid LSH with respect to S, by using a general inequality S2<=R<=S2−S. Our worst case analysis can show that MinHash significantly outperforms SimHash in high similarity region. Interestingly, our intensive experiments reveal that MinHash is also substantially better than SimHash even in datasets where most of the data points are not too similar to each other. This is partly because, in practical data, often R>=Sz−S holds where z is only slightly larger than 2 (e.g., z<=2.1). Our restricted worst case analysis by assuming Sz−S<=R<=S2−S shows that MinHash indeed significantly outperforms SimHash even in low similarity region. We believe the results in this paper will provide valuable guidelines for search in practice, especially when the data are sparse.

Links and resources

Tags

community

  • @ncrn-cornell
  • @dblp
@ncrn-cornell's tags highlighted