High-performance KNN similarity functions in Python, optimized for sparse matrices.
SimilariPy is primarily designed for Recommender Systems and Information Retrieval (IR) tasks, but can be applied to other domains as well.
The package also includes a set of normalization functions useful for pre-processing data before the similarity computation.
The official documentations is available at 📘 SimilariPy Guide
SimilariPy provides a range of high-performance similarity functions for sparse matrices.
All functions are multi-threaded and implemented in Cython + OpenMP for fast parallel computation on CSR matrixes.
- Dot Product – Simple raw inner product between vectors.
- Cosine – Normalized dot product based on L2 norm.
- Asymmetric Cosine – Skewed cosine similarity using an
alpha
parameter. - Jaccard, Dice, Tversky – Set-based generalized similarities.
- P3α – Graph-based similarity computed through random walk propagation with exponentiation.
- RP3β – Similar to P3α but includes popularity penalization using a
beta
parameter.
- S-Plus – A hybrid model combining Tversky and Cosine components, with full control over weights and smoothing.
For mathematical definitions and parameter details, see the 📘 SimilariPy Guide.
SimilariPy provides a suite of normalization functions for sparse matrix pre-processing.
All functions are implemented in Cython and can operate in-place on CSR matrixes for maximum performance and memory efficiency.
- L1, L2 – Applies row- or column-wise normalization.
- TF-IDF – Computes TF-IDF weighting with customizable term-frequency and IDF modes.
- BM25 – Applies classic BM25 weighting used in information retrieval.
- BM25+ – Variant of BM25 with additive smoothing for low-frequency terms.
For more details, check the 📘 SimilariPy Guide.
Here’s a minimal example to get you up and running with SimilariPy:
import similaripy as sim
import scipy.sparse as sps
# Create a random User-Rating Matrix (URM)
urm = sps.random(1000, 2000, density=0.025)
# Normalize the URM using BM25
urm = sim.normalization.bm25(urm)
# Train an item-item cosine similarity model
similarity_matrix = sim.cosine(urm.T, k=50)
# Compute recommendations for user 1, 14, 8
# filtering out already-seen items
recommendations = sim.dot_product(
urm,
similarity_matrix.T,
k=100,
target_rows=[1, 14, 8],
filter_cols=urm
)
SimilariPy can be installed from PyPI with:
pip install similaripy
To install the package and compile the Cython code, a GCC-compatible compiler with OpenMP is required.
Install the official dev-tools:
sudo apt update && sudo apt install build-essential
Install GCC with homebrew:
brew install gcc
Install the official Visual C++ Build Tools.
For Intel CPUs, using SciPy/Numpy with MKL (Math Kernel Library) is highly recommended for best performance. The easiest way to achieve this is to install them via Anaconda.
Package | Version |
---|---|
numpy | >= 1.21 |
scipy | >= 1.10.1 |
tqdm | >= 4.65.2 |
This library originated during the Spotify Recsys Challenge 2018.
Our team, The Creamy Fireflies, faced major challenges computing large similarity models on a dataset with over 66 million interactions. Standard Python/Numpy solutions were too slow as a whole day was required to compute one single model.
To overcome this, I developed high-performance versions of the core similarity functions in Cython and OpenMP. Encouraged by my teammates, I open-sourced this work to help others solve similar challenges.
Thanks to my Creamy Fireflies friends for the support! 🙏
This project is released under the MIT License.
If you use SimilariPy in your research, please cite:
@misc{boglio_simone_similaripy,
author = {Boglio Simone},
title = {bogliosimone/similaripy},
doi = {10.5281/zenodo.2583851},
url = {https://doi.org/10.5281/zenodo.2583851}
}