You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Binary quantization and hamming distance are critical for scaling multi-vector representations (i.e., Colbert).
It looks as though hamming for binary vectors has already been implemented.
While a hamming-based maxsim can be implemented over this with a postgres function per approach here, is this something that might be supported/optimized within the library?
Beyond this, is an unpack_bits operation to convert a binary vector into a float representation (to improve accuracy in a subsequent rerank step) something contemplated?
The text was updated successfully, but these errors were encountered:
michaelbridge
changed the title
Binary quantization and hamming distance
Binary maxsim via hamming
Feb 17, 2025
Ah, looks like RaBitQ is an implementation of binary quant, so perhaps this question is better restated as, how can Colbert/Colpali late interaction be optimized within this framework?
Binary quantization and hamming distance are critical for scaling multi-vector representations (i.e., Colbert).
It looks as though hamming for binary vectors has already been implemented.
While a hamming-based maxsim can be implemented over this with a postgres function per approach here, is this something that might be supported/optimized within the library?
Beyond this, is an
unpack_bits
operation to convert a binary vector into a float representation (to improve accuracy in a subsequent rerank step) something contemplated?The text was updated successfully, but these errors were encountered: