▲ | janalsncm 12 hours ago | ||||||||||||||||
I don’t think the author understands the purpose of reranking. During vector retrieval, we retrieve documents in sublinear time from a vector index. This allows us to reduce the number of documents from potentially billions to a much smaller number. The purpose of re-ranking is to allow high powered models to evaluate docs much more closely. It is true that we can attempt to distill that reranking signal into a vector index. Most search engines already do this. But there is no replacement for using the high powered behavior based models in reranking. | |||||||||||||||||
▲ | supo 5 hours ago | parent | next [-] | ||||||||||||||||
If you could wave a magic wand and push all the ranking signals into retrieval and that index would be fast to update and not that expensive to operate - you would do that and you would delete all your reranking systems, wouldn't you? | |||||||||||||||||
▲ | _QrE 12 hours ago | parent | prev | next [-] | ||||||||||||||||
I agree. > "The real challenge in traditional vector search isn't just poor re-ranking; it's weak initial retrieval. If the first layer of results misses the right signals, no amount of re-sorting will fix it. That's where Superlinked changes the game." Currently a lot of RAG pipelines use the BM25 algorithm for retrieval, which is very good. You then use an agent to rerank stuff only after you've got your top 5-25 results, which is not that slow or expensive, if you've done a good job with your chunking. Using metadata is also not really a 'new' approach (well, in LLM time at least) - it's more about what metadata you use and how you use them. | |||||||||||||||||
| |||||||||||||||||
▲ | laszlo_cravens 10 hours ago | parent | prev | next [-] | ||||||||||||||||
I agree as well. Especially in the context of recommendation systems, the decoupling of retrieval from a heavy ranker has a lot of benefits. It allows for 1) faster experimentation, and 2) the use of different retrieval sources. In reality, the retrieval might consist of a healthy mix of different algorithms (collaborative filtering, personalized page rank, word2vec/2tower embeddings, popular items near the user, etc.) and fallback heuristics | |||||||||||||||||
| |||||||||||||||||
▲ | nostrebored 12 hours ago | parent | prev [-] | ||||||||||||||||
That "much smaller number" is the tricky part. Most rerankers degrade substantially in quality over a few hundred candidates. No amount of powerful rerankers will make "high powered behavior based models" more effective. Those behavioral signals and intents have to be encoded in the query and the latent space. | |||||||||||||||||
|