Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: retrieval Clear Filter

Rerank-2.5 and rerank-2.5-lite: instruction-following rerankers

TL;DR – We are excited to introduce the rerank-2.5 series, which significantly improves upon rerank-2 ’s performance while also introducing instruction-following capabilities for the first time. On our standard suite of 93 retrieval datasets spanning multiple domains, rerank-2.5 and rerank-2.5-lite improve retrieval accuracy by 7.94% and 7.16% over Cohere Rerank v3.5. Furthermore, the new instruction-following feature allows users to steer the model’s output relevance scores using natural langua

rerank-2.5 and rerank-2.5-lite: instruction-following rerankers

TL;DR – We are excited to introduce the rerank-2.5 series, which significantly improves upon rerank-2 ’s performance while also introducing instruction-following capabilities for the first time. On our standard suite of 93 retrieval datasets spanning multiple domains, rerank-2.5 and rerank-2.5-lite improve retrieval accuracy by 7.94% and 7.16% over Cohere Rerank v3.5. Furthermore, the new instruction-following feature allows users to steer the model’s output relevance scores using natural langua

Beyond Retrieval: The Expanding Universe of Augmented Generation in AI

Introduction Standard large language models (LLMs) possess vast knowledge but struggle with limitations like hallucinations and accessing real-time information due to their static training data. This has spurred the development of dynamic AI architectures. Retrieval-Augmented Generation (RAG) has emerged as a key solution, integrating external knowledge into the generation process. However, the field is rapidly evolving beyond basic RAG. Newer models like RASG (Retrieval-Augmented Self-Generate

Muvera: Making multi-vector retrieval as fast as single-vector search

Neural embedding models have become a cornerstone of modern information retrieval (IR). Given a query from a user (e.g., “How tall is Mt Everest?”), the goal of IR is to find information relevant to the query from a very large collection of data (e.g., the billions of documents, images, or videos on the Web). Embedding models transform each datapoint into a single-vector “embedding”, such that semantically similar datapoints are transformed into mathematically similar vectors. The embeddings are