🚀 Try Zilliz Cloud, the fully managed Milvus, for free—experience 10x faster performance! Try Now>>

Milvus
Zilliz
  • Home
  • AI Reference
  • What are the latest advances in zero-shot retrieval for semantic search?

What are the latest advances in zero-shot retrieval for semantic search?

Recent advances in zero-shot retrieval for semantic search focus on improving how models generalize to unseen tasks without requiring task-specific training data. A key development is the use of contrastive learning with synthetic data to train models that better capture semantic relationships. For example, Microsoft’s E5 model employs asymmetric contrastive learning, where synthetic query-document pairs are generated from a large text corpus. The model is trained to distinguish between relevant and irrelevant pairs by embedding queries and documents into a shared space, even when no labeled data exists for the target task. This approach allows the model to generalize to new domains by leveraging the broad patterns learned during pre-training. Models like E5 have shown strong performance on benchmarks like BEIR, which evaluates zero-shot retrieval across diverse datasets.

Another advancement involves refining model architectures to handle zero-shot scenarios more effectively. SGPT (GPT-based models for search) uses prompt-based methods to generate embeddings by feeding instructions or task descriptions into a GPT-style model. For instance, a prompt like "Find documents relevant to [query]" guides the model to produce embeddings that align with the retrieval objective. This method taps into the model’s existing knowledge from pre-training, eliminating the need for fine-tuning on specific tasks. Similarly, ColBERTer modifies the ColBERT architecture by enabling token-level interactions between queries and documents while reducing computational overhead. This allows it to efficiently rank documents based on semantic similarity without prior exposure to the target dataset, making it practical for scenarios where training data is unavailable.

Finally, hybrid approaches combining bi-encoders and cross-encoders have improved zero-shot retrieval pipelines. Bi-encoders (e.g., models that separately encode queries and documents) enable fast candidate retrieval, while cross-encoders (which process query-document pairs jointly) refine rankings with higher accuracy. For example, a system might use a bi-encoder like E5 to fetch an initial set of documents, then apply a cross-encoder like MiniLM to rerank them. This two-step process balances speed and precision, especially when dealing with unseen domains. Additionally, instruction-tuned models like InstructOR embed task descriptions (e.g., “Retrieve scientific papers about climate change”) directly into the embedding space, enabling flexible adaptation to new tasks. These innovations collectively address the challenge of generalizing across domains without labeled data, making zero-shot retrieval more practical for real-world applications.

Like the article? Spread the word