Skip to content

FlagEmbedding LLM

LLM-based reranking via BAAI FlagEmbedding.

Overview

Field Value
Type LLM-based
Library FlagEmbedding
Default Model BAAI/bge-reranker-v2-gemma

Installation

pip install "autorag-research[gpu]"
# or
uv add "autorag-research[gpu]"

Configuration

_target_: autorag_research.rerankers.flag_embedding_llm.FlagEmbeddingLLMReranker
model_name: BAAI/bge-reranker-v2-gemma

Options

Option Type Default Description
model_name str BAAI/bge-reranker-v2-gemma FlagEmbedding LLM model name
use_fp16 bool False Use FP16 for inference
batch_size int 64 Batch size for multiple queries

Usage

from autorag_research.rerankers import FlagEmbeddingLLMReranker

reranker = FlagEmbeddingLLMReranker()
results = reranker.rerank("What is RAG?", ["doc1", "doc2", "doc3"], top_k=2)

When to Use

Good for higher quality reranking when compute resources are available. Uses LLM-scale models (e.g., Gemma) for more nuanced relevance scoring.