Skip to main content

AI Models Overview

Sercha Core supports optional integration with external AI models for enhanced search capabilities. These are entirely optional - Sercha Core works without any AI services using pure keyword (BM25) search.

Model Categories

Embedding Models

Embedding models convert text into numerical vectors, enabling semantic search. When configured, Sercha Core can find documents by meaning, not just keywords.

Use cases:

  • Semantic search ("find documents about machine learning" matches "ML algorithms")
  • Similar document discovery
  • Hybrid search combining keywords and meaning

Large Language Models

LLMs enhance query understanding and document processing. When configured, Sercha Core can rewrite queries and generate summaries.

Use cases:

  • Query expansion (adding synonyms, fixing typos)
  • Document summarization
  • Natural language query understanding

Graceful Degradation

Both model types are optional. Without them:

FeatureWith AI ModelsWithout AI Models
Keyword searchYesYes
Semantic searchYesNo
Query rewritingYesNo
SummarizationYesNo

Sercha Core degrades gracefully - core search functionality is always available.

Configuration

AI services are configured via environment variables:

# Embedding
EMBEDDING_PROVIDER=openai # openai | ollama | (empty for disabled)
EMBEDDING_API_KEY=sk-...
EMBEDDING_MODEL=text-embedding-3-small

# LLM
LLM_PROVIDER=openai # openai | anthropic | ollama | (empty for disabled)
LLM_API_KEY=sk-...
LLM_MODEL=gpt-4o-mini