Retrieval-Augmented Generation

RAG Infrastructure

The retrieval layer your AI applications need. Give your models accurate, current context from your own data.

Trusted by leading companies

How RAG works

Retrieval-Augmented Generation combines search with AI to deliver accurate, grounded answers.

01

User asks a question

Natural language input from your chat interface.

02

Meilisearch retrieves context

Hybrid search finds the most relevant documents instantly.

03

LLM generates answer

AI synthesizes a response grounded in your actual content.

AI

"Based on your documentation, the recommended approach is to use the /search endpoint with the following parameters…"

Source: API ReferenceSource: Best Practices Guide

Everything RAG needs

A complete retrieval foundation for AI applications. No assembly required.

Vector storage

Store embeddings alongside your documents. No separate vector database needed.

  • Embeddings handled for you
  • Scales to millions of docs
  • 8+ AI providers

Hybrid retrieval

Combine keyword precision with semantic understanding for better context.

  • Tunable keyword/semantic mix
  • Best of both approaches
  • Results under 50ms

Context formatting

Control exactly what content gets sent to your LLM.

  • Pick which fields to send
  • Use simple templates
  • Include any metadata

The complete RAG toolkit

Everything you need to build AI applications with accurate, grounded responses.

Vector search

Store and query embeddings alongside your data. Stays fast at millions of documents.

Hybrid retrieval

Combine keyword and semantic search to fetch the right context. Better inputs lead to better AI answers.

Similar documents

Surface the closest matches to any item with one call. Perfect for "more like this" and related content.

Control what the LLM sees

Choose exactly which fields go into each prompt. Cleaner context produces sharper answers.

LangChain retriever

First-class integration. Drop into existing RAG pipelines with one import.

MCP server

Model Context Protocol for AI agents. Tool-using assistants with live data.

Fits your AI stack

First-class integrations with the tools you already use.

LangChain

Official retriever integration for Python and JavaScript.

from langchain_meilisearch import MeilisearchRetriever

LlamaIndex

Use as a vector store in your LlamaIndex pipelines.

from llama_index.vector_stores import MeilisearchVectorStore

MCP server

Model Context Protocol for Claude and other AI agents.

npx @modelcontextprotocol/server-meilisearch

Works with all major LLM providers

Native integrations for OpenAI, Anthropic, Mistral, Google, and more.

58 models from 13 providers

OpenAI

OpenAI

6 models

GPT-5
GPT-5 mini
GPT-4o
GPT-4o mini
o3
o4-mini
Anthropic

Anthropic

3 models

Claude Opus 4.7
Claude Sonnet 4.6
Claude Haiku 4.5
Google

Google

5 models

Gemini 3.1 Pro
Gemini 3 Flash
Gemini 2.5 Pro
Gemini 2.5 Flash
Gemini 2.5 Flash-Lite
Mistral

Mistral AI

7 models

Mistral Large 3
Mistral Small 4
Pixtral Large
Magistral Medium
Magistral Small
Devstral 2
Codestral 2501
Cohere

Cohere

2 models

Command A
Command R
DeepSeek

DeepSeek

3 models

DeepSeek V4 Pro
DeepSeek V4 Flash
DeepSeek R1
Bedrock

AWS Bedrock

4 models

Amazon Nova 2 Pro
Amazon Nova 2 Lite
Amazon Nova Micro
Amazon Nova 2 Sonic
HuggingFace

Hugging Face

6 models

Llama 4 Maverick
Llama 4 Scout
Qwen 3.5 72B
Gemma 4
Phi-4
Phi-4-mini
Ollama

Ollama

8 models

Llama 4 Scout
Llama 4 Maverick
Kimi K2.6
Qwen 3.5
Gemma 4
Mistral
Phi-4
DeepSeek R1
together.ai

Together AI

4 models

Llama 4 Maverick
Qwen 3.5 397B A17B
Gemma 4 31B
DeepSeek V4 Flash
Fireworks

Fireworks AI

4 models

Llama 4 Scout
Kimi K2.5
Qwen 3.5
DeepSeek V4 Flash
Cloudflare

Cloudflare AI

4 models

Llama 4 Scout
Kimi K2.6
Gemma 4
Mistral 7B
MoonshotAI

Moonshot AI

2 models

Kimi K2.6
Kimi K2.5
+

Custom

Any provider

Meilisearch is compatible with any model offering a REST API and tool calling capabilities.

Built for AI applications

The retrieval layer for any AI-powered experience.

Chatbots & assistants

Grounded answers from your actual content.

Semantic search

Find documents by meaning, not just keywords. "budget phones" finds "affordable devices".

Recommendations

Similar document API for "more like this" features. Products, articles, content.

Question answering

Extract answers from documents with source attribution. Full transparency.

Content discovery

Help users explore related content based on what they're viewing.

AI agents

Give tool-using AI assistants access to your data via MCP.

Frequently asked questions

RAG (Retrieval-Augmented Generation) is an AI architecture that combines search with language models. Before generating a response, the system retrieves relevant context from your data, grounding the AI's answer in real information rather than relying solely on training data.

Ready to get started?

Run RAG Infrastructure on Meilisearch Cloud, or self-host the open-source engine.