← Simulators

🔍 RAG Pipeline

Documents → embeddings → vector DB → query → top-k search → LLM.

RAG = Retrieval Augmented Generation. Flow:

1
DocumentsYour corpus (PDFs, docs, etc.)
2
EmbedConvert chunks → vectors
3
Vector DBStore embeddings (e.g. Pinecone, Weaviate)
4
QueryUser question → query embedding
5
SearchFind top-k closest vectors (ANN)
6
LLMSend query + retrieved chunks → generate answer