Evidence observability for RAG systems
SourceMapR traces every answer back to the exact document evidence that produced it — and shows how that evidence was created (parse → chunk → embed → retrieve → answer).
Your app stays the same — SourceMapR shows the evidence lineage.
Drop-in instrumentation for the most popular RAG frameworks
Full pipeline instrumentation
Callback-based tracing
Track token usage, latency, and costs
More frameworks coming soon: Haystack · Semantic Kernel · DSPy
SourceMapR traces every step of your RAG pipeline
For every answer, SourceMapR shows you: which chunks were retrieved, their similarity scores, where they came from in the original document (with PDF highlighting), what prompt was sent to the LLM, and how many tokens were used. Debug your RAG pipeline without guessing.
Trace every response to the exact chunks that were retrieved. See similarity scores and rankings.
Click any chunk to see it highlighted in the original PDF. Split view shows chunk and source side by side.
See the exact prompt sent to the model, the response, token counts, and latency for every query.
Organize runs into experiments. Compare "chunk-size-256" vs "chunk-size-512" side by side.
Add SourceMapR to your existing pipeline. Your code stays the same — we just watch.
# pip install sourcemapr llama-index
from sourcemapr import init_tracing, stop_tracing
init_tracing(endpoint="http://localhost:5000")
# Your existing LlamaIndex code — unchanged
from llama_index.core import SimpleDirectoryReader, VectorStoreIndex
documents = SimpleDirectoryReader("./papers").load_data()
index = VectorStoreIndex.from_documents(documents)
response = index.as_query_engine().query("What is attention?")
print(response)
stop_tracing()
http://localhost:5000 to see the full evidence lineage.
Built for developers who are tired of print-debugging RAG pipelines. Run locally. Own your data. See your evidence.