Live

Trace is live — score groundedness in 280 ms.

DEVELOPERS

Build with Latence in minutes.

Real components shipping today. A typed SDK, a REST API, a CLI, and the open-source retrieval stack underneath them.

install
pip install latence
QUICKSTART

Pipeline to retrieval, in one file.

Three ways to integrate. The SDK, a plain REST call, or the CLI. Same concepts, same primitives.

quickstart.py
from latence import Latence

client = Latence()

# 1. Run the intelligence pipeline
pkg = client.pipeline.run(
    files=["contract_q4.pdf", "pricing.xlsx"]
)

# 2. Ingest the retrieval package into voyager-index
index = client.search.ingest(pkg, graph_sidecar=True)

# 3. Query with MaxSim, hybrid BM25, optional rerank
results = index.search(
    "Q4 renewal obligations", k=10, rerank=True
)

for r in results:
    print(r.score, r.source, r.snippet)
OPEN SOURCE

The stack underneath is open.

Real technical proof — not slideware. Every repo is maintained and publicly benchmarked.

REPO

voyager-index

On-prem retrieval engine. Late-interaction MaxSim, hybrid BM25, FastAPI.

REPO

vllm-factory

High-throughput encoder serving. ColBERT, ColPali, GLiNER plugins.

REPO

latenceai-dataset-intelligence

Pipeline that emits retrieval packages. Parse, chunk, extract, graph.

REPO

latence-python

Typed SDK. Async-first. Builder API for pipelines and search.

COMMUNITY

Built in public. Benchmarked in public.