anirag.io

SOLUTIONS/DEVELOPER · API · TECH TEAMS

REST + TS-SDK + PYTHON-SDK · OPENAPI 3.1 · P99 < 800 MS

RAG that doesn't lock you in. Multi-LLM. Self-hostable. Open-source stack.

REST + TypeScript and Python SDKs, OpenAPI 3.1, p99 under 800 ms under load. Qdrant under Apache-2.0, Helm charts for self-hosting, multi-LLM router (Anthropic, OpenAI Ireland, Mistral, Aleph Alpha, Llama-EU). Token-level observability built in.

Public status page · Versioned changelog · Roadmap on Linear, public

PROBLEM · SOLUTION

What you know. How we solve it.

PAIN · 01

Build vs buy: own stack = six months (embedding pipeline, vector DB ops, eval, monitoring).

JTBD

API library integrated in two days, multi-tenant + RBAC + audit out of the box.

PAIN · 02

Avoid LLM provider lock-in — today OpenAI, tomorrow Claude, next month Mistral.

JTBD

Multi-LLM router with BYOK per provider. Routing rules by language, latency, cost, workspace.

PAIN · 03

'Does our RAG work?' isn't answerable without gold-standard datasets.

JTBD

Built-in eval framework: golden datasets, recall-at-k, citation accuracy, hallucination score.

PAIN · 04

Cost predictability — Pinecone + OpenAI embedding bills explode at scale.

JTBD

Token-level observability, hard cost caps per workspace, embedding cache built in.

Two days to production

Vom Anlass zur ersten Antwort.

  1. STEP · 01

    npm i @anirag/sdk

    Install the TypeScript or Python SDK, grab an API key (60 seconds in the web console), kick off your first embedding pipeline.

  2. STEP · 02

    Set up workspace + sources

    Multi-tenant workspaces via API. RBAC, SSO, audit log out of the box. Sources via REST or connectors.

  3. STEP · 03

    Route in production

    Configure multi-LLM router: language → model, workspace → region, latency budget → provider. Token-level observability into Datadog/Grafana.

Built for engineers

What's built in.

01

TypeScript + Python SDK

Both first-class. Postman/Bruno collection, asciinema quickstarts, GitHub examples repo.

02

OpenAPI 3.1 + gRPC

Full OpenAPI 3.1 spec, optional gRPC for high-throughput. Versioned, breaking changes via major bump.

03

Multi-LLM router

Anthropic, OpenAI Ireland, Mistral, Aleph Alpha, Llama-EU. BYOK. Routing by language, latency, cost.

04

Token observability

Token-level logs, latency histogram p50/p95/p99, embedding-cache hit rate. Datadog, Grafana, OpenTelemetry export.

05

Eval framework

Golden datasets, recall-at-k, citation accuracy, hallucination score. CI integration via @anirag/eval.

06

Self-hostable

Helm charts for Kubernetes. Open-source stack: Qdrant (Apache-2.0), Postgres, Redis. Sovereign plan.

Engineering-grade compliance

Compliance ist die Basis, nicht der Nachschlag.

  • Public status page with p50/p95/p99 per region — historical 99.96% uptime over the last 12 months.
  • Versioned changelog, breaking changes via major bump, 90-day deprecation period.
  • OpenTelemetry export for your own observability stack.
  • Open-source components documented: Qdrant Apache-2.0, Postgres, Redis. No vendor lock-in.
  • Public roadmap on Linear — feature requests and voting open.
What you save on build vs buy

Was Sie konkret einsparen.

Engineers who'd otherwise build RAG3
Weeks of engineering (own stack)24
Fully-loaded engineer hourly rate130

EINSPARUNG · ANNUALISIERT

374.400

2.880 Stunden pro Jahr freigesetzt

Assumption: 24 weeks build time at 3 engineers (40h week) at €130/h fully loaded. anirag integration: 2 days, 1 engineer.

Plugs into your stack

Integriert in Ihren Stack.

GIGitHub Actions
DADatadog
GRGrafana
OPOpenTelemetry
VEVercel
RARailway
FLFly.io
AWAWS Frankfurt
STIMMEN

What practitioners say.

Two days from npm install to production. Clean SDK, complete docs, eval framework caught two regression bugs for us."
A.T.Tech Lead · B2B SaaS, Series A
Multi-LLM router was the killer feature — Claude for DE, GPT for EN, Mistral for latency-critical. Routing in 12 lines of config."
P.K.Staff Engineer · Logistics SaaS
FAQ

What your compliance, IT and business teams ask.

Do you have a real TS/Python SDK or do I have to call REST raw?

Both first-class. @anirag/sdk (TS) and anirag-py (Python). Plus OpenAPI 3.1 for other languages. Postman/Bruno collection in our GitHub examples repo.

Self-hosting option if we grow?

Yes, Sovereign plan. Helm charts for Kubernetes, Terraform modules, OCI-compliant containers. Stack is Qdrant + Postgres + Redis — all open source.

What are your p50/p95/p99 latencies under load?

Public status page has historicals. p50 ~250ms, p95 ~600ms, p99 ~800ms (Frankfurt → Frankfurt, mid-tier models). Worst-case documented per model.

How do I debug why retrieval pulled chunk X over Y?

Trace view per query: embedding vector, top-k candidates with scores, reranking path, final context. Replay API for local repro.

Open-source components or proprietary black box?

Qdrant (Apache-2.0), Postgres, Redis are open-source. Reranker and eval framework are proprietary but ship as @anirag/eval, runnable locally.

Get your API key in 2 minutes.

Free tier, no credit card. 100k embeddings, 10k queries — enough for a real prototype.