Technical
Infrastructure
The high-precision instrument powering enterprise document intelligence. Built for scale, security, and sub-second latency.
Response Latency
Global Edge Network Performance
terminalKernel Stack
"Architecture optimized for asynchronous I/O and vector search embedding throughput."
System Topology
API + Auth Layer
FastAPI on Uvicorn (ASGI). OIDC JWT validated against ZITADEL JWKS endpoint on every request. Permission cache in Redis for sub-millisecond lookups.
AI Processing
Celery async workers: text extraction (pdfplumber/python-docx), LangChain classification + summarization, OpenAI embeddings, pgvector chunk indexing.
Data Layer
PostgreSQL 16 + pgvector for primary data and semantic search. Redis 7 for Celery queues and permission cache. MinIO (S3-compatible) for all file objects.
Real-Time Layer
WebSocket endpoint for live presence, cursor tracking, and lock notifications. SSE streaming for document Q&A token-by-token delivery.
Scalability & Resource Profiles
| Tier | Compute Node | Storage Throughput | Max Concurrency | Deployment |
|---|---|---|---|---|
| Sandbox | 2 vCPU / 4 GB RAM | 250 MB/s | 50 ops/sec | Docker Compose |
| Standard | 8 vCPU / 32 GB RAM | 1.2 GB/s | 450 ops/sec | K8s / Helm v3 |
| Enterprise High-Perf | 32 vCPU / 128 GB RAM | 5.5 GB/s (NVMe) | 2,500+ ops/sec | Bare Metal / K8s |
AI Processing Throughput
The AI pipeline runs as a Celery async task. Classification and summarization complete within 5 minutes of upload. Files are streamed directly to MinIO — the API server never buffers file bytes in memory. Permission lookups are cached in Redis and invalidated immediately on mutation.
- Alembic zero-downtime DB migrations (expand/contract)
- Full version history — no version ever deleted
- Multi-tenant ORM-level org_id isolation
- LangChain pluggable AI provider (OpenAI / Claude / Gemini / Ollama)
Ready for Infrastructure Integration?
Detailed API documentation, OpenAPI schemas, and deployment scripts are available in the partner portal.