Coming soon
Contenox is yours.
Services runs the hard parts so you don't have to.
Contenox stays Apache 2.0.
Contenox Services is what you reach for when you don't want to run the infra.
When you use Contenox Services, your plans still live on your machine. Your context doesn't move. Your execution traces stay local. You're not handing your copilot to a cloud — you're pointing one config line at a different compute backend. The copilot stays yours; we just run the infrastructure.
Contenox connects to any provider: Ollama, OpenAI, Gemini, Vertex, vLLM. When you'd rather not operate GPU nodes or skill servers yourself, Contenox Services slots in as a provider config line. Same plans. Same copilot. Different compute backend.
Layer 1 — Hosted inference
Contenox Services runs inference on demand. You use it exactly like any other model provider: add it to your backend list and Contenox routes there when you need it.
- Demand-pull pricing: you pay for tokens, not idle capacity
- No GPU node to provision, patch, or restart
- Switch back to a local or third-party provider in one config change
- Same model selection semantics as any other provider you configure
This makes sense when a team reaches the point where running a spot A100 is cheaper than paying per-token to a major API — but doesn't want to operate the infra themselves.
Layer 2 — Hosted skills
Skills are HTTP services your plans can call — vector search, OCR, embeddings, document analytics. Running them yourself requires managing Vald clusters, ingestion pipelines, and worker processes. Hosted skills expose them as endpoints Contenox already knows how to reach.
- Vector search over your documents without a self-hosted Vald cluster
- OCR and document extraction as a callable tool in any plan
- Embeddings pipeline — index once, query from any Contenox instance
- Analytics: usage, latency, and plan execution traces
You can self-host any of these if you prefer. Hosted skills exist for teams that want the capability without the operational overhead.
Layer 3 — Contenox for Teams
Contenox for Teams gives everyone on your org their own account, shared plans, and tools that react to real events. It includes document search, bots, background workers, and an OpenAI-compatible API. Self-host it or let us run it as a managed service.
- Multi-user teams with roles and permissions — not just one developer's machine
- RAG pipeline with Vald — retrieval over your own documents and data
- Bots: Telegram, GitHub, and custom integrations that react to real events
- OpenAI-compatible API — drop-in for tools that already speak OpenAI
- On-prem or managed, your security boundary determines deployment
Early access
We're onboarding design partners for Contenox Services and Contenox for Teams.
Each pilot is scoped individually — reach out and we'll figure out what makes sense for your team.
See the full pricing breakdown on the Pricing page.