Multi-model AI platform

Build, deploy, and observe AI products from one developer-native platform.

VeloxAI unifies model routing, agents, tools, knowledge bases, workflow automation, billing, and observability behind clean OpenAI-compatible APIs.

/v1

Versioned API

SSE

Streaming ready

Qdrant

RAG vectors

OpenAI-compatible request

POST /v1/chat/completions
Authorization: Bearer pk_live_...
{
  "model": "gpt-4o-mini",
  "stream": true,
  "messages": [{ "role": "user", "content": "Summarize this ticket" }]
}

Agent Builder

Publish tool-using assistants

Knowledge Base

Cited answers from private docs

Workflow

Queue-backed AI automations

Analytics

Tokens, cost, logs, alerts

15 min

JWT access token TTL

/v1

Versioned public API

24h

Image URL TTL contract

0

Plaintext API keys stored

Platform

Everything needed to ship AI features, not just call a model.

VeloxAI keeps auth, API keys, model routing, agents, tools, RAG, billing, and observability aligned behind scoped services and typed contracts.

View platform
OpenAI-compatible

Unified Chat API

Route requests across OpenAI, Anthropic, Google, Mistral, and local models through one /v1/chat/completions contract with SSE streaming.

Learn more
Tools + memory

Agent Builder

Create draft agents, publish stable versions, attach tools and knowledge bases, then expose agent chat endpoints for apps.

Learn more
Private RAG

Knowledge Base

Ingest documents, URLs, and raw text into PostgreSQL metadata and Qdrant vectors for semantic search and cited answers.

Learn more
AI-native runs

Workflow Automation

Trigger workflows manually or by webhook, run AI and agent nodes, persist every node result, and queue execution with Redis.

Learn more
Async jobs

Image Tools

Generate, edit, upscale, remove backgrounds, describe, and detect image content with job tracking and storage-ready outputs.

Learn more
Quota-aware

Analytics + Billing

Track requests, latency, tokens, credit usage, errors, logs, alert rules, and plan limits from the same control plane.

Learn more
Model routing

Choose the right model for every request.

Use one request shape across premium hosted models and OpenAI-compatible local backends. Keep model entitlement, token usage, and cost visible from day one.

OpenAI

Fast general reasoning, multimodal chat, and broad application compatibility.

gpt-4ogpt-4o-minigpt-4-turboo1o3-mini

Anthropic

Long-form reasoning, safer assistant behavior, and agent orchestration workloads.

claude-opus-4claude-sonnet-4claude-haiku-4

Google

Low-latency multimodal tasks and broad context workflows.

gemini-2.0-flashgemini-2.5-pro

Mistral + Local

Regional deployments, open-weight model routing, and local OpenAI-compatible backends.

mistral-largemistral-smallllama-3.1-70bqwen2.5-72b
Agent Builder

Publish agents that can reason, retrieve, and take action.

Agents combine an LLM, system prompt, tools, memory, and knowledge bases. Draft safely, publish versions, and track sessions with usage and sources.

Draft

Edit prompt, model, memory, tools, and guardrails without touching live traffic.

Publish

Snapshot a reviewed configuration into immutable agent_versions.

Deploy

Expose /v1/agents/:id/chat with scoped API key access.

Observe

Capture sessions, messages, tool calls, sources, and token usage.

Pricing

Start free. Scale with credits, limits, and clear controls.

Free

$0

For local prototypes and API exploration.

  • - 100 credits/month
  • - 2 API keys
  • - 3 agents
  • - 20 RPM
  • - Starter model access
Starter

$29

For small teams shipping their first AI product.

  • - 3,000 credits/month
  • - 10 API keys
  • - 20 agents
  • - 100 RPM
  • - All public models
Most popular

$99

For production teams with heavier traffic.

  • - 12,000 credits/month
  • - 100 API keys
  • - Unlimited agents
  • - 500 RPM
  • - SSO and audit foundations
Enterprise

Custom

For organizations needing custom models and deployment controls.

  • - Custom credits
  • - Custom rate limits
  • - Dedicated support
  • - On-prem options
  • - SLA review
Q&A

Questions developers ask before shipping.

Short answers for architecture, security, billing, and AI workflow decisions.

Is VeloxAI OpenAI-compatible?

Yes. The core chat endpoint is /v1/chat/completions and returns OpenAI-style responses, including SSE chunks ending with data: [DONE].

Where are knowledge base vectors stored?

Vectors live in Qdrant. PostgreSQL stores knowledge base, document, and chunk metadata so search remains scalable and auditable.

Can agents call tools safely?

Agents can use built-in and custom tools. Custom code execution stays disabled until a hardened sandbox is configured.

Do API keys reveal full secrets later?

No. Full API keys are shown only on create or rotate. VeloxAI stores only hashes and displays prefixes afterward.

Does the platform include billing limits?

Yes. Requests pass through rate limit, credit, resource, and model entitlement checks before expensive work starts.

Ready to wire AI into your product?

Create an organization, verify email, generate a scoped API key, and call VeloxAI through production-shaped contracts.

Start building