CORS handled
Configure your origins. Call the API from the browser.
Last updated:
Behest is the AI backend you don't have to build. CORS handled, auth included, PII scrubbed, prompts defended — call from your frontend, ship in minutes.
Add AI to Your App is Behest's developer surface: an OpenAI-compatible /v1/chat/completions endpoint you call from the browser with CORS allowlists, JWT auth, Presidio PII redaction, Sentinel prompt-injection checks, Redis conversation memory, per-session cost capture, and token budgets — no bespoke Node proxy required.
✓ Includes employee token usage tracking
One API call. Behest handles the rest — so you focus on the features your users see.
Configure your origins. Call the API from the browser.
Multi-tenant JWT + per-project API keys, RBAC built in.
Persistent context across sessions — Redis sorted sets, per user.
Microsoft Presidio scrubs sensitive data before the LLM sees it.
Prompt-injection defense in-line on every request.
Includes employee token usage tracking and cost attribution out of the box.
The same gateway developers use on this page powers enterprise programs: finance-grade attribution and caps on Token FinOps, plus policy and audit controls on AI Governance.
Enforce hard budgets, attribute costs per session, and get complete visibility into your enterprise AI spend.
No backend proxy. Same OpenAI-compatible request shape your code already speaks.
// No backend proxy needed — CORS handled
// Call Behest directly from your browser
const response = await fetch(
"https://your-project.behest.app/v1/chat/completions",
{
method: "POST",
headers: {
"Authorization": `Bearer ${userToken}`,
"Content-Type": "application/json",
"X-End-User-Id": userId,
"X-Session-Id": sessionId,
},
body: JSON.stringify({
model: "gemini-2.5-flash",
messages: [{ role: "user", content: message }],
}),
}
);
// Conversation Memory persists across sessions
// PII scrubbed before reaching the LLM
// Per-user rate limits + token budgets enforced
// Per-session cost attribution captured automaticallyStop building auth, rate limiting, CORS proxies, and memory management. Behest deploys the complete AI backend in your cloud. You write the frontend.
Call from your browser — no backend proxy needed
Multi-tenant JWT + API keys with tenant isolation
Persistent conversation context across sessions
Automatic detection and scrubbing before it reaches the LLM
Sentinel blocks injection attacks with pattern detection
Three-tier limits per IP, project, and user
Traces, metrics, logs — correlated in Grafana
Deploy in your cloud — data never leaves your infra
Free Developer Edition. No credit card. No SDK required — standard fetch from any frontend.
✓ Includes employee token usage tracking
Enterprise Token FinOps: Enforce hard budgets and attribute costs per session.