What costs credits
Only successful generation runs consume credits:POST /excel/generatePOST /word/generatePOST /slides/generatePOST /excel/analyzePOST /text/humanize,/text/rewrite,/text/new-content
GET /runs,GET /runs/{run_id},GET /tasks/{task_id}GET /verify,GET /webhooksPOST /files(file upload),DELETE /images/{id}POST /preflight(preflight-only validation)GET /runs/{run_id}/downloadand edit-URL rendering- File storage and bandwidth (not billed separately in v1)
How a run is priced
Thecredits_used field returned on every completed response is the authoritative charge for that run. It’s calculated from the actual model-token usage across all LLM turns inside the run, plus the agent’s tool activity — not from a fixed per-format rate.
Prices depend on:
- Which models were used — this varies by request shape and your account tier.
- How much the agent had to iterate — a short prompt on a single sheet costs a fraction of a multi-sheet financial model that took 20 agent turns.
- Whether the run resumed from state — follow-up turns on an existing
run_idare cheaper because the agent doesn’t re-inspect the full document.
Overten doesn’t publish a static per-format rate card in v1. If you need budgeting numbers for procurement or capacity planning, contact sales@overten.ai for the pricing worksheet for your tier.
Balance check before a run
Before a generation request is accepted, Overten verifies that your balance is above a safety floor. If it isn’t:Required value is a conservative estimate — enough to complete the work. The actual deduction after a successful run is usually equal to or less than this estimate, because it reflects real usage rather than a worst-case projection.
This check runs before preflight and before any agent work starts, so a 402 response never costs credits.
When you’re charged
| Scenario | Credits charged? |
|---|---|
| Request rejected by balance check (402) | No |
Request rejected by preflight (preflight_failed) | No |
| Request rejected by auth or permissions | No |
| Agent runs and succeeds | Yes — see credits_used on the response |
| Agent runs and fails mid-execution | No (v1 behavior) |
| Sync call auto-promoted to async, then succeeds on the worker | Yes |
| Async worker job fails | No |
In v1, runs that crash after starting are not charged. This is intentionally customer-friendly behavior. If this policy changes, we’ll update this page and the changelog before enforcing it.
Checking your balance and topping up
Your credit balance is visible at app.overten.ai → Settings → Billing. Top-ups are applied immediately — the next request sees the updated balance. Every completed response includes acredits_used field, so you can reconcile spend from your own logs without needing dashboard access.
Audit trail
Every credit movement — deduction, top-up, refund — is recorded against your org. If you need to investigate a specific charge, reference therun_id from the response payload when opening a support ticket. The Overten team can pull the matching record and walk you through the breakdown.
Refunds
If a completed run produced something unusable — wrong format, corrupted output, or an edge case that slipped through preflight — open a support ticket with therun_id. We’ll review the internal record and issue a credit adjustment on a case-by-case basis.
Refunds are not issued for results that were generated as requested but didn’t match your expectations. If the agent produced an artifact for the prompt you submitted, the fix is a better prompt or a follow-up turn — not a refund. Work with support on patterns that misbehave and we’ll tune the agents.
Observability (enterprise)
Per-run cost attribution — which agent turn, which tool call, which model — is available to our team via our internal observability backend. File a support ticket with arun_id if you want a breakdown of why a specific run cost what it did.
Customer-side observability export (your own Langfuse or Datadog destination) is available on the enterprise tier. Contact sales@overten.ai to set it up.
