Show HN: Opsmeter–attribute LLM spend to endpoints and prompt versions(no proxy)

Hi HN — I built Opsmeter, a lightweight LLM telemetry tool focused on cost attribution + budget control.

Provider dashboards mostly show totals. Opsmeter shows what caused the bill by breaking spend down by endpointTag, promptVersion, and optionally userId — plus latency and success/error rates.

It’s no-proxy: Opsmeter doesn’t sit in your request path. After each LLM call, you send a small telemetry payload to /v1/ingest/llm-request (provider, model, endpointTag, promptVersion, token counts, latency, status). Opsmeter normalizes cost via a provider/model pricing table and surfaces trends + regressions.

Links:

Home: https://opsmeter.io

Docs: https://opsmeter.io/docs

Pricing: https://opsmeter.io/pricing

If you try it and share anonymized screenshots/feedback, I’m happy to help you interpret the results — e.g.

which endpoints drive spend

which prompt versions increased tokens/cost (deploy regressions)

which users (optional) are the biggest cost drivers

suggested budget thresholds (80% warning / 100% exceeded) and alerting setup

Feedback welcome — especially on what you’d want next: staying telemetry-first, and potentially adding an optional gateway mode later.

2 points | by opsmeter 3 hours ago

0 comments