Vercel Analytics is genuinely useful. It tells you how many people visited your app, which pages they hit, and how your Core Web Vitals are trending. The moment a request hits an API route, Vercel Analytics goes dark.
What Vercel Analytics Actually Tracks
To be fair to Vercel, both products do exactly what they say. Web Analytics tracks visitor behavior — page views, bounce rate, referrers, and demographic breakdowns. Speed Insights gives you Core Web Vitals per route (LCP, FCP, CLS, INP) measured from real browser sessions.
- Page view counts and unique visitor estimates per route
- Top referrers and traffic sources
- Core Web Vitals (LCP, FCP, CLS, INP) from real browser sessions
- P75, P90, P95, P99 percentile breakdowns — for frontend page load performance only
- Custom events you manually instrument in client-side code
The key phrase is 'browser sessions.' Both products measure what happens from the user's browser to the first painted pixel. They are intentionally and explicitly frontend-only tools.
Your API Routes Are a Blind Spot
Every request that hits /api/* in your Next.js app, your Express backend, or your Hono worker — Vercel has no native tooling to tell you how those performed. No latency data. No error rate. No throughput. You can check your function logs, but that's not monitoring, it's archaeology.
- How long your /api/checkout route takes at P95 or P99
- What percentage of API requests are returning 4xx or 5xx errors
- Whether your error rate spiked at 2am on a Tuesday
- Which routes got hammered during a traffic spike
- How a slow database query is affecting specific endpoints
If your API route is timing out for 1 in 100 users, your Vercel dashboard will show a normal bounce rate. Your Core Web Vitals might even look fine — right up until users stop converting.
The Real Cost of Flying Blind in Production
Most production incidents don't announce themselves. They degrade slowly — a 200ms creep in P99 latency, a 2% error rate that climbs to 8% overnight. Without API monitoring, you're relying on users to file support tickets before you know something is broken.
The 30-minute detection window is brutal. In a typical web app, that's thousands of failed requests before anyone knows there's a problem. And the majority of teams are in that window.
Averages Lie — Why You Need Percentile Latency Per Route
Average response time is a lie. If 95% of your requests complete in 80ms but 5% take 2 seconds, your average might be 180ms — which sounds fine. It isn't. Those 5% are real users on real sessions who just experienced a broken product.
This is the P99 problem. Google added 500ms to search results in an experiment — traffic dropped 20%. Amazon found every 100ms of latency cost 1% in sales. These aren't enterprise problems; they happen at any scale where users have alternatives.
You cannot find this with Vercel Analytics. You need per-route, percentile-aware monitoring on the server side — and you need it running continuously, not just when you remember to check logs.
Adding API Monitoring in One Middleware Line
This doesn't require a Datadog contract or a week of configuration. For an Express or Next.js app, it's one import and one middleware call:
import * as statvisor from "@statvisor/sdk"
// Express
app.use(statvisor.express({ apiKey: process.env.STATVISOR_API_KEY! }))
// Hono
app.use(statvisor.hono({ apiKey: process.env.STATVISOR_API_KEY! }))
// Next.js — wrap your route handler
export const GET = statvisor.nextjs(handler, { apiKey: process.env.STATVISOR_API_KEY! })Statvisor tracks P50/P95/P99 latency per route and error rates in real time — no agents, no YAML, no 45-minute onboarding. If your /api/checkout route degrades at 2am, you'll see it in the dashboard within seconds of the first affected request.
Vercel Analytics and API Monitoring Are Complementary
This isn't about replacing Vercel Analytics — it's a solid tool for what it does. But a production API needs server-side visibility as a separate concern: which routes are slow, which are erroring, and where to look when something breaks.
- P50/P95/P99 latency per route — not global averages
- Error rate tracking — 4xx and 5xx trends over time
- Throughput monitoring to catch traffic anomalies before they cascade
- Route-level slow detection before users start filing complaints
Vercel Analytics for frontend behavior. Dedicated API monitoring for backend health. That's the full picture — and one of those halves is missing by default on every Vercel deployment.
Ready to monitor your API in production?
Statvisor gives you latency percentiles, error rates, and request volume for every route — in minutes, not days.
Get started free →