Capacity rehearsals
Push critical paths to 5x, 10x, 20x normal traffic. Find the third-party tag, the slow query, or the cache miss that breaks first.
Push your site to a target concurrent user count. Measure what real browsers actually see. Get a per-session report on every one of those users, with Web Vitals, network, console, and video. Not a request script's guess at it.
Performance Testing is live now. Testing Suite and Monitoring are coming.
What Evaluat calls performance testing is what other tools call load tests, stress tests, or spike tests. Different traffic shapes, same configuration. We use the umbrella term because the differences are knobs on one test, not separate features. Set the ramp-up, the steady-state, the ramp-down. Pick the scenarios. Hit go.
Every performance test in Evaluat is built from the same six pieces. Same controls on Starter as on Enterprise.
A scenario you build for a 1,000-user performance test runs as a deployment smoke test in CI and as a 5-minute production monitor. Same definition, same configuration UI, three different lifecycles. The maintenance burden is one place, not three.
Parameterise the scenario with datasets so virtual users follow different paths. Wire a project-level popup handler once so the cookie banner doesn't break every script.
More on scenariosEvery test produces a five-view report. Aggregate enough for the executive summary. Detailed enough to find why something broke for 14 users out of 42,000.
Reports have stable URLs and can be shared read-only with people outside the team.
The patterns we hear most often on demo calls. Different shapes of the same test, with the same forensic detail in the report.
Push critical paths to 5x, 10x, 20x normal traffic. Find the third-party tag, the slow query, or the cache miss that breaks first.
Compare two test runs side-by-side. If LCP regressed by 400ms after this deploy, you'll see it before customers do.
Set thresholds on LCP, INP, CLS. The test fails when the budget breaks. Stop shipping regressions.
Analytics, A/B testing, consent banners. Measure exactly how much each one costs you in INP under load.
Run from multiple regions with the right timezone and locale. See what customers in Amsterdam actually experience.
Something failed for 12 sessions out of 40,000. Open one. Watch the video. One-hour root cause, not one-week.