Pick your workload and enter expected demand. We add headroom, overhead, and growth automatically safely. See recommended vCPUs and RAM, then export instantly below.
| Scenario | Peak RPS | CPU ms / req | Users | Headroom | Growth | Suggested starting point |
|---|---|---|---|---|---|---|
| Small web API | 40 | 2.5 | 600 | 25% | 20% | 2 vCPU, 4 GB RAM |
| Moderate SaaS | 150 | 3.0 | 2,000 | 30% | 25% | 4 vCPU, 8–12 GB RAM |
| DB-heavy service | 90 | 4.0 | 1,200 | 35% | 30% | 8 vCPU, 16 GB RAM + premium disk |
RPS = (Users × RequestsPerMinutePerUser) ÷ 60Cores100 = RPS × (CpuMsPerRequest ÷ 1000)CoresBase = (Cores100 × (1 + CpuOverhead%)) ÷ TargetCpuUtil%RamBase = BaseAppRamGB + (Users × RamPerUserMB ÷ 1024) + OsOverheadGBRamUtil = RamBase ÷ TargetRamUtil%Multiplier = (1 + Headroom%) × (1 + Growth%)PerInstance = (Base × Multiplier) ÷ Instances| Profile | vCPU | RAM (GB) | Typical use |
|---|---|---|---|
| Nano | 1 | 1 | Low-traffic services, dev workloads |
| Micro | 2 | 2 | Small web apps, light APIs |
| Small | 2 | 4 | General web, small databases |
| Medium | 4 | 8 | Production apps, moderate DB |
| Large | 8 | 16 | Heavier services, analytics |
| XL | 16 | 32 | High load, large caches |
| 2XL | 32 | 64 | Big data nodes, high concurrency |
CPU demand is estimated from peak requests per second and average CPU milliseconds per request. For example, 150 RPS at 3 ms consumes 0.45 core at 100% utilization. After overhead and a 65% target, the baseline rises to roughly 0.73 vCPU before headroom and growth are applied.
RAM is calculated using application baseline memory plus per-user allocations, then adjusted for an operating-system buffer. Keeping a 70–80% memory target reduces swapping risk, lowers garbage-collection pressure, and helps caches remain warm. If your app uses in-memory sessions, increase per-user MB to reflect real heap usage.
Headroom protects performance during bursts, deploy spikes, and noisy-neighbor variance. Growth margin accounts for business expansion, new features, and traffic seasonality. A practical starting point is 20–35% headroom and 15–30% growth for fast-moving products. Mature services can reduce growth but should keep headroom for reliability.
When you scale horizontally, total demand is split across instances, producing a smaller per-instance recommendation. Availability modes influence total planned capacity: N+1 reserves enough capacity to survive a single node failure, while active-active can be modeled conservatively as doubling capacity to absorb failover without degradation.
Storage size impacts retention and log growth, but performance is often limited by IOPS and latency. If targets exceed ~2,000 IOPS, premium tiers are typically appropriate; above ~5,000 IOPS, provisioned performance becomes more predictable. Network throughput requirements above 300 Mbps usually justify higher bandwidth options and careful egress budgeting.
Treat the recommended VM size as a first-pass baseline, then validate with APM traces, load tests, and system metrics. Monitor CPU steal, memory paging, disk queue depth, and p95 latency during peak. Right-size after one to two release cycles, and keep a repeatable sizing record using the built-in CSV and PDF exports.
Use Direct RPS when you have monitoring or load-test data. Use Users when demand is early-stage and you can estimate per-user request rates during peak windows.
Start with APM averages from peak traffic. If you lack data, use 2–5 ms for lightweight APIs and 5–15 ms for heavier logic, then calibrate using load tests.
Running at 100% is unstable for most services. Targets like 60–70% CPU and 70–80% RAM preserve latency, handle spikes, and reduce risk from background tasks and contention.
N+1 increases total planned capacity to survive one instance failure. It does not increase the recommended per-instance size; it ensures remaining instances can carry the load.
Confirm disk latency and IOPS needs. Databases and log-heavy systems often require higher-performance tiers even with modest GB. Benchmark with real workloads to validate queue depth and p95 latency.
No. It is provider-agnostic and focuses on capacity math. Map the result to your platform’s instance families, then validate with metrics because vCPU performance differs by generation, region, and throttling policies.
Important Note: All the Calculators listed in this site are for educational purpose only and we do not guarentee the accuracy of results. Please do consult with other sources as well.