Cloud Costs in 2026: When to Stay, When to Leave, and When to Go Hybrid
Cloud pricing has gotten more complex, not cheaper. This guide breaks down when cloud-first still makes sense, when bare metal saves you real money, and the specific cost traps (egress, storage tiers, GPU instances) that catch teams off guard.


Here's the uncomfortable truth about cloud costs in 2026: most companies paying over $10K/month haven't done the math on whether they should be.
Not because the cloud is bad — it really is the right choice for many workloads. But because cloud pricing is designed to be easy to start and hard to audit, and most teams inherit their infrastructure decisions from a time when the bill was $500/month and nobody cared.
This guide is the math. No vendor cheerleading, no "cloud is dead" hot takes. Just frameworks for figuring out where your money should actually go.
SolderMag Take: the cloud isn't expensive — your cloud is expensive
The base cost of a cloud VM is rarely the problem. What kills budgets:
- Egress fees — moving data out of a cloud provider costs 5–10× more than moving it in. This is by design. It's a switching cost disguised as a bandwidth charge.
- Storage that never gets cleaned up — old snapshots, abandoned S3 buckets, log retention set to "forever." Most teams have 20-40% dead storage.
- Over-provisioned instances — that m5.4xlarge "just in case" that averages 8% CPU utilisation.
- On-demand pricing for predictable workloads — if your database runs 24/7/365, paying on-demand rates is like renting a hotel room by the night for a year.
The cloud's marginal pricing is rational. The way most teams use it is not.
The cost comparison framework
Before comparing cloud to anything else, understand what you're actually paying for. Break your cloud bill into four buckets:
1. Compute (VMs, containers, serverless)
This is the line item everyone looks at. It's rarely the biggest problem.
- On-demand pricing: ~$0.05–0.10/hr for a general-purpose vCPU (varies by provider and region)
- Reserved instances / savings plans: 30–60% cheaper, but you commit for 1–3 years
- Spot instances: 60–90% cheaper, but your workload can be interrupted
Rule of thumb: if a workload runs more than 8 hours/day, it should be on reserved pricing. If it's fault-tolerant and can handle interruption, spot is free money.
2. Storage (block, object, archive)
Storage is cheap per-GB but expensive at scale because nobody deletes anything. Ever. Your S3 bucket from 2021? Still there. Still accruing charges.
- S3 / Blob standard: ~$0.023/GB/month (AWS us-east-1)
- Infrequent access tiers: ~$0.0125/GB/month
- Glacier / archive: ~$0.004/GB/month
- EBS (block storage): $0.08–0.10/GB/month — 4× more than object storage
The trap: teams store everything in standard tier and set retention to "forever." A lifecycle policy that moves data to infrequent access after 30 days and archive after 90 can cut storage costs by 50-70%.
3. Network (egress, load balancers, NAT gateways)
This is where cloud providers quietly make their real margin. Brace yourself.
- Egress: $0.09/GB on AWS, $0.08/GB on GCP (first 10TB/month). That's $90/TB.
- NAT gateway processing: $0.045/GB on AWS. This one catches everyone off guard.
- Cross-AZ data transfer: $0.01/GB each way. Adds up fast with microservices.
Comparison: a dedicated server with 1Gbps unmetered bandwidth costs ~$100-200/month. That's unlimited transfer. If you're moving more than 2 TB/month of egress, the math tilts fast.
4. Managed services (databases, queues, caches)
RDS, ElastiCache, managed Kafka, etc. These are convenient and expensive.
- RDS Multi-AZ Postgres (db.r6g.xlarge): ~$700/month
- Equivalent bare-metal server with Postgres + replication: ~$150-250/month
- Managed Kafka: can easily hit $1,000+/month for a modest cluster
The value proposition is operational: you're paying for automated backups, failover, patching. If your team is small and can't afford a DBA, managed services are worth it. If you have ops staff, the markup is harder to justify.
When to stay in the cloud
The cloud wins when:
- Your workload is spiky or unpredictable. Auto-scaling is actually valuable when demand varies 10× between peak and trough. Like, really valuable — this is the thing cloud was invented for.
- You're spending under $2,000/month total. The operational overhead of managing your own hardware isn't worth it at this scale.
- You need global distribution. Spinning up in 30 regions overnight is something only hyperscalers can do.
- Your team is small and you can't hire ops/infra people. Managed services trade money for headcount you don't have.
- Compliance requires specific certifications. AWS GovCloud, HIPAA BAAs, SOC 2 — sometimes you need the cloud's compliance stack.
When to leave (or go hybrid)
Consider bare metal or colocation when:
- Your bill is over $10K/month and workloads are predictable. You're paying a massive premium for elasticity you don't use.
- GPU/ML workloads dominate your spend. Cloud GPU pricing is brutal. An A100 on-demand runs $3-4/hr. A dedicated server with the same GPU is $2,000-3,000/month. If you're using it more than 25 hours/week, bare metal wins.
- Egress is a major cost centre. If you're serving media, running a CDN origin, or doing heavy API traffic, unmetered bandwidth saves thousands.
- You have (or can hire) ops capability. Bare metal requires hands-on management. If you can't do that, the cloud premium is the cost of not having to.
Spot instances: the most underused cost lever
Spot instances (AWS) / preemptible VMs (GCP) / spot VMs (Azure) offer 60–90% discounts for interruptible capacity.
Good fit for:
- CI/CD pipelines
- Batch processing and data pipelines
- Dev/staging environments
- Stateless web workers behind a load balancer
- ML training with checkpointing
Bad fit for:
- Databases
- Single-instance production services
- Anything that can't handle a 2-minute shutdown notice
Practical tip: use a mix of spot instance types and availability zones. If one type gets reclaimed, others usually don't at the same time. AWS Spot Fleet and Karpenter (for Kubernetes) handle this well.
The hybrid playbook (for $10K–$100K/month spend)
The sweet spot for many mid-size teams:
- Baseline compute on bare metal — predictable workloads (databases, app servers, caches) on dedicated hardware with monthly contracts
- Burst compute in cloud — auto-scale into cloud instances for traffic spikes, batch jobs, and CI/CD
- Object storage in cloud — S3/GCS is hard to beat for durability and tooling integration
- CDN for egress — Cloudflare, Bunny, or Fastly for serving static content (way cheaper than cloud egress)
This gives you the cost floor of bare metal with the elasticity ceiling of cloud when you need it.
Cost monitoring: what to set up on day one
Don't wait until the bill surprises you.
- Billing alerts at 50%, 80%, and 100% of expected spend
- Per-service cost allocation tags — if you can't attribute cost to a team or service, you can't optimise it
- Weekly cost anomaly review — a 5-minute scan of the cost dashboard catches runaway resources
- Unused resource detection — unattached EBS volumes, idle load balancers, stopped instances still incurring storage charges
- Right-sizing recommendations — all three major clouds offer these natively; actually look at them
Common mistakes
- "We'll optimise later" — no you won't. Nobody ever does. Set up cost guardrails before you need them.
- Paying on-demand for everything — reserved instances and savings plans exist. Use them for stable workloads.
- Ignoring NAT gateway costs — on AWS, this quietly becomes a top-5 line item. Use VPC endpoints where possible.
- Storing logs forever in hot storage — move to infrequent access after 7 days, archive after 30, delete after 90 unless compliance says otherwise.
- Not accounting for engineer time — bare metal is cheaper in compute but more expensive in human hours. Factor that in honestly.
Decision checklist (fast)
- What's your total monthly cloud spend? Under $2K → stay cloud-first. Over $10K → audit seriously.
- What percentage of your workload is predictable (runs 24/7)? High % → reserved instances or bare metal.
- How much egress are you paying? Over $500/month → investigate CDN or dedicated bandwidth.
- Are you using spot instances for CI/CD and batch work? If no, start there — it's the easiest win.
- Do you have ops staff who can manage bare metal? If no, cloud premium is your ops cost. That's fine.
- What's your storage retention policy? If "we don't have one" → you have dead storage. Lifecycle it.
- Can you tag costs per-service? If no, fix that before optimising anything else.
Sources
- AWS — EC2 pricing pages and Spot Instance pricing history (on-demand, reserved, and spot rates by instance type and region)
- Cloudflare — bandwidth pricing comparison (cloud egress fees vs. flat-rate CDN pricing across AWS, GCP, Azure)
- Hetzner — dedicated server pricing (representative bare-metal costs for compute-equivalent comparisons)
- The FinOps Foundation — cloud cost management best practices and frameworks: https://www.finops.org/