DEV Community

VoltageGPU
VoltageGPU

Posted on

Per-Second vs Hourly GPU Billing: I Saved 40% — Here Is the Math

Quick Answer: AWS charges $3.43/hr for A100 GPUs with hourly billing. VoltageGPU’s per-second billing cuts costs to $2.02/hr — 41% cheaper. For a 30-minute task, you pay $1.01 instead of $3.43.

TL;DR: I tested 100 AI inference jobs across VoltageGPU and AWS. Average runtime: 22 minutes. With per-second billing, I saved $40/hour. Limitation: Minimum 60-second billing period (if your task takes 59 seconds, you still pay for a minute).


Why Per-Second Billing Matters for AI Workloads

Cloud providers charge for GPUs in two ways:

  1. Hourly billing: You pay for full hours, even if you use the GPU for 10 minutes.
  2. Per-second billing: You pay for exact usage time, rounded up to the nearest second (with a 60-second minimum).

In 2023, AI training and inference jobs often take minutes, not hours. Hourly billing wastes money. I spent 3 weeks benchmarking VoltageGPU’s per-second model vs AWS and RunPod’s hourly pricing.


Real-World Pricing Data

VoltageGPU’s per-second billing costs:

  • H100 80 GB: $2.77/hr (per-second) — live data
  • A100 80 GB: $2.02/hr (per-second) — live data

Competitors’ hourly billing:

Key insight: VoltageGPU’s A100 is 41% cheaper than AWS. RunPod’s A100 is 21% cheaper than VoltageGPU — but only if you need hourly billing.


The Math: 30-Minute Task Example

Provider Billing Model Cost for 30 Minutes Savings vs Hourly
VoltageGPU Per-second $1.01 (2.77/2) $2.42
AWS Hourly $3.43 (full hour)
RunPod Hourly $1.64 (full hour)

Note: VoltageGPU’s per-second billing charges $2.77/60 = $0.046 per second. For 30 minutes (1,800 seconds): 1,800 × $0.046 = $1.01.


How I Saved 40% on 100 AI Jobs

I ran 100 inference tasks (average 22 minutes) on VoltageGPU and AWS:

  • VoltageGPU total cost: 100 × $1.05 = $105
  • AWS total cost: 100 × $3.43 = $343
  • Savings: $343 - $105 = $238 (69% off)

Limitation: Per-second billing has a 60-second minimum. If your task takes 59 seconds, you pay for 1 minute. Hourly billing avoids this, but costs 3–4x more for partial hours.


Code: Run Your Own Test

from openai import OpenAI
client = OpenAI(
    base_url="https://api.voltagegpu.com/v1/confidential",
    api_key="vgpu_YOUR_KEY"
)
response = client.chat.completions.create(
    model="contract-analyst",
    messages=[{"role": "user", "content": "Analyze this document..."}]
)
print(response.choices[0].message.content)
Enter fullscreen mode Exit fullscreen mode

Use this with VoltageGPU’s per-second pricing to avoid hourly traps. Live demo available.


Honest Comparison: When Hourly Wins

  • Long-running tasks: If your job takes 50+ minutes, hourly billing on RunPod ($1.64/hr) is cheaper than VoltageGPU’s per-second ($2.02/hr).
  • No cold start: Hourly billing avoids VoltageGPU’s 30–60s cold start on the Starter plan.

Final Take: Per-Second is Better for Short Jobs

Use Case Best Option Cost per Hour
<30-minute inference VoltageGPU $2.02
1-hour training RunPod $1.64
5-minute batch jobs VoltageGPU $0.23

Don’t trust me. Test it. 5 free agent requests/day -> voltagegpu.com

For law firms: GDPR-compliant AI

For developers: Confidential API guide

For benchmarks: VoltageGPU vs AWS

Top comments (0)