DEV Community

Shrijith Venkatramana
Shrijith Venkatramana

Posted on • Edited on

The Economics of Training Frontier Models

Hello, I'm Shrijith. I'm building git-lrc, an AI code reviewer that runs on every commit. It is free, unlimited, and source-available on Github. Star Us to help devs discover the project. Do give it a try and share your feedback for improving the product.

Presently, I am building LiveAPI, a super-convenient engineering productivity tool. LiveAPI processes your code repositories at scale and automatically produces beautiful API docs in minutes.

As I build LiveAPI, I am also making an effort to learn about various economic matters & share it here with you.

--- was released to the world on 30 November 2022.

I am reading the following report, published on 31 May 2024, almost 1.5 years after the original "ChatGPT moment".

The rising costs of training frontier AI models

As they say, money makes the world go round, so let's try to learn some data and insights about the costs of developing serious (or frontier) models.

Components of Training Cost Models

  1. Hardware
  2. Energy
  3. Cloud rental
  4. Staff Expenses

Rough Estimates for GPT-4 and Gemini

  1. AI Accelerator Chips (37% - 29.5%)
  2. Staff Costs (37% - 29.5%)
  3. Server Components (15-22%)
  4. Cluster-Level Interconnect (9-13%)
  5. Energy Consumption (2-6%)

Since 2016, the absolute cost of training frontier models has increased 2.4x every year. Assuming such a trend continues, larger model trainings will cost more than 1 billion by 2027.

The Data: GPT-4 Training cost $40M, Gemini Ultra Cost $30M

Hardware + Energy Cost Evolution

Hardware + Energy Cost Evolution

Cloud Compute Cost Evolution

Cloud Compute Cost Evolution

Hardware Acquisition Cost

![Har### Energy/Hardware Costs Breakdown for All The Major Models

Conclusions From the Study

  1. Half of ammortized hardware capex + energy cost is for AI chips
  2. The third biggest cost apart from the above two is for employing R&D staff
  3. Training costs are increasing exponentially, year by year
  4. Securing chips and power are going to be bottlenecks in the future for AI development

git-lrc
*AI agents write code fast. They also silently remove logic, change behavior, and introduce bugs -- without telling you. You often find out in production.

git-lrc fixes this. It hooks into git commit and reviews every diff before it lands. 60-second setup. Completely free.*

Any feedback or contributors are welcome! It's online, source-available, and ready for anyone to use.

⭐ Star it on GitHub:

GitHub logo HexmosTech / git-lrc

Free, Unlimited AI Code Reviews That Run on Commit

git-lrc logo

git-lrc

Free, Unlimited AI Code Reviews That Run on Commit



git-lrc - Free, unlimited AI code reviews that run on commit | Product Hunt



AI agents write code fast. They also silently remove logic, change behavior, and introduce bugs -- without telling you. You often find out in production.

git-lrc fixes this. It hooks into git commit and reviews every diff before it lands. 60-second setup. Completely free.

See It In Action

See git-lrc catch serious security issues such as leaked credentials, expensive cloud operations, and sensitive material in log statements

git-lrc-intro-60s.mp4

Why

  • 🤖 AI agents silently break things. Code removed. Logic changed. Edge cases gone. You won't notice until production.
  • 🔍 Catch it before it ships. AI-powered inline comments show you exactly what changed and what looks wrong.
  • 🔁 Build a habit, ship better code. Regular review → fewer bugs → more robust code → better results in your team.
  • 🔗 Why git? Git is universal. Every editor, every IDE, every AI…

Top comments (0)