Aiconomy

Pre-Training

The initial phase of training an AI model on a large, general-purpose dataset to learn broad knowledge and patterns before it is fine-tuned for specific tasks.

Pre-training is the most expensive phase of building large AI models. GPT-4's pre-training reportedly cost $78-191 million in compute alone, processing trillions of tokens from books, websites, and code. The pre-training paradigm, popularized by BERT (2018) and GPT-2 (2019), enables transfer learning — training once, then adapting cheaply to many tasks. Pre-training datasets have grown from millions of documents to trillions of tokens. The quality and composition of pre-training data is now recognized as equally important as model architecture and scale.

Live Data

10,023,606,624GPU Hours Consumed by AI Today

AI Economy Pulse

Every Friday: the 3 AI data points that actually matter this week. Free, forever.

Built on data from Stanford HAI, IEA, OECD & IMF

Latest: “AI Investment Hits $42B in Q1 2026 — Here's Where It Went”

No spam, ever. Unsubscribe anytime.