What Really Caused the AI Boom

Fetch.ai
Fetch.ai
Published in
3 min readOct 31, 2023

The path to artificial intelligence is one for the history books. A labyrinth of innovations has collectively breathed life into the AI we know today.

While the intricate dance between hardware, algorithms, and data has been ongoing for years, the rhythm has recently intensified, propelling AI into the limelight. But what were the factors that led to it?

The Era of Gigantic Neural Networks

Let’s rewind a bit. Neural networks before 2012 were like quaint little villages — small and manageable, sporting just a few thousand neurons. The post-2012 era, however, saw them grow into bustling metropolises: large, deep, and complex. This growth in size and depth wasn’t merely for show. Larger models demonstrated improved performance across a multitude of tasks, hinting at the potential of scaling towards Artificial General Intelligence (AGI) and possibly — even Superintelligent AI.

GPUs: Gaming Chips Turn AI Powerhouses

Interestingly, the video game industry can claim credit for the growth of AI. Graphic Processing Units (GPUs), initially developed to satisfy gamers’ demands for better graphics, soon became AI researchers’ workhorses.

While the architectural skeleton of GPUs remained relatively consistent, the scale, speed, and efficiency ballooned. Modern GPUs, like the V100, can churn out computations at a staggering rate, rivaling the supercomputers from just a decade ago.

When companies unleash an army of these GPUs on voluminous datasets, the computational prowess is akin to having a fleet of supercomputers from the early 2000s.

Data and Architecture: The Unsung Duo

Behind every successful AI model lies an often-overlooked hero: data. Data is a fundamental piece of the AI puzzle. As the internet has grown, so has the availability of vast datasets.

A notable example is ImageNet, which provided millions of labeled images. With more data at their disposal, networks like AlexNet managed to significantly improve accuracy rates. This success spotlighted the potential of techniques such as stochastic gradient descent and set off a rapid evolution in neural network architectures.

Architectural innovations have also played a pivotal role in the growth of AI. Long Short-Term Memory Networks (LSTMs) dramatically improved speech-to-text applications, bringing error rates from a frustratingly high 25% to below 10%. The user experience of dictating a text message or asking a virtual assistant a question transformed from an exercise in patience to a seamless interaction. Following LSTMs, Transformer models changed the game. Born from the principle that ‘Attention is all you need,’ Transformers displayed an impressive ability to retain and process large chunks of text.

The Convergence of Multiple Forces

It is safe to conclude that the current AI surge isn’t attributed to just one factor. It’s a blend of various elements coming together. The explosion of high-quality data, powered by the spread of the internet and smartphones, laid the groundwork. Then, the dramatic improvement in computing power, combined with cost reductions, added momentum.

Technologies like CUDA and CuDNN were also a catalyst — as they enabled efficient deep learning on hardware, facilitating the training of deeper models. Innovations like Rectified Linear Units (ReLU) and optimizers like Adam made it feasible and cost-effective to train these mammoth models. Even as the nuts and bolts of AI, such as word embeddings and attention mechanisms, underwent refinements, the overarching vision remained consistent — creating machines that understand and respond with unparalleled accuracy.

The AI surge of the modern era is a result of a collection of technological advancements, each contributing its part. As we look to the future and continue pushing the limits, we’re left to wonder: What’s the next big leap in AI?

--

--

Fetch.ai
Fetch.ai

Build, deploy and monetize AI apps and services.