AIMultiple ResearchAIMultiple Research

Top 10 AI Chip Makers of 2024: In-depth Guide

Top 10 AI Chip Makers of 2024: In-depth GuideTop 10 AI Chip Makers of 2024: In-depth Guide

As the figure above illustrates, the number of parameters (consequently the width and depth) of the neural networks and therefore the model size is increasing. To build better deep learning models and power generative AI applications, organizations require increased computing power and memory bandwidth. 

Powerful general-purpose chips (such as CPUs) cannot support highly parallelized deep learning models. Therefore, AI chips that enable parallel computing capabilities are increasingly in demand, and according to McKinsey, this trend will continue.

However, even Intel, which has numerous world-class engineers and a strong research background, needed three years of work to develop its own AI chip. Therefore, for most companies, buying chips from these vendors or renting capacity from cloud GPU providers is the only way to develop powerful deep learning models. This article will introduce top AI chip vendors to help companies choose the right one.

Which are the leading AI chip producers?

1. Nvidia

Source: MarketWatch

Nvidia has been producing graphics processing units (GPUs) for the gaming sector since 1990s. The PlayStation 3 and Xbox both use Nvidia graphics arrays. The company also makes AI chips such as Volta, Xavier, and Tesla. Thanks to the generative AI boom, NVIDIA had excellent results in Q2 2023, reached a trillion in valuation and solidified its status as the leader of GPU and AI hardware markets.

NVIDIA’s chipsets are designed to solve business problems in various industries. Xavier, for example, is the basis for an autonomous driving solution, while Volta is aimed at data centers. DGX™ A100 and H100 are the flagship AI chips of Nvidia which are designed for AI training and inference in data centers. A100 integrates 8 GPUs and up to 640GB GPU memory. Nvidia Grace is the new AI chip model that the company released for the HPC market in 2023.

For AI workloads on the cloud, Nvidia almost has a monopoly with most cloud players offering only Nvidia GPUs as cloud GPUs. Nvidia also launched its DGX Cloud offering providing cloud GPU infrastructure directly to enterprises.

2. Advanced Micro Devices (AMD)

AMD is a chip manufacturer that has CPU, GPU and AI accelerator products. For instance, AMD’s Alveo U50 data center accelerator card has 50 billion transistors. Accelerator can run 10 million embedding datasets and perform graph algorithms in milliseconds.

AMD launched MI300 for AI training workloads in June 2023 and will be competing with NVIDIA for marketshare in that market.1 There are startups, research institutes, enterprises and tech giants that adopted AMD hardware in 2023 since Nvidia AI hardware has been difficult to procure due to rapidly increasing demand with the rise of generative AI as demonstrated by ChatGPT.2345

AMD is also working with machine learning companies like Hugging Face to enable data scientists to use their hardware more efficiently.6

The software ecosystem is critical as hardware performance relies heavily on software optimization. For example, AMD and NVIDIA had a public disagreement over benchmarking H100 and MI300. The focus of the disagreement was the package and floating point to use in the benchmark. According to the latest benchmarks, it appears that MI300 is better or on par with H100 for inferencing on a 70B LLM.78

3. Intel

Intel is the largest player in the CPU market and has a long history of semiconductor development. In 2017, Intel became the first AI chip company in the world to break the $1 billion sales barrier. 

Intel’s Xeon CPUs are appropriate for a variety of jobs, including processing in data centers and have had an impact on the company’s commercial success.

Gaudi3 is the latest AI accelerator processor from Intel. Since it will be publicly released in 2024, there are currently limited benchmarks on its performance.9

Which public cloud providers produce AI chips?

4. Alphabet / Google Cloud Platform

Google Cloud TPU is the purpose-built machine learning accelerator chip that powers Google products like Translate, Photos, Search, Assistant, and Gmail. It can be used via the Google Cloud as well. Google announced TPUs in 2016.10

Edge TPU, another accelerator chip from Google Alphabet, is smaller than a one-cent coin and is designed for edge devices such as smartphones, tablets, and IoT devices.

5. AWS

AWS produces Tranium chips for model training and Inferentia chips for inference. Though AWS is the market leader in public cloud, it started building its own chips after Google.

6. IBM

IBM announced its latest deep learning chip, artificial intelligence unit (AIU), in 2022.11. IBM is considering using these chips to power its watson.x generative AI platform.12

AIU builds on “IBM Telum Processor” which powers AI processing capabilities of IBM Z mainframe servers. At launch, Telum processors’ highlighted use cases included fraud detection.13

IBM also demonstrated that merging compute and memory can lead to efficiencies. These were demonstrated in the NorthPole processor prototype.14

7. Alibaba

Alibaba produces chips like Hanguang 800 for inference. However, some North American, European and Australian organizations (e.g. those in the defense industry) may not prefer to use Alibaba Cloud for geopolitical reasons.

Who are the leading AI chip startups?

We would also like to introduce some startups in the AI chip industry whose names we may hear more often in the near future. Even though these companies were founded only recently, they have already raised millions of dollars.

Image shows the fundings of leading AI chip startups. According to recent data SambaNova has funding over 1 billion USD. It is followed by Cerebras Systems and Graphcore.
Figure 2: Total funding for AI chip makers, Source: Statista 15 and Reuters 16

8. SambaNova Systems

SambaNova Systems was founded in 2017 with the goal of developing high-performance, high-precision hardware-software systems for high volume generative AI workloads. The company has developed the SN40L chip and raised more than $1.1 billion in funding.1718

It is important to note that SambaNova Systems also leases its platform to businesses. AI platform as service approach of SambaNova Systems makes their systems easier to adopt and encourages hardware reuse for circular economy

9. Cerebras Systems

Cerebras Systems was founded in 2015. In April 2021, the company announced its new AI chip model, Cerebras WSE-2, which has 850,000 cores and 2.6 trillion transistors. Undoubtedly, the WSE-2 is a big improvement over the WSE-1, which has 1.2 trillion transistors and 400,000 processing cores. 

Celebra’s system works with many pharmaceutical companies such as AstraZeneca and GlaxoSmithKline because the effective technology of WSE-1 accelerates genetic and genomic research and shortens the time for drug discovery.

10. Groq

Groq has been founded by former Google employees. The company represents LPUs, a new model for AI chip architecture, that aims to make it easier for companies to adopt their systems. The startup has already raised around $350 million and produced its first models such as GroqChip™ Processor, GroqCard™ Accelerator, etc.

The company is focused on LLM inference and released benchmarks for Llama-2 70B.19

In Q1 2024, the company shared that 70k developer signed up on its cloud platform and built 19k new applications.20

On the March 1, 2022, Groq acquired Maxeler, which has high performance computing (HPC) solutions for financial services.

What are upcoming AI hardware producers?

Though these are compelling AI hardware solutions, there are currently limited benchmarks on their effectiveness since they are newcomers to the market.

Meta

Meta Training and Inference Accelerator (MTIA) is a family of processors for AI workloads such as training Meta’s LLaMa models.

The latest model is Next Gen MTIA which is based on TSMC 5nm technology and is claimed to have 3x improved performance vs. MTIA v1. MTIA will be hosted in racks containing up to 72 accelerators.21

MTIA is currently for Meta’s internal usage. However, in the future, if Meta launched a LLaMa based enterprise generative AI offering, these chips could power such an offering.

Microsoft Azure

Microsoft launched Maia AI Accelerator in November 202322

Rebellions

Korea based startup raised $124M in 2024 and is focused on LLM inference.23

What are other AI chip producers?

Graphcore

Graphcore is a British company founded in 2016. The company announced its flagship AI chip as IPU-POD256. Graphcore has already been funded with around $700 million.

Company has strategic partnerships between data storage corporations like DDN, Pure Storage and Vast Data. Graphcore works with research institutes around the globe like Oxford-Man Institute of Quantitative Finance, University of Bristol and Berkeley University of California are other reputable research organizations that use Graphcore’s AI chips.

As of October 2023, the company’s long term viability is at risk as it is losing ~$200M per year and ~$160M in assets as of Jan 1st, 2023.2425

Mythic

Mythic was founded in 2012 and is focused on edge AI. Mythic follows an unconventional path, an analog compute architecture, that aims to deliver power-efficient edge AI computing.

It developed products such as M1076 AMP, MM1076 key card, etc., and has already raised about $165 million in funding.26

Mythic laid of most of its staff and restructured its business with its funding round in March 2023.27

What are companies reported to be working on AI hardware?

These companies are yet to launch their AI hardware.

_etched

The team claims to have built the world’s first transformer supercomputer however AIMultiple team has not yet come across any benchmarks or client references.

OpenAI

OpenAI is reported to be raising funds to build its own AI hardware. 28

You can also check our sortable list of companies working on AI chips.

You might enjoy reading our articles on TinyML and accelerated computing.

If you have questions about how AI hardware can help your business, we can help:

Find the Right Vendors

References

Access Cem's 2 decades of B2B tech experience as a tech consultant, enterprise leader, startup entrepreneur & industry analyst. Leverage insights informing top Fortune 500 every month.
Cem Dilmegani
Principal Analyst
Follow on

Cem Dilmegani
Principal Analyst

Cem has been the principal analyst at AIMultiple since 2017. AIMultiple informs hundreds of thousands of businesses (as per similarWeb) including 60% of Fortune 500 every month.

Cem's work has been cited by leading global publications including Business Insider, Forbes, Washington Post, global firms like Deloitte, HPE, NGOs like World Economic Forum and supranational organizations like European Commission. You can see more reputable companies and media that referenced AIMultiple.

Throughout his career, Cem served as a tech consultant, tech buyer and tech entrepreneur. He advised businesses on their enterprise software, automation, cloud, AI / ML and other technology related decisions at McKinsey & Company and Altman Solon for more than a decade. He also published a McKinsey report on digitalization.

He led technology strategy and procurement of a telco while reporting to the CEO. He has also led commercial growth of deep tech company Hypatos that reached a 7 digit annual recurring revenue and a 9 digit valuation from 0 within 2 years. Cem's work in Hypatos was covered by leading technology publications like TechCrunch and Business Insider.

Cem regularly speaks at international technology conferences. He graduated from Bogazici University as a computer engineer and holds an MBA from Columbia Business School.

To stay up-to-date on B2B tech & accelerate your enterprise:

Follow on

Next to Read

Comments

Your email address will not be published. All fields are required.

2 Comments
Dave
Aug 29, 2022 at 05:49

You forgot to include Tesla with their DOJO supercomputer. From the ground-up, the supercomputer was specifically designed for machine learning and image recognition – which means that every component was designed for it including, but not limited to, PCI board design, CPU, RAM, cooling, power, scalable hardware design and software. If I’m not mistaken, the AI is also the second most widely tested and used in the “wild”, just below that of Google due to Google using it in their Search.

Bardia Eshghi
Sep 06, 2022 at 13:52

Thank you for your feedback, Dave!

thayyil
Mar 19, 2022 at 11:48

surprised that brainchip (akida) missing in this report. any reasons?

Cem Dilmegani
Nov 18, 2022 at 07:36

All included companies here raised $100+M. Last time we collected the data, that wasn’t the case for akida. Why don’t you reach out to us at info@aimultiple.com and let’s discuss why it should be included.

Thank you!