Why is Apple making a data center chip?

Jay Goldberg

Posts: 98   +2
Staff
Editor's take: Of course, Apple is designing its own AI chips for its data centers. This will give them an end-to-end, cloud-to-edge AI system that its competitors will not be able to match anytime soon.

A report from Bloomberg's Mark Gurman last week indicated that Apple was going to put its own chips in the data center, and a few days later, the Journal confirmed this with its own reporting. The information was initially met with a high degree of skepticism, particularly from the Journal, which later changed its stance on the subject.

Many said that Apple was "not a cloud company" and did not see the point in them pursuing this path. This struck us as odd since we know Apple has been looking at this for years. Here, we want to lay out some of the reasons we think Apple would want its own data center chip.

Editor's Note:
Guest author Jonathan Goldberg is the founder of D2D Advisory, a multi-functional consulting firm. Jonathan has developed growth strategies and alliances for companies in the mobile, networking, gaming, and software industries.

For starters, Apple is building a chip because they can. In a world of constrained leading-edge capacity, Apple is TSMC's largest customer and can presumably get most of the capacity they want. If they actually had a good use for the chip, now is the time. Everyone else is scrambling, and Apple could conceivably build up a big head start.

It should also be clear that Apple is very much a provider of cloud services – they literally have a product called iCloud. Part of the confusion here rests in the fact that Apple says very little about what it does with its data centers. They never make it entirely clear how they deliver their cloud services, which seems to be some combination of Google Cloud, Azure, and maybe some AWS, but also their own data centers. When we talk about the "hyperscalers," we silently assume that Apple is one of those ten largest data center operators. And if all the other hyperscalers are designing their own chips, it makes sense that Apple would as well.

The big question is what Apple is going to do with that capacity. This is 2024, so the simple summary is that, of course, they are going to do AI.

The more nuanced answer is where this gets more interesting. We have long bemoaned the lack of compelling consumer uses for generative AI, and we often add that the industry needs Apple to show everyone else what can be done. We are not convinced that Apple has, in fact, figured it out, but given all the rumors that next month's WWDC is going to heavily focus on AI, we are intrigued.

So let's assume Apple has come up with some impressive AI applications, or at least applications Apple thinks are impressive. We know that they have been working on AI for years – from Siri's steadily improving speech recognition to under-the-hood machine learning improvements to photo and video images, to the simple fact that the A series phone processor has had a "Neural Engine" since 2018.

They brought transformer support to their M-Series chips before ChatGPT. They launched that support so soon after those models became widely popular that it seems likely they had been working on the idea for some time already. Apple has a capable AI team.

Moreover, it seems likely that team has built its own foundational model. Apple, of course, owns all of its critical software and a huge amount of data (I.e., they know what you do with your phone). We are not clear how Apple trained this model, but we suspect they used Nvidia like everyone else.

However, as Apple rolls out its AI apps, they are going to need to do a lot of inference. They will run some of that on-device (a.k.a. "the edge," a.k.a. consumers pay for the capex and opex), but they will also run some amount of inference in the cloud.

From all the reporting, it is not entirely clear what chips Apple will be using in the data center. We have seen reports they will use their M2 CPU, their M4 CPU, or a brand-new chip called ACDC (Apple Chips in Data Center, which would be the clear winner in 2024's chip naming competition). Whatever they end up using, they will benefit from having a common software layer connecting the cloud and edge inference.

Many companies have talked about having this kind of offering, but Apple is likely to be the first to put one in large-scale production. Having everything connected in this way offers meaningful power savings but also probably allows for a higher degree of privacy, a big marketing theme for the company. We imagine the pitch will go something like, "Have your phone run amazing AI, and only you will see your data." So there are clear benefits to having their own silicon along the entire software chain.

That leaves one question: how will Apple position this in their marketing?

As we mentioned above, Apple never talks about their data centers, but now they might be willing to, as it conveys a message that benefits them. In fact, given the amount of press coverage coming out on the subject, we suspect that Apple may have deliberately leaked a bit of this to make sure we are paying attention.

When we boil it all down, it looks a lot like Apple has an interesting story to tell about the breadth of its AI offering. This is not just an AI Mac or an AI iPhone but a complete system offering some form of differentiated AI. If this is true, and we may be wrong, it will be one more example of Apple stealing a beat on the competition, because this is something that its competitors will struggle to match.

Permalink to story:

 
Not to be too reductive but I think it's simple:

1) All AI models as ambitious as the one Apple is planning basically need to steal data to train.

2) In order to not get caught for the longest possible time with your hands on the stolen data you need to own as much of the entire process chain as possible. This means owning the entire data center if you can help it

3) Data centers are not cheap propositions: it's why Microsoft makes most of it's money today from Azure and AWS is about as big if not bigger with Google not being far behind. Only way even the richest company in the world can justify such a massive investment is to cut down the cost anywhere they can so their own datacenter chips become kind of necessary.

Of course you can argue that even without point 1 being necessarily true at all (But is not like Apple will ever show their models to refute it) Point 3 it's still kind of maintained: Azure and AWS are not really designed to launch a massive operation like running an AI model for potentially millions of simultaneous users. Well to clarify, they are designed to scale up that much but their price is definitively not designed to not put you our of business at the rate Apple would need to expand.

Basically if you require so much compute that a cloud provider would probably have to create or rent entire new data centers just for you, it will be far too expensive to just use someone else as a provider you have to provide your own "cloud" at that point.
 
Apple is absolutely not the first to put their own custom chips into large scale production. One of the pain points Intel has had of late is having this drop off of Xeon sales -- because Google, Facebook, Amazon, several of the Chinese cloud providers, they all are making their own chips. Better performance per watt; they don't have to pay a penny (and have wasted die space and possibly power use) for features they know they won't use (and can put in features even if unusual that they think they may make use of); and of course no markup (high initial costs for chip design; but then they're just paying TSMC directly for production, versus buying AMD or Intel CPUs where I'm sure they get a price break but Intel and AMD are still expecting some profit margin on the chips.)


Anyway... Apple definitely runs their own cloud setup, and it indeed makes sense for them to use their own CPUs in it. I'm surprised they weren't already.
 
Back