11 points

It’s a no from me. I suspect as the US gets more deregulated for AI, it will be more no’s from people around the world.

permalink
report
reply
12 points

And it’s hard to tell what the difference is. Apples ‘built from the ground up for AI’ chips just have more RAM. What’s the difference with CPUs? Do they just have more onboard graphics processing that can also be used for matrix multiplication?

permalink
report
reply
3 points

Basically yes. They come with an NPU (Neural processing unit) which is hardware acceleration for matrix multiplications. It cannot do graphics. Slap whatever NPU into the chip, boom: AI laptop!

permalink
report
parent
reply
2 points

Matrix multiplication is also largely what graphics cards do, I wonder how the npus are different.

permalink
report
parent
reply
1 point

Modern graphics cards pack a lot of functionality. Shading units, Ray tracing, video encoding/deciding. NPU is just the part needed to accelerat Neural nets.

permalink
report
parent
reply
7 points

The stupid difference is supposed to be that they have some tensor math accelerators like the ones that have been on GPUs for three generations now. Except they’re small and slow and can barely run anything locally, so if you care about “AI” you’re probably using a dedicated GPU instead of a “NPU”.

And because local AI features have been largely useless, so far there is no software that will, say, take advantage of NPU processing for stuff like image upscaling while using the GPU tensor calculations for in-game raytracing or whatever. You’re not even offloading any workload to the NPU when you’re using your GPU, regardless of what you’re using it for.

For Apple stuff where it’s all integrated it’s probably closer to what you describe, just using the integrated GPU acceleration. I think there are some specific optimizations for the kind of tensor math used in AI as opposed to graphics, but it’s mostly the same thing.

permalink
report
parent
reply
-3 points

Seems silly to try to get the CPU to do GPU stuff, just upgrade the GPU.

permalink
report
parent
reply
4 points

The idea is having tensor acceleration built into SoCs for portable devices so they can run models locally on laptops, tablets and phones.

Because, you know, server-side ML model calculations are expensive, so offloading compute to the client makes them cheaper.

But this gen can’t really run anything useful locally so far, as far as I can tell. Most of the demos during the ramp-up to these were thoroughly underwhelming and nowhere near what you get from server-side services.

Of course they could have just called the “NPU” a new GPU feature and make it work closer to how this is run on dedicated GPUs, but I suppose somebody thought that branding this as a separate device was more marketable.

permalink
report
parent
reply
2 points

The Apple chips also have a wide interface to the RAM. That means you can run chatbots (LLMs) and other AI workloads that are memory-bound at crazy speeds compared to an Intel (or AMD) computer.

permalink
report
parent
reply
3 points

Really? How fast is the memory bus compared to x86? And did they just double the bus bandwidth by doubling the memory?

I’m dubious because they only now went to 16gb ram as base, which has been standard on x86 for almost a decade.

permalink
report
parent
reply
2 points
*

Depending on the chip, they have somewhere from 100 to 400 GB/s. I’m not sure on the numbers on Intel processors. I think the consumer processors have about 50 - 80 GB/s. (~Alder Lake, dual channel DDR5) Mine seems to have way less. And a recent GPU will be somewhere in the range of 400 to 1000 GB/s. But consumer graphics cards stop at 24GB of VRAM and these flagship models are super expensive. Even compared to Apple products.

The people from the llama.cpp project did some measurements and I believe the Apple “Metal” framework seems to outperform the x86 computers by an order of magnitude or so. I’m not sure, it’s been some time since i skimmed the discussions on their Github page.

permalink
report
parent
reply
1 point
*

Apple is also much faster because the integrated graphics are actually usable for LLMs.

The base M is just a big faster than an Intel/AMD laptop if you can get their graphics working. The M Pro is 2x is fast (as its memory bus is 2x as wide). The M Max is 4x as fast.

AMD is coming out with something more competitive in 2025 though, Strix Halo.

permalink
report
parent
reply
7 points
*

”However, if it is performance you are concerned about, “it’s important to note that GPUs still far outperform NPUs in terms of raw performance,” Jessop said, while NPUs are more power-efficient and better suited for running perpetually.”

Ok, so if you want to run your local LLM on your desktop, use your GPU. If you’re doing that on a laptop in a cafe, get a laptop with an NPU. If you don’t care about either, you don’t need to think about these AI PCs.

permalink
report
reply
2 points

Or use a laptop with a GPU? An npu seems to just be slightly upgraded onboard graphics.

permalink
report
parent
reply
1 point

It’s a power efficiency thing. According to the article, a GPU gets the job done, but uses more energy to get there. Probably not a big deal unless charging opportunities are scarce.

permalink
report
parent
reply
34 points

Just saying “built on AI” or whatever isn’t a convincing sales pitch. What can I actually do with AI that will improve my day to day life? Not a single advert or pitch has told me a single use case for this that applies to what anyone would use for a personal computer, and they’re too risky to buy for employees in a work environment unless you can afford to be the guinea pig for this unproven line of hardware (in the sense that I know a ThinkPad will last 10 years but I have no idea how long a copilot pc will last, how often I need to replace the battery or ram or anything else). I’m aware of tech, I know what these laptops are, but as far as I can see the market for them just does not exist and I don’t understand why anyone would think otherwise.

permalink
report
reply
8 points

Have they tried forcing people to upgrade to AI PCs in order to receive security updates by checking to see if your PC is an AI PC? You know just to prove people really want AI PCs?

/s

permalink
report
reply

Community stats

  • 1.9K

    Monthly active users

  • 723

    Posts

  • 3.2K

    Comments