AMD will bring its “Ryzen AI” processors to standard desktop PCs for first time (arstechnica.com)

223 points by Bender 3 days ago

bilekas 9 hours ago

> This makes them AMD’s first desktop chips to qualify for Microsoft’s Copilot+ PC label, which enables a handful of unique Windows 11 features like Recall and Click to Do.

This is not the selling point they think it is.

The problem I see with the AM5 socket is simply the fact that DD5 RAM to support it is just too expensive. So this will not really make the big impact they were hopin for.

davidmurdoch 8 hours ago

I have a hard time believing ANYONE thinks this is a selling point. Literally anyone, including marketing and execs at Microsoft. I think they have just sunk too much money in it to quit, so they keep doubling down.

giantg2 7 hours ago

"The problem I see with the AM5 socket is simply the fact that DD5 RAM to support it is just too expensive."

DDR4 is basically just as expensive. At least DDR5 gives on-chip error correction (not as good as full ECC).

For a geneal computer, there's not that much difference between AM4 and AM5 unless you really want the extra speed of DDR5, PCI 5, and the newest processors. You can build a very capable AM4 machine for slightly less money, but that savings is found on the CPU and motherboard, not on the RAM.

sitkack 15 minutes ago

The ECC in DDR5 is there just to make it work because of the errors caused by the density and the data rates. The ECC isn’t there for you, it is for the manufacturers.

HerbManic 2 hours ago

I am still running DDR2 & DDR3 machines! I was going to finally make the big upgrade this year but am now holding off until the market finds a little bit more sanity.

deltoidmaximus 6 hours ago

DDR4 looks to be around half the price of DDR5 on the used market to me. I wouldn't call that slightly less money unless you weren't planning to install much RAM.

giantg2 5 hours ago

tuckerman 4 hours ago

I think for brand new computers/builds that's correct but where it hit me was wanting to upgrade an existing desktop. I already have more DDR4 RAM than I need and would have been willing to purchase a new CPU/motherboard and being forced to also purchase new RAM at the same time made it too big of a price tag all at once. I just found the best zen 3 cpu I could on ebay and called it a day.

I think your point still stands overall for AMD's business though, I assume a vast majority of CPUs are purchased in new desktops?

wing-_-nuts an hour ago

I built a workstation / gaming pc in 2024, and I feel like I was on the last chopper out of 'nam

mahirsaid 4 hours ago

Agree, i just built a desktop for the first time in ages, it is a leap and change from using laptops with numerous components pplugged into them. i made the leap to desktop. everything was comparably reasonable except the RAM or anything that has memory chip on it ( RAM, NVME etc) so i did some research just to make sure. All in all i happy with the result i went with AMD 9900x no graphics card in this option, i skipped the graphics card for now.

mahirsaid 4 hours ago

giancarlostoro 7 hours ago

If I can use it with Linux in any meaningful way, that would be a better selling point.

jakogut 6 hours ago

You in fact can now! In the past week, a transformer framework called FastFlowLM [0] supporting XDNA 2 NPUs officially started supporting Linux.

I posted it here the same day I found and started using it, to almost no reaction.

[0] https://github.com/FastFlowLM https://fastflowlm.com/ https://huggingface.co/FastFlowLM

giancarlostoro 4 hours ago

wing-_-nuts an hour ago

vyr 5 hours ago

SunshineTheCat 7 hours ago

Now all the features you don't use can perform 20% faster!

mikepurvis 7 hours ago

Upgrading to AM5 wasn't compelling to me even last summer before things went bonkers; I'm still very content with my 5800x and 64GB of DDR4.

Trying to take the plunge on that now sounds like a nightmare.

hu3 an hour ago

I got lucky on the timing and got 9800X3D + 64GB DDR5 before prices increased.

Your machine is sweet and probably runs just as fast on most tasks. I wouldn't be in n a hurry to upgrade.

I upgraded from an old Intel i7.

jmward01 3 hours ago

To really take advantage of those gpu cores you need memory bandwidth. Modern transformer based LLMs are really bandwidth hungry. I am really happy to see this first push. NVIDIA having discrete GPU/memory/etc is an option, but not great for a lot of different reasons. Unified memory architectures like what AMD and Apple have are the way to go for the future. Put 256GB of ram on the main board and be able to access it at speed for LLM use please.

downrightmike 7 hours ago

Just like the previous generation of AI PC, consumers just need a usb/pcie NPU,

Mass adoption won't happen until we get those cheap, because there are no mass prosumers making software for them that is massively popular.

u8080 7 hours ago

No, AI inference is mainly RAM/RAM speed constrained, we need more fast RAM to make local AI thrive.

downrightmike 6 hours ago

tuukkah 16 hours ago

Meanwhile, the corresponding "non-standard" desktop PC is the Framework Desktop, which with the Ryzen AI Max+ 395 can use 120GB of its 128GB RAM for the GPU: How to Run a One Trillion-Parameter LLM Locally: An AMD Ryzen™ AI Max+ Cluster Guide https://www.amd.com/en/developer/resources/technical-article...

hedgehog 15 hours ago

Minisforum MS-S1 is the same chip but has a PCIe slot suitable for a network card.

jsheard 12 hours ago

The Framework Desktop motherboard does actually have a PCIe x4 slot, their case just doesn't expose it for whatever reason. But you can buy the board separately and put it in your own choice of Mini-ITX case which does.

gpff 11 hours ago

AmVess 10 hours ago

hedgehog 11 hours ago

SkyMarshal 13 hours ago

Or a Beelink GTR9 Pro with same chip and memory, and two 10GbE LAN ports built in for clustering several together.

hedgehog 11 hours ago

tuukkah 12 hours ago

In their article, AMD chose to use the built-in 5Gbe NIC, but I read you could use the two USB4 @40Gbps ports as interconnects too (or for USB NICs).

qalmakka 14 hours ago

Yeah, but all of this is pointless when RAM is as expensive as two CPUs by itself - if it's even in stock. AMD/Intel should focus on that first if they want to save their DIY business at all - which I'm starting to doubt they don't

Ajedi32 6 hours ago

The RAM shortage is... a shortage. It's temporary by nature. RAM didn't suddenly get 4x more expensive to produce, it's just in high demand right now. Supply will eventually catch up even if it takes a few years.

mort96 an hour ago

Most of us are not really doing detailed half-decade hardware purchasing plans, for most of us "the next few years" is really the only relevant time frame.

Aldipower 41 minutes ago

andriy_koval 4 hours ago

> Supply will eventually catch up even if it takes a few years.

I am wondering if this so true. What resources and time are needed to increase supply by N times to catch demand.

Ajedi32 3 hours ago

wmf 3 hours ago

littlestymaar 3 hours ago

zardo 4 hours ago

Or they'll just enjoy the high margins and not invest significantly more than they normally would in new production capacity.

hamdingers 3 hours ago

Ajedi32 3 hours ago

DiabloD3 13 hours ago

They can't fix stupid.

Let me describe this in the most simple terms possible: You have speculators speculating about AI products. The speculators are not very smart when it comes to technology, and think RAM is RAM. There is at least three kinds of RAM that are important to this: DDR for system RAM, GDDR for GPUs, and HBM for high density enterprise products, and they are not interchangeable, there is no one-die-fits-all solution.

So, these speculators are like "oh no, more GPUs requires more RAM!", and then just start speculating on all RAM. Which of these RAMs are the ones that they need to worry about? Exclusively HBM, which is a minority in production, DDR and GDDR dominate production.

If you're into inference, and have older machines, you're buying Hxxx or Bxxx cards that use HBM, fit into dual slot x16 configurations, and you're jamming (optimally) 8 of them in. If you're into hardware that is newer, somewhere in the middle of the inference boom, you're using MXM cards. In either situation, the host machine has DDR, but if you're OpenAI, Anthropic, Microsoft, or Google, you're not building (more) inference machines like this.

The first two are buying Nvidia's all in one SBC solution: unified HBM, onboard ARM CPU to babysit the dual GPUs, has its own dual QSFP network controller that can RDMA, etc. No DDR or GDDR involved. Any machines built before this platform are being phased out entirely.

Microsoft is doing the same, but with AMD's products, the MI series that co-locates Epyc-grade Zen 4/5 CCDs with CDNA compute chiplets, running the entire thing off HBM, thus also unified and no DDR/GDDR needed. They, too, are phasing out machines older than this.

Google has a mix: they offer Nvidia all in one SBCs as part of GCP for legacy inference tasks (so your stack that can't run on AMD yet still can run), but also offer the same MI products that Microsoft offers via Azure's inference product, but also has their own TPUs that some of Gemini runs on; the TPUs run on HBM afiact. No DDR or GDDR here.

So, what does AMD or Intel do here? Lets say they waste fab time to make their own dies on the wrong process (TSMC and Intel-Foundry do not have for-RAM optimized processes)... they would be producing DDR and GDDR for a market that almost has its entire demand met. Intel lacks the die stacking technology required to build HBM, and TSMC I think can't do it for that many layers (HBM has 8 to 16 layers in current gen stuff iirc).

Micron, for example, already is bringing two large factories online here in the US to meet the projected growth in demand for the next 20+ years. When these factories finally start producing, it will not change the minds of speculators: they still seem to think AI datacenters need RAM, of any kind, and refuse to understand even the most basics of nuance. Also, when they come online, HBM will be a minority product; the AI inference boom is still just a bump in the road for them.

Nvidia kinda screwed their consumer partners, btw: they no longer bundle the GDDR required for the card with the purchase of the die. There is a slight short term bump in GDDR spot prices as partners are building up warchests to push series 60 GPUs into production, and once that is done, spot prices return to normal (outside of the wild speculation manipulation).

One last thing: what about LPDDR, used by AMD Strix Halo and Apple stuff? Speculation seems to have not actually effected it. I consider it as a sub-category of DDR (and some dies seem to work as either DDR or LPDDR as of DDR5, due to the merger of the specs by JEDEC), but since it isn't something you find in datacenters, it seems to have avoided speculation.

The Ryzen Max CPUs mentioned in the linked article? Uses LPDDR. Doubling down on the Ryzen Max product line might be a brilliant move.

PunchyHamster 13 hours ago

> The speculators are not very smart when it comes to technology, and think RAM is RAM. There is at least three kinds of RAM that are important to this: DDR for system RAM, GDDR for GPUs, and HBM for high density enterprise products, and they are not interchangeable, there is no one-die-fits-all solution.

The commenter is also not very smart and does not realize companies making the RAM can trade capacity of one for another and any re-tooling at current price is still profitable.

The commenter also does not realize that is also true for lines currently making SSDs

DiabloD3 12 hours ago

nopurpose 12 hours ago

cheschire 9 hours ago

Those micron factories won’t even be targeted at consumer-grade RAM though, right?

DiabloD3 9 hours ago

throw2847575 12 hours ago

> what about LPDDR, used by AMD Strix Halo and Apple stuff? Speculation seems to have not actually effected it

Good luck actually finding them on stock with 128GB+ RAM. I got strix laptop while ago, now price in EU is technically the same, but no stock. Maybe month or three

There is also claw hype. And large gwen3.5 models can run very well on DDR5 CPUs or mac minis...

corimaith 11 hours ago

I find the panic over RAM prices to be overestimated. 32GB DDR5 RAM is around $500 which is comparable to to the 9800x3D. Sure it sucks that it increases by around 4x, but when you factor in the overall price of a top end PC at around 1000-2000, especially for the lion's sum of the GPU, the increase is marginal.

This only effects a very narrow slice of highly budget conscious consumers trying to build high end PCs at razor thin margins.

zozbot234 10 hours ago

$500 for 32GB is about $15/GB which is a high we haven't seen since the mid-2000s. This is a big deal, it turns RAM and to some extent storage (especially fast storage) into a massive economic bottleneck.

BoredomIsFun 10 hours ago

Sohcahtoa82 4 hours ago

> 32GB DDR5 RAM is around $500 which is comparable to to the 9800x3D.

Apples to oranges. Why are you comparing RAM prices to CPU prices? It's different hardware.

$500 for 32 GB is insane. Just 18 months ago, I bought 128 GB of DDR5 for only $480.

no_ja 11 hours ago

I disagree with you. The issue does not only affect a “very narrow slice” of consumers. https://www.techspot.com/news/111472-hp-warns-ram-now-makes-... A major brand is now suggesting that this is a “new normal” and one solution is to just offer systems with less ram. This is an issue when lots of modern software seems to expect an unending supply.

smcleod 10 hours ago

That is an insane amount of money for just 32GB of RAM! That's what we were paying back when it was hard to use more than 32-64GB in a desktop setting. These days with all the electron and node bloatware, containers everywhere and AI - 32GB doesn't get you far.

Forgeties79 9 hours ago

$500 is 5x what it cost less than a year ago, just for context. It turns a $1600 computer build into a $2000 one. That’s a huge difference.

Edit: I don’t get your math. If we’re using a very generous definition of “top end,” even neglecting Nvidia and going AMD - which some would argue makes it not top end - you’re talking conservatively: $600 for a GPU, $500 for 32gb of ram, and $500 for a CPU. $1600 before PSU, case, SSD, fan(s), mobo…there’s no world in which you’re coming in under $2k. The SSD and board will put you over immediately.

You’re talking 3/2025 prices, not 3/2026. A compromise, mid-range computer is $1500 to build now.

mahirsaid 4 hours ago

FpUser 11 hours ago

>"overall price of a top end PC at around 1000-2000"

All 4 of my "top end PCs" have 128GB RAM. Me server (I self host everything is 512GB). Lucky for me all were bought before that insanity.

EtienneK 6 hours ago

In time for Windows 12 that reportedly will require an NPU: https://tech4gamers.com/windows-12-reportedly-relasing-2026-...

wtallis 3 hours ago

That's a ridiculously implausible and sensationalized rumor. At most, Microsoft may make a NPU a requirement for OEMs to use the Windows 12 logo on new PCs. Actually refusing to support the existing install base of recent and hughly-capable desktops is not at all likely. It would be far more drastic than the hardware deprecations brought by Windows 11, which were already quite controversial (and loosely enforced).

snovv_crash 16 hours ago

Hoe much dedicated cache do these NPUs have? Because it's easy enough to saturate the memory bandwidth using the CPU for compute, never mind the GPU. Adding dark silicon for some special operations isn't going to make out memory bandwidth faster.

bjackman 14 hours ago

Does a cache help with inference workloads anyway?

I don't know much about it but my mental model is that for transformers you need random access to billions of parameters.

fc417fc802 11 hours ago

It's streaming access, and no not as far as I'm aware. APUs have always been hilariously bottlenecked on memory bandwidth as soon as your task actually needed to pull in data. The only exception I know of is the PS5 because it uses GDDR instead of desktop memory.

zigzag312 11 hours ago

Are we going to see more memory channels for consumer desktop at some point from AMD or Intel? Apple seems to be the only one that offers it.

kristianp 2 hours ago

Am6 Socket isn't due for a couple of years. The current sTR5 used by Threadripper processors supports up to 8 channels of memory.

hypercube33 10 hours ago

I'm not sure what you mean - I think the mobile 300 series can do quad channel already for its APU at least. I'd assume it can do more but do you not need more slots beyond that?

wtallis 2 hours ago

zigzag312 9 hours ago

zozbot234 10 hours ago

NPUs are more useful for prefill than decode anyway. Memory bandwidth is not the bottleneck for prefill.

Aardwolf 7 hours ago

What exactly is this:

Is this fast GPU like instructions that anyone can use in any operating system to run any open sourced LLMs on CPU with all your RAM rather than on a discrete GPU with its own limited amount of VRAM?

Or is this a proprietary thing that only works in Windows for some specific use cases and irrelevant for Linux users?

bpavuk 7 hours ago

it's a distinct piece of hardware based on AMD XDNA architecture, which, coincidentally, much like CPUs, can tap into your RAM pool. there are XDNA drivers (`amdxdna`) for Linux.

mathisfun123 6 hours ago

> can tap into your RAM pool

lol no it can't - there's a small (40MB) SRAM that can DMA to DRAM and then each of the tiles is another DMA away from that SRAM.

cebert 3 days ago

AMD marketing is hoping the “AI” branding is a positive. Antidotally, I know many consumers who are not sold on AI. This branding could actually hurt sales.

marcosdumay 8 hours ago

Well, as long as it's one product of many, they may be able to segment the market and suffer no drawback.

aljgz 16 hours ago

We are dealing with a hype, but the reality is that AI would change everything we do. Local models will start being helpful in [more] unobtrusive ways. Machines with decent local NPUs would be usable for longer before they feel too slow.

ezst 14 hours ago

> the reality is that AI would change everything we do

Your true believer convictions don't matter here. Those AI accelerators are merely just marketing stunts. They won't help your local inference because they are not general purpose enough for that, they are too weak to be impactful, most people won't ever run local inference because it sucks and is a resource hog most can't afford, and it goes against the interests of those scammy unprofitable corporations who are selling us LLMs as AI as the silver bullet to every problem and got us there in the first place (they are already successful in that, by making computing unaffordable). There's little to no economical and functional meaning to those NPUs.

squidbeak 8 hours ago

otabdeveloper4 10 hours ago

BoredomIsFun 10 hours ago

upboundspiral 5 hours ago

I am bullish on AI being used in all sorts of useful and discreet and non-discreet ways in the present and future. However I am exceedingly skeptical of NPUs being some winning bet.

No one is running LLMs on current gen NPUs so if we will in the future its a long time coming. Unless they can demonstrate some real (and not marketing) wins I remain skeptical that a large NPU for LLMs is the future.

I can totally see NPU accelerating simple tasks, but to be worth the silicon they have a ways to go imo.

99% of people don't need or want a dev workstation. My travel laptop is 7+ years old and I couldn't tell you the difference between it and a current flagship in terms of browsing and everyday tasks.

I will not lie, I find LLMs useful but the desktop experience is pretty polished already. NPUs seem to be an attempt to ride the AI bandwagon with very little to show for it so far.

vbezhenar 16 hours ago

For some people maybe. I don't want to use local AI and NPU will be dead weight for me. Can't imagine a single task in my workflow that would benefit from AI.

It's similar to performance/effiency cores. I don't need power efficiency and I'd actually buy CPU that doesn't make that distinction.

wtallis 16 hours ago

fodkodrasz 16 hours ago

orbital-decay 16 hours ago

g947o 10 hours ago

Your comment is almost completely irrelevant to what the parent is saying. "AI would change everything we do" has nothing to do with "This new chip along with bloat from Windows enables new workflows for you". If you have been paying attention, you'd know that NPUs from these new CPUs barely made any difference from a consumer's perspective.

saithound 13 hours ago

This is irrelevant when the question is whether marketing your CPU with "AI" will help sales.

Toilets also changed everything we do and are helpful in unobtrusive ways, but that won't make the "Ryzen Crapper" a customer favorite.

wood_spirit 16 hours ago

So I’ve got a lot warmer to believing that AI can be a better programmer than most programmers these days. That is a low bar :). The current approach to AI can definitely change how effective a programmer is: but then it is up to the market to decide if we need so many programmers. The talk about how each company is going to keep all the existing programmers and just expect productivity multipliers is just what execs are currently telling programmers; that might change when the same is execs are talking to shareholders etc.

But does this extrapolate to the current way of doing AI being in normal life in a good way that ends up being popular? The way Microsoft etc is trying to put AI in everything is kinda saying no it isn’t actually what users want.

I’d like voice control in my PC or phone. That’s a use for these NPUs. But I imagine it is like AR- what we all want until it arrives and it’s meh.

snovv_crash 16 hours ago

PunchyHamster 13 hours ago

Doubt it, and competition does the same so not like they are going intel coz of that

himata4113 17 hours ago

I'd actually love to have an NPU that isn't useless on my 285k.

shakna 13 hours ago

Or hoping to get ahead of the Windows 12 requirements.

skirmish 16 hours ago

Indeed, I was buying a laptop for my wife, and she was viscerally against "Ryzen AI": I don't want a CPU with builtin AI to spy on my screen all the time!

kijin 16 hours ago

They can just buy a regular Ryzen 9000 series CPU, then. Maybe add a real graphics card if they're into gaming.

Havoc 14 hours ago

Some of these supporting ECC is of some interest. Though fast udimm ecc ram is going to be extremely expensive

2001zhaozhao 3 hours ago

Interestingly, this tops off at 8 cores. Where are the 16 core versions that already exist on laptops?

Buttons840 16 hours ago

Do we expect special AI processors to diverge from GPUs? Like, processors that can do parallel neural network computations but cannot draw graphics?

c0balt 16 hours ago

That is already the case with datacenter "GPUs". A A100, MI300 or Intel PVC/Gaudi does not have useful graphics performance nor capabilities. Coprocessors ala NPU/VPU are also on the rise again for CPUs.

elcritch 15 hours ago

Great now I’m envisioning a rich guy using an A100 as his desktop GPU just to show off. Which begs the question if that’s even possible.

userbinator 15 hours ago

vel0city 9 hours ago

amelius 9 hours ago

I'm expecting in the not so distant future we'll even have LLMs baked into an ASIC, in our PCs.

dagmx 16 hours ago

That’s already the norm no?

Pretty much every hardware vendor has an NPU

pjmlp 14 hours ago

Yes, this has already been the case for years on mobile devices, CoPilot+ PC design requires this approach as well.

Additionally, GPUs are going back to the early days, by becoming general purpose parallel compute devices, where you can use the old software rendering techniques, now hardware accelerated.

PunchyHamster 13 hours ago

We kinda already have it with NPU/TPUs, tho they are usually attached to CPUs (and for some reason come with near zero proper documentation).

I can see separate cards for datacenter use but for consumers they will probably come on same SOC as CPU

jiggawatts 16 hours ago

Yes.

Even the latest NVIDIA Blackwell GPUs are general purpose, albeit with negligible "graphics" capabilites. They can run fairly arbitrary C/C++ code with only some limitations, and the area of the chip dedicated to matrix products (the "tensor units") is relatively small: less than 20% of the area!

Conversely, the Google TPUs dedicate a large area of each chip to pure tensor ops, hence the name.

This is partly why Google's Gemini is 4x cheaper than OpenAI's GPT5 models to serve.

Jensen Huang has said in recent interviews that he stands by the decision to keep the NVIDIA GPUs more general purpose, because this makes them flexible and able to be adapted to future AI designs, not just the current architectures.

That may or may not pan out.

I strongly suspect that the winning chip architecture will have about 80% of its area dedicated to tensor units, very little onboard cache, and model weights streamed in from High Bandwidth Flash (HBF). This would be dramatically lower power and cost compared to the current hardware that's typically used.

Something to consider is that as the size of matrices scales up in a model, the compute needed to perform matrix multiplications goes up as the cube of their size, but the other miscellaneous operations such as softmax, relu, etc.. scale up linearly with the size of the vectors being multiplied.

Hence, as models scale into the trillions of parameters, the matrix multiplications ("tensor" ops) dominate everything else.

kcb 5 hours ago

The 100 class Nvidia chips are targeted at training. With Nvidia buying Groq it will further move in that direction.

IsTom 9 hours ago

I'm not following the whole LLM space, but

> the compute needed to perform matrix multiplications goes up as the cube of their size,

are they really not using even Strassen multiplication?

jcranmer 8 hours ago

jiggawatts 8 hours ago

mcraiha 16 hours ago

These are mobile chips shoehorned into AM5. They aren't very good e.g. for gaming purposes. https://videocardz.com/newz/amd-ryzen-ai-400-does-not-suppor...

anticorporate 10 hours ago

I guess it depends on what games you play. I have an AI Max 395 (Framework Desktop) and it runs every game in my library flawlessly. I'm sure if I played this year's most resource-intensive games it might stutter, but I don't. For me, it's an amazing low power minpc doing triple duty as a gaming PC, development box, and running my self-hosted services for the rest of the house.

AmVess 10 hours ago

Yeah. I have one. People buying the Framework Desktop mainboard aren't buying it just for gaming. There are better and far cheaper options for gaming. What this does is everything, though. Good enough for 1440p gaming. 16c/32t powerful CPU, it can run LLM's. SFF main pc that can do everything in a tiny space is a win.

bcraven 15 hours ago

Presumably that's why the subheading is:

>First wave of Ryzen AI desktop CPUs targets business PCs rather than DIYers.

iso-logi 17 hours ago

8 Core/16 Thread, boosting up to 5.1GHz with iGPU would be pretty neat for a Plex Server or Proxmox Server with a few VMs.

zeroflow 15 hours ago

As far as I can find, Plex does not support AMD iGPU for transcoding. Jellyfin will work, but support seems rather spotty. For other AI/ML work, it seems like ROCm is up and coming, but support - e.g. for Frigate object detection - is still a work in progress, especially for newer chips.

hamdingers 3 hours ago

Jellyfin supports it, but the resulting quality is noticeably poor compared to Intel QuickSync or software transcoding. Perhaps the newer chips are better, but if you're building a media server from scratch you'd probably build around an Intel CPU or ARC GPU anyway.

shiroiuma 14 hours ago

I have an AM4 AMD iGPU I use with Jellyfin; it works fine.

happymellon 13 hours ago

jeroenhd 13 hours ago

My proxmox server with a few VMs works perfectly fine with much less compute.

My homelab setup runs out of memory much faster than it does CPU cores.

threetonesun 10 hours ago

Depends what you’re doing on the VMs, I run one as a desktop PC so have 4/6 cores and all the GPU access is important.

Mashimo 15 hours ago

Maybe also Immich, for face and object recognition.

rafaelmn 7 hours ago

PC desktop chips have 128 bit busses to right ? From what I know theoretical maximum memory bandwidth of chips is less than 100gbps - which is less than like base M5/M4 chips.

So no matter how much compute you stuff in there it's going to be shit for AI ?

PC architecture is not adapting to AI workloads at all, and no signs of that changing in years to come. I would not be surprised if your phone was more capable of running AI models than an average desktop - especially given gpu pricing.

HighGoldstein 5 hours ago

Apple has their high-bandwidth chips, the rest of the commercial desktop market is effectively running Windows, and Microsoft has no incentive to move towards local AI, their ideal case is that you use their cloud-based services and pay for them forever (you being enterprise clients with thousands of PCs).

lelanthran 15 hours ago

It doesn't sound as impressive as I wanted :-(

I wanted a better strix halo (which has 128GB unified RAM and 40cu on the 8080s (or something) iGPU).

This looks like normal Ryzen mobile chips + but with fewer cus.

wtallis 15 hours ago

Putting Strix Halo into the AM5 socket would make no sense. Half the memory controllers would be orphaned and the GPU would be severely bandwidth-starved (assuming that the memory controller on Strix Halo actually supports DDR5 and not just LPDDR5).

PunchyHamster 13 hours ago

They could put it on threadripper socket, very similar memory bandwidth

undersuit 3 hours ago

noelwelsh 15 hours ago

Yeah the next generation of Strix Halo is what would get me excited. I think right now TSMC has no capacity, so maybe we have to wait another year. Kinda ironic that all CPU/RAM capacity is being sold to LLM companies, and as a result we can't get the hardware needed for good local LLMs.

qalmakka 13 hours ago

> all CPU/RAM capacity is being sold to LLM companies, and as a result we can't get the hardware needed for good local LLMs.

yeah... Ironic I guess. It's as if they've realised that it's only a matter of time until we get a "good enough" FOSS model that runs on consumer hardware. The fact that such a thing would demolish their entire business of getting VC hyped while giving out their service for a loss surely got lost to them. Surely they and Nvidia have not realised that the only thing that could stop this is to make good hardware unreachable for anything smaller than a massive corp

Mark my words: in less than one year, we'll probably get something akin to Opus 4.6 FOSS. China is putting as much money into that as they can because they know this would crash the US economy, which is in the green only thanks to big tech pumping up AI. China wants Trump either gone or neutered as soon as possible, which they know they can do by making Republicans as unelectable as possible - something that will probably do if the economy crashes and a recession happens

mixxit 10 hours ago

I can use this in OpenWeb on Unraid? Save me buying a pascal card?

ilovechaz 9 hours ago

50 tops ain’t all that much.

craftkiller 8 hours ago

Since this is for desktop, the NPU is irrelevant. Consumer-grade NPUs are not made for high performance. They are optimized for low power consumption, which makes sense when you are trying to run basic AI tasks on a laptop without turning it into a frying pan. On consumer-grade gear, the GPU will outperform the NPU, and since a desktop is not constrained by battery and can dissipate far more heat, the NPU is mostly irrelevant (aside from using it to develop software for laptop NPUs).

DeathArrow 13 hours ago

"AI" branding applied to subpar products hoping to boost sales.

poly2it 16 hours ago

The Ryzen AI line is actually great if deployed to an entire org as the bottom tier, as it garuantees every device has a 50 TOPs NPU. We deploy local software at $STARTUP and this makes deployment to a Windows corp more predictable.

heraldgeezer 6 hours ago

So happy I got my 9800X3D + 64GB RAM this fall instead of now lmaooo

FpUser 16 hours ago

Well, for me personally it is a meh until RAM prices go down. Suddenly, decent PC has turned from a tool accessible to average Joe to a luxury item

bitwize 16 hours ago

Narrator: The RAM prices did not, in fact, go down.

wosined 10 hours ago

Looks like someone has run out of money.

a012 15 hours ago

> This makes them AMD’s first desktop chips to qualify for Microsoft’s Copilot+ PC label, which enables a handful of unique Windows 11 features like Recall and Click to Do.

Microsoft: "Friendship ended with Intel, now AMD is my best friend"

pjmlp 14 hours ago

Actually it is Qualcom, as they keep trying to push for ARM, but due to the way PC ecosystem has been going since the IBM PC clones started, no one is rushing out to adopt ARM.