I realize this has not much to do with CPU choice per se, but I'm still gonna leave this recommendation here for people who like to build PCs to get stuff done with :) Since I've been able to afford it and the market has had them available, I've been buying desktop systems with proper ECC support.
I've been chasing flimsy but very annoying stability problems (some, of course, due to overclocking during my younger years, when it still had a tangible payoff) enough times on systems I had built that taking this one BIG potential cause out of the equation is worth the few dozens of extra bucks I have to spend on ECC-capable gear many times over.
Trying to validate an ECC-less platform's stability is surprisingly hard, because memtest and friends just aren't very reliably detecting more subtle problems. PRIME95, y-cruncher and linpack (in increasing order of effectiveness) are better than specialzied memory testing software in my experience, but they are not perfect, either.
Most AMD CPUs (but not their APUs with potent iGPUs - there, you will have to buy the "PRO" variants) these days have full support for ECC UDIMMs. If your mainboard vendor also plays ball - annoyingly, only a minority of them enables ECC support in their firmware, so always check for that before buying! - there's not much that can prevent you from having that stability enhancement and reassuring peace of mind.
So I'm trying to learn more about this stuff, but aren't there multiple ECC flavors and the AMD consumer CPUs only support one of them (not the one you'd have on servers?)
Does anyone maintain a list with de-facto support of amd chips and mainboards? That partlist site only shows official support IIRC, so it won't give you any results.
dijit · 20h ago
> only a minority of them enables ECC support in their firmware, so always check for that before buying!
This is the annoying part.
That AMD permits ECC is a truly fantastic situation, but if it's supported by the motherboard is often unlikely and worse: it's not advertised even when it's available.
I have an ASUS PRIME TRX40 PRO and the tech specs say that it can run ECC and non-ECC but not if ECC will be available to the operating system, merely that the DIMMS will work.
It's much more hit and miss in reality than it should be, though this motherboard was a pricey one: one can't use price as a proxy for features.
c0l0 · 20h ago
Usually, if a vendor's spec sheet for a (SOHO/consumer-grade) motherboard mentions ECC-UDIMM explicitly in its memory compatibility section, and (but this is a more recent development afaict) DOES NOT specify something like "operating in non-ECC mode only" at the same time, then you will have proper ECC (and therefore EDAC and RAS) support in Linux, if the kernel version you have can already deal with ECC on your platform in general.
I would assume your particular motherboard to operate with proper SECDED+-level ECC if you have capable, compatible DIMM, enable ECC mode in the firmware, and boot an OS kernel that can make sense of it all.
consp · 20h ago
Isn't it mostly an ease of mind thing? I've never seen a ECC error on my home server which has plenty of memory in use and runs longer than my desktop. Maybe it's more common with higher clocked, near the limit, desktop PC's.
Also: DDR5 has some false ecc marketing due to the memory standard having an error correction scheme build in. Don't fall for it.
wpm · 2m ago
I had a somewhat dodgy stick of used RAM (DDR4 UDIMM) in a Supermicro X11 board. This board is running my NAS, all ZFS, so RAM corruption can equal data corruption. The OS alerted me to recoverable errors on DIMM B2. Swapped it and another DIMM, rebooted, saw DIMM error on slot B1. Swapped it for a spare stick. No more errors.
This was running at like, 1866 or something. It's a pretty barebones 8th gen i3 with a beefier chipset, but ECC still came in clutch. I won't buy hardware for server purposes without it.
c0l0 · 19h ago
I see a particular ECC error at least weekly on my home desktop system, because one of my DIMMs doesn't like the (out of spec) clock rate that I make it operate at. Looks like this:
94 2025-08-26 01:49:40 +0200 error: Corrected error, no action required., CPU 2, bank Unified Memory Controller (bank=18), mcg mcgstatus=0, mci CECC, memory_channel=1,csrow=0, mcgcap=0x0000011c, status=0x9c2040000000011b, addr=0x36e701dc0, misc=0xd01a000101000000, walltime=0x68aea758, cpuid=0x00a50f00, bank=0x00000012
95 2025-09-01 09:41:50 +0200 error: Corrected error, no action required., CPU 2, bank Unified Memory Controller (bank=18), mcg mcgstatus=0, mci CECC, memory_channel=1,csrow=0, mcgcap=0x0000011c, status=0x9c2040000000011b, addr=0x36e701dc0, misc=0xd01a000101000000, walltime=0x68b80667, cpuid=0x00a50f00, bank=0x00000012
(this is `sudo ras-mc-ctl --errors` output)
It's always the same address, and always a Corrected Error (obviously, otherwise my kernel would panic). However, operating my system's memory at this clock and latency boosts x265 encoding performance (just one of the benchmarks I picked when trying to figure out how to handle this particular tradeoff) by about 12%. That is an improvement I am willing to stomach the extra risk of effectively overclocking the memory module beyond its comformt zone for, given that I can fully mitigate it by virtue of properly working ECC.
ainiriand · 1h ago
I think you've found a particularly weak memory cell, I would start thinking about replacing that module. The consistent memory_channel=1, csrow=0 pattern confirms it's the same physical location failing predictably.
Hendrikto · 19h ago
Running your RAM so far out of spec that it breaks down regularly, where do you take the confidence that ECC will still work correctly?
Also: Could you not have just bought slightly faste RAM, given the premium for ECC?
c0l0 · 16h ago
"Breaks down" is a strong choice of words for a single, corrected bit error. ECC works as designed, and demonstrates that it does by detecting this re-occurring error. I take the confidence mostly from experience ;)
And no, as ECC UDIMM for the speed (3600MHz) I run mine at simply does not exist - it is outside of what JEDEC ratified for the DDR4 spec.
adithyassekhar · 16h ago
JEDEC rated DDR4 at only 2400mhz right? And anything higher is technically over clocking?
DDR4-3200 (PC4-25600) (the highest supported in the DDR4 generation)
What's *NOT* supported are some enthusiast ones that typically require more than 1.2v for example: 3600 MT/s, 4000 MT/s & 4266 MT/s
c0l0 · 16h ago
JEDEC specifies rates up to 3200MT/s, what's officially referred to as DDR4-3200 (PC4-25600).
kderbe · 12h ago
I would loosen the memory timings a bit and see if that resolves the ECC errors. x265 performance shouldn't fall since it generally benefits more from memory clock rate than latency.
Also, could you share some relevant info about your processor, mainboard, and UEFI? I see many internet commenters question whether their ECC is working (or ask if a particular setup would work), and far fewer that report a successful ECC consumer desktop build. So it would be nice to know some specific product combinations that really work.
BikiniPrince · 3h ago
I pick up old serves for my garage system. With edac it is a dream to isolate the fault and be instantly aware. It also lets you determine the severity of the issue. Dimms can run for years with just the one error or overnight explode into streams of corrections. I keep spares so it’s fairly easy to isolate any faults. It’s just how do you want to spend your time?
hedora · 3h ago
You have to go pretty far down the rabbit hole to make sure you’ve actually got ECC with [LP]DDR5
Some vendors use hamming codes with “holes” in them, and you need the CPU to also run ECC (or at least error detection) between ram and the cache hierarchy.
Those things are optional in the spec, because we can’t have nice things.
Scramblejams · 3h ago
I run a handful of servers and I have a couple that pop ECC errors every year or three, so YMMV.
immibis · 16h ago
I saw a corrected memory error logged every few hours when my current machine was new. It seems to have gone away now, so either some burn-in effect, or ECC accidentally got switched off and all my data is now corrupted. Threadripper 7000 series, 4x64GB DDR5.
Edit: it's probably because I switched it to "energy efficiency mode" instead of "performance mode" because it would occasionally lock up in performance mode. Presumably with the same root cause.
swinglock · 20h ago
Excellent point. It's a shame and a travesty that data integrity is still mostly locked away inside servers, leaving most other computing devices effectively toys, the early prototype demo thing but then never finished and sold forever at inflated prices.
I wish AMD would make ECC a properly advertised feature with clear motherboard support. At least DDR5 has some level of ECC.
wpm · 51s ago
I wish AMD wouldn't gate APU ECC support behind unobtainium "PRO" SKUs they only give out, seemingly, to your typical "business" OEMs and the rare Chinese miniPC company.
kevin_thibedeau · 4h ago
> At least DDR5 has some level of ECC.
That is mostly to assist manufacturers in selling marginal chips with a few bad bits scattered around. It's really a step backwards in reliability.
wallopinski · 30m ago
Funny and depressing that the AMD/Intel culture war still exists. I remember arguing about it in 1990. Their marketing departments severely brainwashed generations of nerds.
cjpartridge · 22m ago
Try enabling PBO and finding a setting for the curve optimizer that works for you, each CPU is different but -10/-15 is generally achievable - should reduce temperatures across the board and potentially give you some more performance.
danieldk · 21h ago
I feel like both Intel and AMD are not doing great in the desktop CPU stability department. I made a machine with a Ryzen 9900X a while back and it had the issue that it would freeze when idling. A few years before I had a 5950X that would regularly crash under load (luckily it was a prebuilt, so it was ultimately fixed).
When you do not have a bunch of components ready to swap out it is also really hard to debug these issues. Sometimes it’s something completely different like the PSU. After the last issues, I decided to buy a prebuilt (ThinkStation) with on-site service. The cooling is a bit worse, etc., but if issues come up, I don’t have to spend a lot of time debugging them.
Random other comment: when comparing CPUs, a sad observation was that even a passively cooled M4 is faster than a lot of desktop CPUs (typically single-threaded, sometimes also multi-threaded).
seec · 8h ago
Your comment about the passively cooled M4 is misleading.
Sure, in single thread, it will be definitely faster. In multithread unless you are going for low end or older CPUs it's basically a lie.
A 10 Core M4 will score around a 14TH gen mobile i5. It will consume much less power but the argument is on performance, so that's beside the point.
And if we are talking about a passively cooled M4 (MacBook Air basically) it will quite heavily throttle relatively quickly, you lose at the very least 30%.
So, let's not misrepresent things, Apple CPUs are very power efficient but they are not magic, if you hit them hard, they still need good cooling. Plenty of people have had the experience with their M4 Max, discovering that actually, if they did use the laptop as a workstation, it will generate a good amount of fan noise, there is no other way around.
Apple stuff is good because most people actually have bursty workload (especially graphic design, video editing and some audio stuff) but if you hammer it for hours on end, it's not that good and the power efficiency point becomes a bit moot.
bob1029 · 18h ago
I've got a 5950x that I can reliably crater with a very specific .NET 8 console app when it would otherwise be stable 24/7/365, even under some pretty crazy workloads like Unity.
I think a lot of it boils down to load profile and power delivery. My 2500VA double conversion UPS seems to have difficulty keeping up with the volatility in load when running that console app. I can tell because its fans ramp up and my lights on the same circuit begin to flicker very perceptibly. It also creates audible PWM noise in the PC which is crazy to me because up til recently I've only ever heard that from a heavily loaded GPU.
heelix · 16h ago
I wonder if cooling/power is really the key here. I've got a 5950x that ended up getting the water loop I'd intended for my next threadripper - only to find they were not selling the blasted things to anyone but a few companies. With the cooling sized for almost twice what the 5950x could put out, it has been a very stable machine for some crazy workloads. That old dog will likely keep the setup when a zen 5 TR gets swapped in.
For a long time, my Achille's heel was my Bride's vacuum. Her Dyson pulled enough amps that the UPS would start singing and trigger the auto shutdown sequence for the half rack. Took way too long to figure out as I was usually not around when she did it.
486sx33 · 5h ago
My 5950 didn’t like liquid cooling and lives very well with air cooling :)
bell-cot · 17h ago
I'm sure there are spec's for how fast a PS should be able to ramp up in response to spikes in demand, how a motherboard should handle sudden load changes, etc.
But if your UPS (or just the electrical outlet you're plugged into) can't cope - dunno if I'd describe that as cratering your CPU.
sunmag · 1h ago
Have had three systems (two 5800x, one 3600x) that reboots/freezes due to WHEA errors. Started after about 3years problem free.
One of the 5800xs so frequently it was trashed.
etempleton · 17h ago
My experience is similar. Modern enthusiast CPUs and hardware compatibility is going backwards. I have a 5900x that randomly crashes on idle, but not under load. My 285K has so far been rock solid and generally feels snappier. I feel like both Intel and AMD are really trying to push the envelope to look good on benchmarks and this is the end result.
naasking · 1h ago
Crash on idle, interesting. Must be some timing issue related to down clocking, or maybe a voltage issue related to shutting off a core.
protocolture · 5h ago
>I made a machine with a Ryzen 9900X a while back and it had the issue that it would freeze when idling
I also have this issue.
c0balt · 4h ago
If you are on Linux, there are long time known problems with low power cpu states. These states can be entered by your CPU when under low/no load.
A common approach is to go into the BIOS/UEFI settings and check that c6 is disabled. To verify and/or temporarily turn c6 off, see https://github.com/r4m0n/ZenStates-Linux
hedora · 2h ago
It’s also worth checking all the autodetected stuff that can be overclocked, like ram speed. That stuff can be wrong, and then you get crazy but similar bugs in linux and windows.
kristopolous · 4h ago
M3 ultra is the more capable chip by quite a bit. For instance: 80 GPU cores versus 10 in the m4.
Twice the memory bandwidth, twice the CPU core count... It's really wacky how they've decided to name things
bee_rider · 1h ago
Is there some tricky edge case here? I thought the “3” and “4” just denoted generations. The Ultra chips are like Apple’s equivalent of a workstation chip, they are always bigger right? It is like comparing the previous generation Xeon to a current gen i5.
dwood_dev · 1h ago
Apple has been iterating IPC as well as increasing core count.
The Ultra is a pair of Max chips. While the core counts didn't increase from M3 to M4 Max, overall performance is in the neighborhood of 5-25% better. Which still puts the M3 Ultra as Apple's top end chip, and the M5 Max might not dethrone it either.
The uplift in IPC and core counts means that my M1 Max MBP has a similar amount of CPU performance as my M3 iPad Air.
johnisgood · 21h ago
This does not fill me with much hope. What am I even ought to buy at this point then, I wonder. I have a ~13 years old Intel CPU which lacks AVX2 (and I need it by now) and I thought of buying a new desktop (items separately, of course), but that is crazy to me that it freezes because of the CPU going idle. It was never an issue in my case. I guess I can only hope it is not going to be a problem once I completed building my PC. :|
On what metric am I ought to buy a CPU these days? Should I care about reviews? I am fine with a middle-end CPU, for what it is worth, and I thought of AMD Ryzen 7 5700 or AMD Ryzen 5 5600GT or anything with a similar price tag. They might even be lower-end by now?
hhh · 21h ago
Just buy an AMD CPU. One person’s experience isn’t the world. Nobody in my circle has had an issue with any chip from AMD in recent time (10 years).
Intel is just bad at the moment and not even worth touching.
hedora · 2h ago
I went further and got an AMD system on chip machine with an integrated gpu. It’s fine for gaming and borderline for LLM inference (I should have put 64GB in instead of 32GB).
The only issues are with an intel Bluetooth chipset, and bios auto detection bugs. Under Linux, the hardware is bug for bug compatible with Windows, and I’m down to zero known issues after doing a bit of hardware debugging.
danieldk · 20h ago
I agree that Intel is bad at the moment (especially with the 13th and 14th gen self-destruct issues). But unfortunately I also know plenty of people with issues with AMD systems.
And it's no bad power quality on mains as someone suggested (it's excellent here) or 'in the air' (whatever that means) if it happens very quickly after buying.
I would guess that a lot of it comes from bad firmware/mainboards, etc. like the recent issue with ASRock mainboards destroying Ryzen 9000-series GPUs: https://www.techspot.com/news/108120-asrock-confirms-ryzen-9... Anyone who uses Linux and has dealt with bad ACPI bugs, etc. knows that a lot of these mainboards probably have crap firmware.
I should also say that I had a Ryzen 3700X and 5900X many years back and two laptops with a Ryzen CPU and they have been awesome.
tester756 · 21h ago
This is funny because recently my AMD Ryzen 7 5700X3D died and I've decided that my next CPU will be Intel
CPUs like Intel Core Ultra 7 265K are pretty close to top Ryzens
Panzer04 · 4h ago
Intel CPUs are decidedly better value when multicore performance is a concern. At the top end they trade blows.
If your workload is pointer-chasing intel's new CPUs aren't great though, and the X3D chips are possibly a good pick (if the workload fits in cache) which is why they get a lot of hype from reviewers who benchmark games and judge the score 90% based on that performance.
That's been year ago, my AMD CPU died very recently
johnisgood · 21h ago
That is what I thought, thanks.
ahofmann · 21h ago
I wouldn't be so hopeless. Intel and AMD CPUs are used in millions of builds and most of them just work.
danieldk · 20h ago
However, the vast majority of PCs out there are not hobbyist builds but Dell/Lenovo/HP/etc. [1] with far fewer possible configurations (and much more testing as a byproduct). I am not saying these machines never have issues, but a high failure rate would not be acceptable to their business customers.
[1] Well, most non-servers are probably laptops today, but the same reasoning applies.
giveita · 19h ago
If you value your time a dell laptop with extended warranty and accidental damage where they replace shit and send people out to fix shit is well worth it. It costs but you can be a dumb user and call "IT" when you need a fix and thats a nice feeling IMO!
homebrewer · 21h ago
It's either bad luck, bad power quality from the mains, or something in the air in that particular area. I know plenty of people running AM5 builds, have done so myself for the last couple of years, and there were no problems with any of them apart from the usual amdgpu bugs in latest kernels (which are "normal" since I'm running mainline kernels — it's easy to solve by just sticking to lts, and it has seemingly improved anyway since 6.15).
scns · 18h ago
> I thought of AMD Ryzen 7 5700
Definetly not that one if you plan to pair with a dedicated GPU! The 5700X has twice the L3 cache. All Ryzen 5000 with a GPU have only 16MB, 5700 has the GPU deactivated.
PartiallyTyped · 21h ago
3 of my last 4 machines have been AMD x NVDA and I have been very happy. The intel x NVDA machine has been my least stable one.
encom · 21h ago
>M4 is faster than a lot of desktop CPUs
Yea, but unfortunately it comes attached to a Mac.
An issue I've encountered often with motherboards, is that they have brain damaged default settings, that run CPU's out of spec. You really have to go through it all with a fine toothed comb and make sure everything is set to conservative stock manufacturer recommended settings. And my stupid MSI board resets everything (every single BIOS setting) to MSI defaults when you upgrade its BIOS.
homebrewer · 20h ago
Also be careful with overclocking, because the usual advice of "just running EXPO/XMP" often results in motherboards setting voltages on very sensitive components to more than 30% over their stock values, and this is somehow considered normal.
It looks completely bonkers to me. I overclocked my system to ~95% of what it is able to do with almost default voltages, using bumps of 1-3% over stock, which (AFAIK) is within acceptable tolerances, but it requires hours and hours of tinkering and stability testing.
Most users just set automatic overclocking, have their motherboards push voltages to insane levels, and then act surprised when their CPUs start bugging out within a couple of years.
Shocking!
danieldk · 20h ago
Unfortunately, some separately purchasable hardware components seem to be optimized completely for gamers these days (overclocking mainboards, RGB on GPUs, etc.).
I'd rather run everything at 90% and get very big power savings and still have pretty stellar performance. I do this with my ThinkStation with Core Ultra 265K now - I set the P-State maximum performance percentage to 90%. Under load it runs almost 20 degrees Celsius cooler. Single core is 8% slower, multicore 4.9%. Well worth the trade-off for me.
(Yes, I know that there are exceptions.)
hedora · 2h ago
I’ve had multiple systems that crash or corrupt data when underclocked, so running at 90% might not be what you want.
You can always play with the CPU governor / disable high power states. That should be well-tested.
anonymars · 1h ago
It sounds like you are conflating undervolting for underclocking. Undervolting runs it out of spec, while underclocking simply runs it slower
mschuster91 · 16h ago
> Unfortunately, some separately purchasable hardware components seem to be optimized completely for gamers these days
It turned out during the shitcoin craze and then AI craze that hardcore gamers, aka boomers with a lot of time and retirement money on their hands and early millennials working in big tech building giant-ass man caves, are a sizeable demographic with very deep pockets.
The wide masses however, they gotta live with the scraps that remain after the AI bros and hardcore gamers have had their pick.
to;dr: they heavily customize BIOS settings, since many BIOSes run CPUs out-of-spec by default. With these customizations there was not much of a difference in failure rate between AMD and Intel at that point in time (even when including Intel 13th and 14th gen).
ahartmetz · 20h ago
Since you mention EXPO/XMP, which are about RAM overclocking: RAM has the least trouble with overvoltage. Raising some of the various CPU voltages is a problem, which RAM overclocking may also do.
eptcyka · 20h ago
The heat is starting to become an issue for DDR5 with higher voltage.
electroglyph · 20h ago
yah, the default overclocking stuff is pretty aggressive these days
danieldk · 20h ago
Yea, but unfortunately it comes attached to a Mac.
Yeah. If Asahi worked on newer Macs and Apple Silicon Macs supported eGPU (yes I know, big ifs), the choice would be simple. I had NixOS on my Mac Studio M1 Ultra for a while and it was pretty glorious.
claudex · 20h ago
>And my stupid MSI board resets everything (every single BIOS setting) to MSI defaults when you upgrade its BIOS.
I had the same issue with my MSI board, next one won't be a MSI.
techpression · 19h ago
My ASUS and Gigabyte did the same too. I think vendors are being lazy and don’t want to write migration code
izacus · 19h ago
Did what exactly? All the ASUS and Gigabytes I've seen had PBO (which I guess you're talking about) disabled by default.
techpression · 17h ago
Reset all the settings with BIOS updates
philistine · 14h ago
I'd bet you don't care that it's attached to a Mac. I bet you don't want to switch OS. Which is understandable. In a couple of years, when Microsoft finally offers Windows as an ARM purchase, Linux is finally full-fledged done implementing support, and Apple resuscitates Boot Camp, I think a lot of people like you will look at Macs like the Mac Mini differently.
timmytokyo · 13h ago
Just get used to the extortionate prices on things like memory and storage. Who wouldn't want to pay $200 to go from a 16GB to a 24GB configuration? Who wouldn't want to pay $600 more for the 2TB storage option? And forget about upgrading after you buy the computer.
api · 17h ago
The M series chips aren’t the absolute fastest in raw speed, though they are toward the top of the list, but they destroy x86 lineage chips on performance per watt.
I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.
I’ve read people here and there arguing that instruction sets don’t matter, that it’s all the same past the decoder anyway. I don’t buy it. The superior energy efficiency of ARM chips is so obvious I find it impossible to believe it’s not due to the ISA since not much else is that different and now they’re often made on the same TSMC fabs.
AnthonyMouse · 1h ago
> they destroy x86 lineage chips on performance per watt.
This isn't really true. On the same process node the difference is negligible. It's just that Intel's process in particular has efficiency problems and Apple buys out the early capacity for TSMC's new process nodes. Then when you compare e.g. the first chips to use 3nm to existing chips which are still using 4 or 5nm, the newer process has somewhat better efficiency. But even then the difference isn't very large.
And the processors made on the same node often make for inconvenient comparisons, e.g. the M4 uses TSMC N3E but the only x86 processor currently using that is Epyc. And then you're obviously not comparing like with like, but as a ballpark estimate, the M4 Pro has a TDP of ~3.2W/core whereas Epyc 9845 is ~2.4W/core. The M4 can mitigate this by having somewhat better performance per core but this is nothing like an unambiguous victory for Apple; it's basically a tie.
> I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.
One of the reasons for this is that Apple has always been willing to run components right up to their temperature spec before turning on the fan. And then even though that's technically in spec, it's right on the line, which is bad for longevity.
In consumer devices it usually doesn't matter because most people rarely put any real load on their machines anyway, but it's something to be aware of if you actually intend to, e.g. there used to be a Mac Mini Server product and then people would put significant load on them and then they would eat the internal hard drives because the fan controller was tuned for acoustics over operating temperature.
ac29 · 13h ago
> I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.
This anecdote perfectly describes my few generation old Intel laptop too. The fans turn on maybe once a month. I dont think its as power efficient as an M-series Apple CPU, but total system power is definitely under 10W during normal usage (including screen, wifi, etc).
adithyassekhar · 16h ago
I'd rather it spin the fan all the time to improve longevity but that's just me.
One of the many reasons why snapdragon windows laptops failed was both amd and Intel (lunar lake) was able to reach the claimed efficiency of those chips. I still think modern x86 can match arm ones in efficiency if someone bothered to tune the os and scheduler for most common activities. M series was based on their phone chips which were designed from the ground up to run on a battery all these years. AMD/Intel just don't see an incentive to do that nor do Microsoft.
hedora · 2h ago
I have a modern AMD system on chip mini desktop that runs Linux (devuan), and have had M1/2/3 laptops. They all seem pretty comparable on power usage, especially at idle. Games and LLM load warm up the desktop and kill the laptop battery. Other than that, power consumption seems fine.
There is one exception: If I run an idle Windows 11 ARM edition VM on the mac, then the fans run pretty much all the time. Idle Linux ARM VMs don’t cause this issue on the mac.
I’ve never used windows 11 for x86. It’s probably also an energy hog.
dagmx · 5h ago
Afaik they are the fastest cores in raw speed. They’re just not available in very high core offerings so eventually fall behind when parallelism wins.
enronmusk · 17h ago
If OP's CPU cooler (Noctua NH-D15 G2) wasn't able to cool down his CPU below 100C, he must have been (intentionally or unintentionally with Asus multi core enhancement) overclocked his CPU. Or he didn't apply thermal paste properly or didn't remove the cooler plastic sticker?
I have followed his blog for years and hold him in high respect so I am surprised he has done that and expected stability at 100C regardless of what Intel claim is okay.
Not to mention that you rapidly hit diminishing returns pass 200W with current gen Intel CPUs, although he mentions caring able idle power usage. Why go from 150W to 300W for a 20% performance increase?
baobabKoodaa · 21h ago
Why is the author showing a chart of room temperatures? CPU temperature is what matters here. Expecting a CPU to be stable at 100C is just asking for problems. Issue probably could have been avoided by making improvements to case airflow.
Jolter · 20h ago
I would expect the CPU to start throttling at high temperatures in order to avoid damage. Allegedly, it never did, and instead died. Do you think that’s acceptable in 2025?
ACCount37 · 20h ago
Thermal throttling originated as a safety feature. The early implementations were basically a "thermal fuse" in function, and cut all power to the system to prevent catastrophic hardware damage. Only later did the more sophisticated versions that do things like "cut down clocks to prevent temps from rising further" appear.
On desktop PCs, thermal throttling is often set up as "just a safety feature" to this very day. Which means: the system does NOT expect to stay at the edge of its thermal limit. I would not trust thermal throttling with keeping a system running safely at a continuous 100C on die.
100C is already a "danger zone", with elevated error rates and faster circuit degradation - and there are only this many thermal sensors a die has. Some under-sensored hotspots may be running a few degrees higher than that. Which may not be enough to kill the die outright - but more than enough to put those hotspots into a "fuck around" zone of increased instability and massively accelerated degradation.
If you're relying on thermal throttling to balance your system's performance, as laptops and smartphones often do, then you seriously need to dial in better temperature thresholds. 100C is way too spicy.
baobabKoodaa · 9h ago
What does room temperature have to do with any of this? Yes, you can lower your CPU temperature by lowering your room temperature. But you can also lower your CPU temperature by a variety of other means; particularly by improving case airflow. CPU temperature is the interesting metric here, not room temperature.
FeepingCreature · 20h ago
No but it's also important to realize that this CPU was running at an insane temperature that should never happen in normal operation. I have a laptop with an undersized fan and if I max out all my cores with full load, I barely cross 80. 100 is mental. It doesn't matter if the manufacturer set the peak temperature wrong, a computer whose cpu reaches 100 degrees celsius is simply built incorrectly.
If nothing else, it very clearly indicates that you can boost your performance significantly by sorting out your cooling because your cpu will be stuck permanently emergency throttling.
izacus · 19h ago
I somehow doubt that, are you looking at the same temperature? I haven't seen a laptop that would have thermal stop under 95 for a long time and any gaming laptop will run at 95 under load for package temps.
FeepingCreature · 19h ago
i7 8550u. Google confirms it stabilizes at 80-85C.
That said, there's a difference between a laptop cpu turbo boosting to 90 for a few minutes and a desktop cpu, which are usually cooler anyway, running at 100 sustained for three hours.
hedora · 2h ago
There’s something very odd with those temps. It’s hitting 100C at > 95% iowait. The CPU core should be idle when that happens.
Maybe the pci bus is eating power, or maybe it’s the drives?
userbinator · 1h ago
Expecting a CPU to be stable at 100C is just asking for problems.
I had an 8th-gen i7 sitting at the thermal limit (~100C) in a laptop for half a decade 24/7 with no problem. As sibling comments have noted, modern CPUs are designed to run "flat-out against the governor".
Voltage-dependent electromigration is the biggest problem and what lead to the failures in Intel CPUs not long ago, perhaps ironically caused by cooling that was "too good" --- the CPU finds that there's still plenty of thermal headroom, so it boosts frequency and accompanying voltage to reach the limit, and went too far with the voltage. If it had hit the thermal limit it would've backed off on the voltage and frequency.
swinglock · 20h ago
The text clearly explains all of this.
baobabKoodaa · 9h ago
No it does not. Which part of the text do you feel explains this?
chmod775 · 4h ago
First off, there's a chart for CPU temperature at the very top and they do talk about it:
> I also double-checked if the CPU temperature of about 100 degrees celsius is too high, but no: [..] Intel specifies a maximum of 110 degrees. So, running at “only” 100 degrees for a few hours should be fine.
Secondly, the article reads:
> Tom’s Hardware recently reported that “Intel Raptor Lake crashes are increasing with rising temperatures in record European heat wave”, which prompted some folks to blame Europe’s general lack of Air Conditioning.
> But in this case, I actually did air-condition the room about half-way through the job (at about 16:00), when I noticed the room was getting hot. Here’s the temperature graph:
> [GRAPH]
> I would say that 25 to 28 degrees celsius are normal temperatures for computers.
So apparently a Tom's Hardware article connected a recent heat wave with crashing computers containing Intel CPUs. They brought that up to rule it out by presenting a graph showing reasonable room temperatures.
I hope this helps.
baobabKoodaa · 35m ago
Hmmh. On a new reading, you're right, the Tom's Hardware reference does justify it. Though I still wouldn't have discussed room temperature at all as it doesn't bring any useful extra information after already monitoring for the CPU temp.
formerly_proven · 20h ago
Strange, laptop CPUs and their thermal solutions are designed in concert to stay at Tjmax when under sustained load and throttle appropriately to maintain maximum temperature (~ power ~ performance).
ACCount37 · 19h ago
And those mobile devices have much more conservative limits, and much more aggressive throttling behavior.
Smartphones have no active cooling and are fully dependent on thermal throttling for survival, but they can start throttling at as low as 50C easily. Laptops with underspecced cooling systems generally try their best to avoid crossing into triple digits - a lot of them max out at 85C to 95C, even under extreme loads.
dezgeg · 17h ago
For handhelds the temperature of the device's case is one factor as well when deciding the thermal limits (so you don't burn the user's hands) - less of a problem on laptops.
perching_aix · 20h ago
> Expecting a CPU to be stable at 100C is just asking for problems.
No. High performance gaming laptops will routinely do this for hours on end for years.
If it can't take it, it shouldn't allow it.
bell-cot · 17h ago
I've not looked at the specifics here - but "stable at X degrees, Y% duty cycle, for Z" years is just another engineering spec.
So, yes - running the CPU that close to its maximum is really not asking for stability, nor longevity.
No reason to doubt your assertion about gaming laptops - but chip binning is a thing, and the manufacturers of those laptops have every reason to pay Intel a premium for CPU's which test to better values of X, Y, and Z.
whyoh · 21h ago
It's crazy how unreliable CPUs have become in the last 5 years or so, both AMD and Intel. And it seems they're all running at their limit from the factory, whereas 10-20 years ago they usually had ample headroom for overclocking.
stavros · 21h ago
That's good, isn't it? I don't want the factory leaving performance on the table.
bell-cot · 21h ago
Depends on your priorities. That "performance on the table" might also be called "engineering safety factor for stability".
makeitdouble · 20h ago
TBF using more conservative energy profiles will bring stability and safety. To that effect in Windows the default profile effectively debuffs the CPU and most people will be fine that way.
therein · 2h ago
So now you're saying just accept the fact that they come pushed past their limits, and the limits are misrepresented. Factory configuration runs them faster than they could in a stable fashion.
That sounds terrible.
stavros · 21h ago
Given that there used to be plenty of room to overclock the cores while still keeping them stable, I think it was more "performance on the table".
formerly_proven · 20h ago
You could also get the idea that vendors sometimes make strange decisions which increase neither performance nor reliability.
For example, various brands of motherboards are / were known to basically blow up AMD CPUs when using AMP/XMP, with the root cause being that they jacked an uncore rail way up. Many people claimed they did this to improve stability, but overclockers now that that rail has a sweet spot for stability and they went way beyond it (so much so that the actual silicon failed and burned a hole in itself with some low-ish probability).
techpression · 19h ago
The 7800X3D is amazing here, runs extremely cool and stable, you can push it far above its defaults and it still won’t get to 80C even with air cooling. Mine was running between 60-70 under load with PBO set to high. Unfortunately it seems its successor is not that great :/
hu3 · 3h ago
Same for 9800X3D here, which is basically the same CPU. Watercooled. Silent. Stupidly fast.
williamDafoe · 18h ago
The 7000 series of CPUs is NOT known for running cool, unlike the AMD 5000 series (which are basically server CPUs repurposed for desktop usage). In the 7000 series, AMD decided to just increase the power of each CPU and that's where most of the performance gains are coming from - but power consumption is 40-50% higher than with similar 5000-series CPUs.
scns · 18h ago
When you use EcoMode with them you only lose ~5% performance, but are still ~30% ahead of the corresponding 5000-series CPU. You can reduce PPT/TDP even further while still ahead.
I specifically singled out the 7800X3D though, it runs incredibly cool and at a very low power draw for the performance you get.
mldbk · 13h ago
> You know, I'm something of a CPU engineer myself :D
Actually almost everything what you wrote is not true, and commenter above already sent you some links.
7800X3D is the GOAT, very power efficient and cool.
Numerlor · 4h ago
The only reason the 7800x3d is power efficient is because it simply can't use much power, and so it runs at a better spot of the efficiency curve. Most of the CPUs won't use more than ~88w without doing manual overclocking (not pbo). Compare that to e.g. a 7600x that's 2 cores fewer on the same architecture and will happily pull over 130w.
And even if could push it higher, they run very hot compared to other CPUs at the same power usage as a combination of AMD's very thick IHS, the compute chiplets being small/power dense and 7000 series X3D cache being on top of the compute chiplet unlike 9000 series that has it on the bottom.
The 9800x3d limited in the same way will be both mildly more power efficient from faster cores and run cooler because of the cache location. The only reason it's hotter is that it's allowed to use significantly more power, usually up to 150w stock, for which you'd have to remove the IHS on the 7800X3D if you didn't want to see magic smoke
mrheosuper · 2h ago
we have unstable "code" generator, so unstable CPU would be natural.
fmajid · 17h ago
I generally prefer AMD Zen5 to Intel due to AVX512 not being gimped by crippled E-cores that really don't belong on a desktop system, SMT (hyperthreading) that actually works and using TSMC processes, but they've also had their issues recently:
Seems like failure in choosing cooling solutions. These high-end chips have obscene cooling needs. My guess would be using something that was not designed for TDP in question.
Sufficient cooler, with sufficient airflow is always needed.
mrheosuper · 2h ago
CPU TDP means nothing now. a 65W tdp cpu can easily consume over 100w during boost.
uniqueuid · 20h ago
For what it's worth, I have an i9-13900K paired with the largest air cooler available at the time (a be quiet! Dark Rock 5 IIRC), and it's incapable of sufficiently cooling that CPU.
The 13900k draws more than 200W initially and thermal throttles after a minute at most, even in an air conditioned room.
I don't think that thermal problems should be pushed to end user to this degree.
michaelt · 18h ago
The "Dark Rock 5" marketing materials say it provides a 210 W TDP [1] and marketers seldom under-sell their products' capabilities.
So if your CPU is drawing "more than 200W" you're pretty much at the limits of your cooler.
Feels like CPU manufacturers should be at least slapping a big warning on if they're selling a CPU that draws more power than any available cooler can dissipate.
anonymars · 1h ago
The idea is that it's for a limited time, after a period of lower-than-that cooling. In other words TDP is time-weighted.
fishtacos · 2h ago
In hindsight, I would have gone for an AMD deskop replacement laptop instead of the Dell Intel-based gaming laptop that I purchased last year. The CPU is the best the Raptor Lake line has to offer in mobile format (i7-13900hx) but there is no conceivable way for the laptop, ast thick as it is, to cool it beyond very bursty workloads.
This affects the laptop with other issues, like severe thermal throttling both in CPU and GPU.
A utility like throttlestop allows me to place maximums on power usage so I don't hit the tjMax during regular use. That is around 65-70W for the CPU - which can burst to 200+W in its allowed "Performance" mode. Absolutely nuts.
SomeoneOnTheWeb · 19h ago
This means your system doesn't have enough airflow if it throttles this quickly.
But I agree this should not be a problem in the first place.
ttyyzz · 20h ago
Agree. Also, use good thermal paste. 100 °C is not safe or sustainable long term. Unfortunately, I think the manufacturer's specifications regarding the maximum temperature are misleading. With proper cooling, however, you'll be well within that limit.
onli · 20h ago
No, those processors clock or shut down if too hot. In no circumstances should they fail because of insufficient cooling. Even without airflow etc.
williamDafoe · 18h ago
A badly optimized CPU will take excessive amounts of power. The "failure in choosing cooling solutions" excuse is just the pot calling the kettle black.
Jnr · 21h ago
I have not had any issues with Intel or AMD CPUs but I have so many issues with AMD APUs, I would steer clear of them. In my experience with different models, they have many graphics issues, broken video transcoding and overall extremely unstable. If you need decent integrated graphics then Intel is the only real option.
sellmesoap · 21h ago
They make a lot of apus for gaming handhelds, I think they do well in that segment. I've had a handful of desktop and laptop apus with no complaints. Even an APU with ecc support, they've all worked without a hitch. I haven't tried transcoding anything on them mind you.
Jnr · 21h ago
Yes, I have Steam Deck and it works great. But I also have 2400G and 5700G and both of those have graphics issues (tested with different recommended RAM sets).
vkazanov · 20h ago
My laptop's AMD is great (Ryzen AI 7 PRO 360 w/ Radeon 880M). Gaming, GPI work, battery, small LLMs - all just work on my Ubuntu.
Don't know about transcoding though.
imiric · 20h ago
I've had the same experience with an 8600G on Linux. Very frequent graphics driver crashes and KDE/Wayland freezes, on old and new kernels alike. I've been submitting error reports for months, and the issues still persist. The RAM passes MemTest, and the system otherwise works fine, but the graphics issues are very annoying. It's not like I'm gaming or doing anything intensive either; it happens during plain desktop usage.
Yet I also use a 7840U in a gaming handheld running Windows, and haven't had any issues there at all. So I think this is related to AMD Linux drivers and/or Wayland. In contrast, my old laptop with an NVIDIA GPU and Xorg has given me zero issues for about a decade now.
So I've decided to just avoid AMD on Linux on my next machine. Intel's upcoming Panther Lake and Nova Lake CPUs seem promising, and their integrated graphics have consistently been improving. I don't think AMD's dominance will continue for much longer.
hedora · 2h ago
Check dmesg after the driver crashes and restarts. If the crash is something about a ringbuffer timeout, use dmidecode to see what the ram is actually clocked at.
Make sure it matches the min of the actual spec of the ram that you bought and what the CPU can do.
I used to get crashes like you are describing on a similar machine. The crashes are in the GPU firmware, making debugging a bit of a crap shoot. If you can run windows with the crashing workload on it, you’ll probably find it crashes the same ways as Linux.
For me, it was a bios bug that underclocked the ram. Memory tests, etc passed.
I suspect there are hard performance deadlines in the GPU stack, and the underclocked memory was causing it to miss them, and assume a hang.
If the ram frequency looks OK, check all the hardware configuration knobs you can think of. Something probably auto-detected wrong.
energy123 · 21h ago
I can't comment on the quality question, but for memory bandwidth sensitive use cases, Intel desktop is superior.
ttyyzz · 20h ago
I'm not convinced, what would be the use case?
energy123 · 18h ago
Data science where you need to keep ~50GB of data in RAM and do intensive things with it (e.g. loop over it repeatedly with numba). You can't get use out of more than 4 cores because memory bandwidth is the only limitation. The data is too big for AMD's cache to be a factor.
Threadripper is built for this. But I am talking about the consumer options if you are on a budget. Intel has significantly more memory bandwidth than AMD in the consumer end. I don't have the numbers on hand, but someone at /r/localllama did a comparison a while ago.
Marsymars · 2h ago
Framework Desktop?
mrlonglong · 21h ago
To make linking go quicker, use mold.
Pass -fuse=mold when building.
positron26 · 21h ago
Do beware when doing WASM with mold. I shipped a broken WASM binary that Firefox could run just fine but Chrome would not.
unsnap_biceps · 12h ago
I also recently swapped to amd and my biggest surprise was how awful their platform is. With intel, getting sensor data just worked without anything special. With amd, it looks like each platform, perhaps model, requires special support. My mobo is a used godlike, and just has zero sensor support.
protocolture · 5h ago
Hmm arent the top end Ryzens dying too? This could be a funny future blog post.
eptcyka · 21h ago
This is rather late, to be quite fair.
discardable_dan · 21h ago
My thoughts exactly: he figured out in 2025 what the rest of us knew in 2022.
positron26 · 21h ago
One of my work computers died and I hadn't checked the CPU market in years. Rode home that night in a taxi with a Ryzen 1700x completely stoked that AMD was back in the game.
If anyone thinks competition isn't good for the market or that also-rans don't have enough of an effect, just take note. Intel is a cautionary tale. I do agree we would have gotten where we are faster with more viable competitors.
M4 is neat. I won't be shocked if x86 finally gives up the ghost as Intel decides playing in Risc V or ARM space is their only hope to get back into an up-cycle. AMD has wanted to do heterogeneous stuff for years. Risc V might be the way.
One thing I'm finding is that compilers are actually leaving a ton on the table for AMD chips, so I think this is an area where AMD and all of the users, from SMEs on down, can benefit tremendously from cooperatively financing the necessary software to make it happen.
jeffbee · 2h ago
The idle power consumption on this guy's rig is completely outrageous. Since almost everything else in my rig is the same, but I use the integrated GPU, I can only conclude that the power floor for GPUs is way too high. Or is it Linux that isn't managing the GPU properly?
rurban · 14h ago
I've burned two of those already, watercooled. The H100 traffic was too much for them
andsoitis · 21h ago
> I would say that 25 to 28 degrees celsius are normal temperatures for computers.
An ideal ambient (room) temperature for running a computer is 15-25 celcius (60-77 Fahrenheit)
And that is an impossibility in most of the world today and it will be even more like that going forward.
nl · 20h ago
Much of the world (for better or worse) uses airconditioning in places they commonly use desktop computers.
trueismywork · 6h ago
No they dont. They don't have the money. I remember my childhood when gaming in summer holidays in India, my PC would run at full tilt because my room was at 36C (and outside was 48C).
em-bee · 19h ago
no they don't. in some countries in europe (maybe in all of them?), installing airconditioning is frowned upon because it is considered a waste of energy. if you want government subsidies for replacing your heating system with a more energy efficient one you are not allowed to have airconditioning. and in the rest of the world only people/countries well of, that don't consider their energy usage, do it. airconditioning is luxury.
using to much airconditioning is also not comfortable. i used to live in singapore. we used to joke that singapore has two seasons: indoors and outdoors. because the airconditioning is powered so high that you had to bring jacket to wear inside. i'd frequently freeze after entering a building. i don't know why they do it, because it doesn't make sense. when i did turn on airconditioning at home i'd go barely below 30. just a few degrees cooler than the outside so it feels more comfortable without making the transition to hard.
qwerpy · 38m ago
> installing airconditioning is frowned upon
Seattle was like this a couple of decades ago when I moved there. People sneered at me when I talked about having air conditioning installed at my house. Having moved from a warmer part of the country, I ignored their smug comments and did it anyway. The next few years I basked in the comfort of my climate-controlled home while my coworkers complained about not being able to sleep due to the heat.
trueismywork · 6h ago
Europe not having some kind of air conditioning (even if heat pump based) is stupidity in my opinion. It lowers productivity
Are they saying this is bad? This Intel CPU has been at it for over a decade. There was a fan issue for half a year and would go up to 80 C for... half a year. Still works perfectly fine but it is outdated, it lacks instruction sets that I need, and it has two cores only, and 1 thread per core.
Maybe today's CPUs would not be able to handle it, I am not sure. One would expect these things to only improve, but seems like this is not the case.
Edit: I misread it, oops! Disregard this comment.
Rohansi · 21h ago
That is your CPU temperature, not ambient (room) temperature.
johnisgood · 21h ago
Oh, I misread. My bad!
imtringued · 20h ago
So you're saying that if you go even 3 degrees Celsius over that temperature range you should expect your CPU to fry itself? Even when the CPU throttled itself to exactly 100°C?
andsoitis · 20h ago
> So you're saying that if you go even 3 degrees Celsius over that temperature range you should expect your CPU to fry itself? Even when the CPU throttled itself to exactly 100°C?
It is actually 2.9999, precisely.
black_puppydog · 18h ago
Layoutparser looks really neat! Glad the author led with this. :D
shmerl · 21h ago
Well, worryingly there were reports of AMD X3D CPUs burning out too. I hope it will be sorted out.
hawshemi · 20h ago
And welcome to USB slow speeds and issues...
amelius · 19h ago
USB issues are driving me nuts. Please, someone show me the path to serenity.
formerly_proven · 20h ago
> Looking at my energy meter statistics, I usually ended up at about 9.x kWh per day for a two-person household, cooking with induction.
> After switching my PC from Intel to AMD, I end up at 10-11 kWh per day.
It's kind of impressive to increase household electricity consumption by 10% by just switching one CPU.
usr1106 · 19h ago
I guess the author runs it at high load for long times, not only for the benchmarks to write this blog post. And less than 10 kWh is a low starting point, many households would be much higher.
don-bright · 18h ago
The amount of power this is using is roughly the same as it takes my car to do my short commute to work
KronisLV · 20h ago
> I also double-checked if the CPU temperature of about 100 degrees celsius is too high, but no: this Tom’s Hardware article shows even higher temperatures, and Intel specifies a maximum of 110 degrees. So, running at “only” 100 degrees for a few hours should be fine.
I'd say that even crashing at max temperatures is still completely unreasonable! You should be able to run at 100C or whatever the max temperature is for a week non-stop if you well damn please. If you can't, then the value has been chosen wrong by the manufacturers. If the CPU can't handle that, the clock rates should just be dialed back accordingly to maintain stability.
It's odd to hear about Core Ultra CPUs failing like that, though - I thought that they were supposed to be more power efficient than the 13th and 14th gen, all while not having their stability issues.
That said, I currently have a Ryzen 7 5800X, OCed with PBO to hit 5 GHz with negative CO offsets per core set. There's also an AIO with two fans and the side panel is off because the case I have is horrible. While gaming the temps usually don't reach past like 82C but Prime95 or anything else that's computationally intensive can make the CPU hit and flatten out at 90C. So odd to have modern desktop class CPUs still bump into thermal limits like that. That's with a pretty decent ambient temperature between 21C to 26C (summer).
williamDafoe · 18h ago
Just FYI Google runs their data centers at 85 degrees F (about 30 degrees C). I think Google probably knows more about how to run Intel CPUs for longest life and lowest cost per CPU cycle. After all they are the #5 computer maker on earth. What Intel is doing and what they are recommending is the act of a desperate corporation incapable of designing energy-efficient CPUs, incapable of progressing their performance in MIPS per Watt of power. This is a sign of a failed corporation.
Panzer04 · 4h ago
Google runs datacenters hot because it's probably cheaper than over-cooling them with AC.
Chips are happy to run at high temperatures, that's not an issue. It's just a tradeoff of expense and performance.
KronisLV · 16h ago
> Just FYI Google runs their data centers at 85 degrees F (about 30 degrees C). I think Google probably knows more about how to run Intel CPUs for longest life and lowest cost per CPU cycle. After all they are the #5 computer maker on earth.
Servers and running things at scale are way different from consumer use cases and the cooling solutions you'll find in the typical desktop tower, esp. considering the average budget and tolerance for noise. Regardless, on a desktop chip, even if you hit tJMax, it shouldn't lead to instability as in the post above, nor should the chips fail.
If they do, then that value was chosen wrong by the manufacturer. The chips should also be clocking back to maintain safe operating temps. Essentially, squeeze out whatever performance is available with a given cooling solution: be it passive (I have some low TDP AM4 chips with passive Alpine radiator blocks), air coolers or AIOs or a custom liquid loop.
> What Intel is doing and what they are recommending is the act of a desperate corporation incapable of designing energy-efficient CPUs, incapable of progressing their performance in MIPS per Watt of power.
I don't disagree with this entirely, but the story is increasingly similar with AMD as well - most consumer chip manufacturers are pushing the chips harder and harder out of the factory, so they can compete on benchmarks. That's why you hear about people limiting the power envelope to 80-90% of stock and dropping close to 10 degrees C in temperatures, similarly you hear about the difficulties of pushing chips all that far past stock in overclocking, because they're already pushed harder than the prior generations.
To sum up: Intel should be less delusional in how far they can push the silicon, take the L and compete against AMD on the pricing, instead of charging an arm and a leg for chips that will burn up. What they were doing with the Arc GPUs compared to the competitors was actually a step in the right direction.
scotty79 · 4h ago
Last Intel desktop CPU that I bought was Pentium 133Mhz. It was also my first PC. Never again ratio of performance to price in my preferred price range favored Intel.
crinkly · 21h ago
I’ve given up both.
Thorrez · 21h ago
What do you use?
aurareturn · 21h ago
I’ve given up on both and use Apple Silicon only. AMD and Intel are simply too power hungry for how slow they are and can’t optimize for power like Apple can.
maciejw · 21h ago
I also switched to cheapest Mac Mini M4 Pro this year (after 20+ years of using Intel CPUs). MacOS has its quirks, but it provides ZSH and it "just works" (unlike manjaro I used in parallel with Windows). I especially like the preview tool - it has useful pdf and photo editing options.
The hardware is impressive - tiny, metal box, always silent, basic speaker built-in and it can be left always on with minimal power consumption.
Drive size for basic models is limited (512gb) - I solved it by moving photos to NAS. I don't use it for gaming, except Hello Kitty Island Adventure. I would say it's a very competitive choice for a desktop PC in 2025 overall.
lostlogin · 21h ago
I just replaced a headless nuc 9 with a headless M4.
Nuc 9 averaged 65-70W power usage, while the m4 is averaging 6.6W.
The Mac is vastly more performant.
mr_windfrog · 21h ago
That's pretty amazing, I've never heard of that before .-_-!
ychompinator · 21h ago
No desktop CPU I’ve ever used has remained stable at 100 degrees.
My 14900k crashes almost immediately at that temp.
3 hours at 100 degrees is obscene.
And any CPU from the last decade will just throttle down if it gets too hot. That's how the entire "Turbo" thing works: go as fast as we can until it gets too hot, after which it throttles down.
swinglock · 20h ago
Then all your desktop CPUs were defective.
Besides AMD CPUs of the early 2000s going up in smokes without working cooling, they all throttle before they become temporarily or permanently unstable. Otherwise they are bad.
I've never had a desktop part fail due to max temperatures, but I don't think I've owned one that advertises nor allows itself to reach or remain at 100c or higher.
If someone sells a CPU that's specified to work at 100 or 110 degrees and it doesn't then it's either defective or fraudulent, no excuses.
clait · 21h ago
Yeah, i can’t believe they think it’s fine.
I would’ve shutdown my PC and rethought my cooling setup the first time it hit 100C tbh
sys_64738 · 16h ago
100C will trigger PROCHOT state in Intel leading to CPU throttling. The CPU will eventually shutdown.
willtemperley · 20h ago
Gaming seems to be the final stronghold of x86 and I imagine that will shrink. Clearly games are able to run well on RISC architectures despite decades of x86 optimisation in game engines. Long term, an architecture that consumes more power and is tightly locked down by licensing looks unsustainable compared to royalty-free RISC alternatives. The instability, presumably because Intel are overclocking their own chips to look OK on benchmarks will not help.
smallpipe · 19h ago
x86 hasn't been CISC in 3 decades anywhere but in the frontend. An architecture doesn't consume power, a design does. I'm all for shitting on intel, but getting the facts right wouldn't hurt.
uncircle · 18h ago
X86 isn’t CISC, sure, but it isn’t a RISC architecture either.
arp242 · 6h ago
Do RISC architectures still exist? ARM has gained tons of stuff and isn't really "RISC" any more either.
Maybe RISC-V? It's right there in the name, but I haven't really looked at it. However, there are no RISC-V chips that have anywhere near the performance x86 or ARM has, so it remains to be seen if RISC-V can be competitive with x86 or ARM for these types of things.
RISC is one of those things that sounds nice and elegant in principle, but works out rather less well in practice.
userbinator · 1h ago
MIPS is as close to a "real RISC" CPU as one can be, and it's "everywhere you don't look", but for reasons entirely unrelated to performance --- it's the choice of SoCs which are too cheap for ARM. I suspect RISC-V is going to become more popular in that market, although it's one which is already filled with various Chinese MIPS/RISC-ish cores that are entirely unimpressive.
immibis · 8h ago
The traditional CISC and RISC division broke down the moment processors started doing more than one thing at a time.
A RISC architecture was actually one with simple control flow and a CISC architecture was one with complex control flow, usually with microcode. This distinction isn't applicable to CPUs past the year 1996 or so, because it doesn't make sense to speak of a CPU having global control flow.
willtemperley · 15h ago
You’re contradicting yourself. The whole reason x86 burns more power is that the CISC front end can’t be avoided.
The CISC decoder is like a "decompressor" that saves memory bandwidth and cache usage.
wqaatwt · 15h ago
Are there any mid-high end RISC-V chips that have comparable performance per watt to x86?
williamDafoe · 18h ago
"Stronghold" is a joke phrase, is it not? Intel had ZERO progress in integrated graphics from 2013-2020. ZERO. That's the reason why "it works so well" - because they NEVER improved the performance or architecture! Sure, they diddled with the number of CU's, but as far as graphics architecture, they never changed it, and it was POOR to begin with (couldn't do 1080p esports very well ...)
nofriend · 4h ago
x86 is the cpu architecture. i don't believe gp was talking about intels igpu solution at all.
willtemperley · 15h ago
Are x86 consoles a joke?
wqaatwt · 15h ago
> Clearly games are able to run well on RISC architectures
Theoretically that’s likely true. But is there any empirical evidence?
Even underclocked Intel desktop chips are massively faster.
willtemperley · 15h ago
Mobile gaming has run well on RISC for a long time and more recently Macs have shown gaming potential.
wqaatwt · 14h ago
Oh sorry. Based on the tone and general fervor in your comment I somehow read it as RISC-V instead of simply RISC (which as other say seems like a mostly meaningless label these days).
Yes, ARM is certainly competitive. But I don’t know how much is that down to Apple being good at making chips instead of the architecture itself.
Qualcomm of course makes decent chips but it’s not like they are that much ahead of x86 on laptops.
Even in Apple’s case, if you only care about raw CPU power instead of performance per watt M series is not that great compared to AMD/Intel.
I've been chasing flimsy but very annoying stability problems (some, of course, due to overclocking during my younger years, when it still had a tangible payoff) enough times on systems I had built that taking this one BIG potential cause out of the equation is worth the few dozens of extra bucks I have to spend on ECC-capable gear many times over.
Trying to validate an ECC-less platform's stability is surprisingly hard, because memtest and friends just aren't very reliably detecting more subtle problems. PRIME95, y-cruncher and linpack (in increasing order of effectiveness) are better than specialzied memory testing software in my experience, but they are not perfect, either.
Most AMD CPUs (but not their APUs with potent iGPUs - there, you will have to buy the "PRO" variants) these days have full support for ECC UDIMMs. If your mainboard vendor also plays ball - annoyingly, only a minority of them enables ECC support in their firmware, so always check for that before buying! - there's not much that can prevent you from having that stability enhancement and reassuring peace of mind.
Quoth DJB (around the very start of this millenium): https://cr.yp.to/hardware/ecc.html :)
Does anyone maintain a list with de-facto support of amd chips and mainboards? That partlist site only shows official support IIRC, so it won't give you any results.
This is the annoying part.
That AMD permits ECC is a truly fantastic situation, but if it's supported by the motherboard is often unlikely and worse: it's not advertised even when it's available.
I have an ASUS PRIME TRX40 PRO and the tech specs say that it can run ECC and non-ECC but not if ECC will be available to the operating system, merely that the DIMMS will work.
It's much more hit and miss in reality than it should be, though this motherboard was a pricey one: one can't use price as a proxy for features.
I would assume your particular motherboard to operate with proper SECDED+-level ECC if you have capable, compatible DIMM, enable ECC mode in the firmware, and boot an OS kernel that can make sense of it all.
Also: DDR5 has some false ecc marketing due to the memory standard having an error correction scheme build in. Don't fall for it.
This was running at like, 1866 or something. It's a pretty barebones 8th gen i3 with a beefier chipset, but ECC still came in clutch. I won't buy hardware for server purposes without it.
It's always the same address, and always a Corrected Error (obviously, otherwise my kernel would panic). However, operating my system's memory at this clock and latency boosts x265 encoding performance (just one of the benchmarks I picked when trying to figure out how to handle this particular tradeoff) by about 12%. That is an improvement I am willing to stomach the extra risk of effectively overclocking the memory module beyond its comformt zone for, given that I can fully mitigate it by virtue of properly working ECC.
Also: Could you not have just bought slightly faste RAM, given the premium for ECC?
And no, as ECC UDIMM for the speed (3600MHz) I run mine at simply does not exist - it is outside of what JEDEC ratified for the DDR4 spec.
DDR4-1600 (PC4-12800)
DDR4-1866 (PC4-14900)
DDR4-2133 (PC4-17000)
DDR4-2400 (PC4-19200)
DDR4-2666 (PC4-21300)
DDR4-2933 (PC4-23466)
DDR4-3200 (PC4-25600) (the highest supported in the DDR4 generation)
What's *NOT* supported are some enthusiast ones that typically require more than 1.2v for example: 3600 MT/s, 4000 MT/s & 4266 MT/s
Also, could you share some relevant info about your processor, mainboard, and UEFI? I see many internet commenters question whether their ECC is working (or ask if a particular setup would work), and far fewer that report a successful ECC consumer desktop build. So it would be nice to know some specific product combinations that really work.
Some vendors use hamming codes with “holes” in them, and you need the CPU to also run ECC (or at least error detection) between ram and the cache hierarchy.
Those things are optional in the spec, because we can’t have nice things.
Edit: it's probably because I switched it to "energy efficiency mode" instead of "performance mode" because it would occasionally lock up in performance mode. Presumably with the same root cause.
I wish AMD would make ECC a properly advertised feature with clear motherboard support. At least DDR5 has some level of ECC.
That is mostly to assist manufacturers in selling marginal chips with a few bad bits scattered around. It's really a step backwards in reliability.
When you do not have a bunch of components ready to swap out it is also really hard to debug these issues. Sometimes it’s something completely different like the PSU. After the last issues, I decided to buy a prebuilt (ThinkStation) with on-site service. The cooling is a bit worse, etc., but if issues come up, I don’t have to spend a lot of time debugging them.
Random other comment: when comparing CPUs, a sad observation was that even a passively cooled M4 is faster than a lot of desktop CPUs (typically single-threaded, sometimes also multi-threaded).
And if we are talking about a passively cooled M4 (MacBook Air basically) it will quite heavily throttle relatively quickly, you lose at the very least 30%.
So, let's not misrepresent things, Apple CPUs are very power efficient but they are not magic, if you hit them hard, they still need good cooling. Plenty of people have had the experience with their M4 Max, discovering that actually, if they did use the laptop as a workstation, it will generate a good amount of fan noise, there is no other way around.
Apple stuff is good because most people actually have bursty workload (especially graphic design, video editing and some audio stuff) but if you hammer it for hours on end, it's not that good and the power efficiency point becomes a bit moot.
I think a lot of it boils down to load profile and power delivery. My 2500VA double conversion UPS seems to have difficulty keeping up with the volatility in load when running that console app. I can tell because its fans ramp up and my lights on the same circuit begin to flicker very perceptibly. It also creates audible PWM noise in the PC which is crazy to me because up til recently I've only ever heard that from a heavily loaded GPU.
For a long time, my Achille's heel was my Bride's vacuum. Her Dyson pulled enough amps that the UPS would start singing and trigger the auto shutdown sequence for the half rack. Took way too long to figure out as I was usually not around when she did it.
But if your UPS (or just the electrical outlet you're plugged into) can't cope - dunno if I'd describe that as cratering your CPU.
I also have this issue.
A common approach is to go into the BIOS/UEFI settings and check that c6 is disabled. To verify and/or temporarily turn c6 off, see https://github.com/r4m0n/ZenStates-Linux
Twice the memory bandwidth, twice the CPU core count... It's really wacky how they've decided to name things
The Ultra is a pair of Max chips. While the core counts didn't increase from M3 to M4 Max, overall performance is in the neighborhood of 5-25% better. Which still puts the M3 Ultra as Apple's top end chip, and the M5 Max might not dethrone it either.
The uplift in IPC and core counts means that my M1 Max MBP has a similar amount of CPU performance as my M3 iPad Air.
On what metric am I ought to buy a CPU these days? Should I care about reviews? I am fine with a middle-end CPU, for what it is worth, and I thought of AMD Ryzen 7 5700 or AMD Ryzen 5 5600GT or anything with a similar price tag. They might even be lower-end by now?
Intel is just bad at the moment and not even worth touching.
The only issues are with an intel Bluetooth chipset, and bios auto detection bugs. Under Linux, the hardware is bug for bug compatible with Windows, and I’m down to zero known issues after doing a bit of hardware debugging.
And it's no bad power quality on mains as someone suggested (it's excellent here) or 'in the air' (whatever that means) if it happens very quickly after buying.
I would guess that a lot of it comes from bad firmware/mainboards, etc. like the recent issue with ASRock mainboards destroying Ryzen 9000-series GPUs: https://www.techspot.com/news/108120-asrock-confirms-ryzen-9... Anyone who uses Linux and has dealt with bad ACPI bugs, etc. knows that a lot of these mainboards probably have crap firmware.
I should also say that I had a Ryzen 3700X and 5900X many years back and two laptops with a Ryzen CPU and they have been awesome.
https://news.ycombinator.com/item?id=45043269
https://www.cpubenchmark.net/cpu_value_alltime.html
CPUs like Intel Core Ultra 7 265K are pretty close to top Ryzens
If your workload is pointer-chasing intel's new CPUs aren't great though, and the X3D chips are possibly a good pick (if the workload fits in cache) which is why they get a lot of hype from reviewers who benchmark games and judge the score 90% based on that performance.
https://youtu.be/OVdmK1UGzGs
https://youtu.be/oAE4NWoyMZk
[1] Well, most non-servers are probably laptops today, but the same reasoning applies.
Definetly not that one if you plan to pair with a dedicated GPU! The 5700X has twice the L3 cache. All Ryzen 5000 with a GPU have only 16MB, 5700 has the GPU deactivated.
Yea, but unfortunately it comes attached to a Mac.
An issue I've encountered often with motherboards, is that they have brain damaged default settings, that run CPU's out of spec. You really have to go through it all with a fine toothed comb and make sure everything is set to conservative stock manufacturer recommended settings. And my stupid MSI board resets everything (every single BIOS setting) to MSI defaults when you upgrade its BIOS.
It looks completely bonkers to me. I overclocked my system to ~95% of what it is able to do with almost default voltages, using bumps of 1-3% over stock, which (AFAIK) is within acceptable tolerances, but it requires hours and hours of tinkering and stability testing.
Most users just set automatic overclocking, have their motherboards push voltages to insane levels, and then act surprised when their CPUs start bugging out within a couple of years.
Shocking!
I'd rather run everything at 90% and get very big power savings and still have pretty stellar performance. I do this with my ThinkStation with Core Ultra 265K now - I set the P-State maximum performance percentage to 90%. Under load it runs almost 20 degrees Celsius cooler. Single core is 8% slower, multicore 4.9%. Well worth the trade-off for me.
(Yes, I know that there are exceptions.)
You can always play with the CPU governor / disable high power states. That should be well-tested.
It turned out during the shitcoin craze and then AI craze that hardcore gamers, aka boomers with a lot of time and retirement money on their hands and early millennials working in big tech building giant-ass man caves, are a sizeable demographic with very deep pockets.
The wide masses however, they gotta live with the scraps that remain after the AI bros and hardcore gamers have had their pick.
https://www.pugetsystems.com/blog/2024/08/02/puget-systems-p...
to;dr: they heavily customize BIOS settings, since many BIOSes run CPUs out-of-spec by default. With these customizations there was not much of a difference in failure rate between AMD and Intel at that point in time (even when including Intel 13th and 14th gen).
Yeah. If Asahi worked on newer Macs and Apple Silicon Macs supported eGPU (yes I know, big ifs), the choice would be simple. I had NixOS on my Mac Studio M1 Ultra for a while and it was pretty glorious.
I had the same issue with my MSI board, next one won't be a MSI.
I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.
I’ve read people here and there arguing that instruction sets don’t matter, that it’s all the same past the decoder anyway. I don’t buy it. The superior energy efficiency of ARM chips is so obvious I find it impossible to believe it’s not due to the ISA since not much else is that different and now they’re often made on the same TSMC fabs.
This isn't really true. On the same process node the difference is negligible. It's just that Intel's process in particular has efficiency problems and Apple buys out the early capacity for TSMC's new process nodes. Then when you compare e.g. the first chips to use 3nm to existing chips which are still using 4 or 5nm, the newer process has somewhat better efficiency. But even then the difference isn't very large.
And the processors made on the same node often make for inconvenient comparisons, e.g. the M4 uses TSMC N3E but the only x86 processor currently using that is Epyc. And then you're obviously not comparing like with like, but as a ballpark estimate, the M4 Pro has a TDP of ~3.2W/core whereas Epyc 9845 is ~2.4W/core. The M4 can mitigate this by having somewhat better performance per core but this is nothing like an unambiguous victory for Apple; it's basically a tie.
> I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.
One of the reasons for this is that Apple has always been willing to run components right up to their temperature spec before turning on the fan. And then even though that's technically in spec, it's right on the line, which is bad for longevity.
In consumer devices it usually doesn't matter because most people rarely put any real load on their machines anyway, but it's something to be aware of if you actually intend to, e.g. there used to be a Mac Mini Server product and then people would put significant load on them and then they would eat the internal hard drives because the fan controller was tuned for acoustics over operating temperature.
This anecdote perfectly describes my few generation old Intel laptop too. The fans turn on maybe once a month. I dont think its as power efficient as an M-series Apple CPU, but total system power is definitely under 10W during normal usage (including screen, wifi, etc).
One of the many reasons why snapdragon windows laptops failed was both amd and Intel (lunar lake) was able to reach the claimed efficiency of those chips. I still think modern x86 can match arm ones in efficiency if someone bothered to tune the os and scheduler for most common activities. M series was based on their phone chips which were designed from the ground up to run on a battery all these years. AMD/Intel just don't see an incentive to do that nor do Microsoft.
There is one exception: If I run an idle Windows 11 ARM edition VM on the mac, then the fans run pretty much all the time. Idle Linux ARM VMs don’t cause this issue on the mac.
I’ve never used windows 11 for x86. It’s probably also an energy hog.
I have followed his blog for years and hold him in high respect so I am surprised he has done that and expected stability at 100C regardless of what Intel claim is okay.
Not to mention that you rapidly hit diminishing returns pass 200W with current gen Intel CPUs, although he mentions caring able idle power usage. Why go from 150W to 300W for a 20% performance increase?
On desktop PCs, thermal throttling is often set up as "just a safety feature" to this very day. Which means: the system does NOT expect to stay at the edge of its thermal limit. I would not trust thermal throttling with keeping a system running safely at a continuous 100C on die.
100C is already a "danger zone", with elevated error rates and faster circuit degradation - and there are only this many thermal sensors a die has. Some under-sensored hotspots may be running a few degrees higher than that. Which may not be enough to kill the die outright - but more than enough to put those hotspots into a "fuck around" zone of increased instability and massively accelerated degradation.
If you're relying on thermal throttling to balance your system's performance, as laptops and smartphones often do, then you seriously need to dial in better temperature thresholds. 100C is way too spicy.
If nothing else, it very clearly indicates that you can boost your performance significantly by sorting out your cooling because your cpu will be stuck permanently emergency throttling.
That said, there's a difference between a laptop cpu turbo boosting to 90 for a few minutes and a desktop cpu, which are usually cooler anyway, running at 100 sustained for three hours.
Maybe the pci bus is eating power, or maybe it’s the drives?
I had an 8th-gen i7 sitting at the thermal limit (~100C) in a laptop for half a decade 24/7 with no problem. As sibling comments have noted, modern CPUs are designed to run "flat-out against the governor".
Voltage-dependent electromigration is the biggest problem and what lead to the failures in Intel CPUs not long ago, perhaps ironically caused by cooling that was "too good" --- the CPU finds that there's still plenty of thermal headroom, so it boosts frequency and accompanying voltage to reach the limit, and went too far with the voltage. If it had hit the thermal limit it would've backed off on the voltage and frequency.
> I also double-checked if the CPU temperature of about 100 degrees celsius is too high, but no: [..] Intel specifies a maximum of 110 degrees. So, running at “only” 100 degrees for a few hours should be fine.
Secondly, the article reads:
> Tom’s Hardware recently reported that “Intel Raptor Lake crashes are increasing with rising temperatures in record European heat wave”, which prompted some folks to blame Europe’s general lack of Air Conditioning.
> But in this case, I actually did air-condition the room about half-way through the job (at about 16:00), when I noticed the room was getting hot. Here’s the temperature graph:
> [GRAPH]
> I would say that 25 to 28 degrees celsius are normal temperatures for computers.
So apparently a Tom's Hardware article connected a recent heat wave with crashing computers containing Intel CPUs. They brought that up to rule it out by presenting a graph showing reasonable room temperatures.
I hope this helps.
Smartphones have no active cooling and are fully dependent on thermal throttling for survival, but they can start throttling at as low as 50C easily. Laptops with underspecced cooling systems generally try their best to avoid crossing into triple digits - a lot of them max out at 85C to 95C, even under extreme loads.
No. High performance gaming laptops will routinely do this for hours on end for years.
If it can't take it, it shouldn't allow it.
Intel's basic 285K spec's - https://www.intel.com/content/www/us/en/products/sku/241060/... - say "Max Operating Temperature 105 °C".
So, yes - running the CPU that close to its maximum is really not asking for stability, nor longevity.
No reason to doubt your assertion about gaming laptops - but chip binning is a thing, and the manufacturers of those laptops have every reason to pay Intel a premium for CPU's which test to better values of X, Y, and Z.
That sounds terrible.
For example, various brands of motherboards are / were known to basically blow up AMD CPUs when using AMP/XMP, with the root cause being that they jacked an uncore rail way up. Many people claimed they did this to improve stability, but overclockers now that that rail has a sweet spot for stability and they went way beyond it (so much so that the actual silicon failed and burned a hole in itself with some low-ish probability).
https://www.computerbase.de/artikel/prozessoren/amd-ryzen-79...
Actually almost everything what you wrote is not true, and commenter above already sent you some links.
7800X3D is the GOAT, very power efficient and cool.
And even if could push it higher, they run very hot compared to other CPUs at the same power usage as a combination of AMD's very thick IHS, the compute chiplets being small/power dense and 7000 series X3D cache being on top of the compute chiplet unlike 9000 series that has it on the bottom.
The 9800x3d limited in the same way will be both mildly more power efficient from faster cores and run cooler because of the cache location. The only reason it's hotter is that it's allowed to use significantly more power, usually up to 150w stock, for which you'd have to remove the IHS on the 7800X3D if you didn't want to see magic smoke
https://www.theregister.com/2025/08/29/amd_ryzen_twice_fails...
Sufficient cooler, with sufficient airflow is always needed.
The 13900k draws more than 200W initially and thermal throttles after a minute at most, even in an air conditioned room.
I don't think that thermal problems should be pushed to end user to this degree.
So if your CPU is drawing "more than 200W" you're pretty much at the limits of your cooler.
[1] https://www.bequiet.com/en/cpucooler/5110
This affects the laptop with other issues, like severe thermal throttling both in CPU and GPU.
A utility like throttlestop allows me to place maximums on power usage so I don't hit the tjMax during regular use. That is around 65-70W for the CPU - which can burst to 200+W in its allowed "Performance" mode. Absolutely nuts.
But I agree this should not be a problem in the first place.
Don't know about transcoding though.
Yet I also use a 7840U in a gaming handheld running Windows, and haven't had any issues there at all. So I think this is related to AMD Linux drivers and/or Wayland. In contrast, my old laptop with an NVIDIA GPU and Xorg has given me zero issues for about a decade now.
So I've decided to just avoid AMD on Linux on my next machine. Intel's upcoming Panther Lake and Nova Lake CPUs seem promising, and their integrated graphics have consistently been improving. I don't think AMD's dominance will continue for much longer.
Make sure it matches the min of the actual spec of the ram that you bought and what the CPU can do.
I used to get crashes like you are describing on a similar machine. The crashes are in the GPU firmware, making debugging a bit of a crap shoot. If you can run windows with the crashing workload on it, you’ll probably find it crashes the same ways as Linux.
For me, it was a bios bug that underclocked the ram. Memory tests, etc passed.
I suspect there are hard performance deadlines in the GPU stack, and the underclocked memory was causing it to miss them, and assume a hang.
If the ram frequency looks OK, check all the hardware configuration knobs you can think of. Something probably auto-detected wrong.
Threadripper is built for this. But I am talking about the consumer options if you are on a budget. Intel has significantly more memory bandwidth than AMD in the consumer end. I don't have the numbers on hand, but someone at /r/localllama did a comparison a while ago.
Pass -fuse=mold when building.
If anyone thinks competition isn't good for the market or that also-rans don't have enough of an effect, just take note. Intel is a cautionary tale. I do agree we would have gotten where we are faster with more viable competitors.
M4 is neat. I won't be shocked if x86 finally gives up the ghost as Intel decides playing in Risc V or ARM space is their only hope to get back into an up-cycle. AMD has wanted to do heterogeneous stuff for years. Risc V might be the way.
One thing I'm finding is that compilers are actually leaving a ton on the table for AMD chips, so I think this is an area where AMD and all of the users, from SMEs on down, can benefit tremendously from cooperatively financing the necessary software to make it happen.
An ideal ambient (room) temperature for running a computer is 15-25 celcius (60-77 Fahrenheit)
Source: https://www.techtarget.com/searchdatacenter/definition/ambie...
using to much airconditioning is also not comfortable. i used to live in singapore. we used to joke that singapore has two seasons: indoors and outdoors. because the airconditioning is powered so high that you had to bring jacket to wear inside. i'd frequently freeze after entering a building. i don't know why they do it, because it doesn't make sense. when i did turn on airconditioning at home i'd go barely below 30. just a few degrees cooler than the outside so it feels more comfortable without making the transition to hard.
Seattle was like this a couple of decades ago when I moved there. People sneered at me when I talked about having air conditioning installed at my house. Having moved from a warmer part of the country, I ignored their smug comments and did it anyway. The next few years I basked in the comfort of my climate-controlled home while my coworkers complained about not being able to sleep due to the heat.
Maybe today's CPUs would not be able to handle it, I am not sure. One would expect these things to only improve, but seems like this is not the case.
Edit: I misread it, oops! Disregard this comment.
It is actually 2.9999, precisely.
> After switching my PC from Intel to AMD, I end up at 10-11 kWh per day.
It's kind of impressive to increase household electricity consumption by 10% by just switching one CPU.
I'd say that even crashing at max temperatures is still completely unreasonable! You should be able to run at 100C or whatever the max temperature is for a week non-stop if you well damn please. If you can't, then the value has been chosen wrong by the manufacturers. If the CPU can't handle that, the clock rates should just be dialed back accordingly to maintain stability.
It's odd to hear about Core Ultra CPUs failing like that, though - I thought that they were supposed to be more power efficient than the 13th and 14th gen, all while not having their stability issues.
That said, I currently have a Ryzen 7 5800X, OCed with PBO to hit 5 GHz with negative CO offsets per core set. There's also an AIO with two fans and the side panel is off because the case I have is horrible. While gaming the temps usually don't reach past like 82C but Prime95 or anything else that's computationally intensive can make the CPU hit and flatten out at 90C. So odd to have modern desktop class CPUs still bump into thermal limits like that. That's with a pretty decent ambient temperature between 21C to 26C (summer).
Chips are happy to run at high temperatures, that's not an issue. It's just a tradeoff of expense and performance.
Servers and running things at scale are way different from consumer use cases and the cooling solutions you'll find in the typical desktop tower, esp. considering the average budget and tolerance for noise. Regardless, on a desktop chip, even if you hit tJMax, it shouldn't lead to instability as in the post above, nor should the chips fail.
If they do, then that value was chosen wrong by the manufacturer. The chips should also be clocking back to maintain safe operating temps. Essentially, squeeze out whatever performance is available with a given cooling solution: be it passive (I have some low TDP AM4 chips with passive Alpine radiator blocks), air coolers or AIOs or a custom liquid loop.
> What Intel is doing and what they are recommending is the act of a desperate corporation incapable of designing energy-efficient CPUs, incapable of progressing their performance in MIPS per Watt of power.
I don't disagree with this entirely, but the story is increasingly similar with AMD as well - most consumer chip manufacturers are pushing the chips harder and harder out of the factory, so they can compete on benchmarks. That's why you hear about people limiting the power envelope to 80-90% of stock and dropping close to 10 degrees C in temperatures, similarly you hear about the difficulties of pushing chips all that far past stock in overclocking, because they're already pushed harder than the prior generations.
To sum up: Intel should be less delusional in how far they can push the silicon, take the L and compete against AMD on the pricing, instead of charging an arm and a leg for chips that will burn up. What they were doing with the Arc GPUs compared to the competitors was actually a step in the right direction.
The hardware is impressive - tiny, metal box, always silent, basic speaker built-in and it can be left always on with minimal power consumption.
Drive size for basic models is limited (512gb) - I solved it by moving photos to NAS. I don't use it for gaming, except Hello Kitty Island Adventure. I would say it's a very competitive choice for a desktop PC in 2025 overall.
Nuc 9 averaged 65-70W power usage, while the m4 is averaging 6.6W.
The Mac is vastly more performant.
Max Operating Temperature: 105 °C
14900k: https://www.intel.com/content/www/us/en/products/sku/236773/...
Max Operating Temperature: 100 °C
Different CPUs, different specs.
And any CPU from the last decade will just throttle down if it gets too hot. That's how the entire "Turbo" thing works: go as fast as we can until it gets too hot, after which it throttles down.
Besides AMD CPUs of the early 2000s going up in smokes without working cooling, they all throttle before they become temporarily or permanently unstable. Otherwise they are bad.
I've never had a desktop part fail due to max temperatures, but I don't think I've owned one that advertises nor allows itself to reach or remain at 100c or higher.
If someone sells a CPU that's specified to work at 100 or 110 degrees and it doesn't then it's either defective or fraudulent, no excuses.
Maybe RISC-V? It's right there in the name, but I haven't really looked at it. However, there are no RISC-V chips that have anywhere near the performance x86 or ARM has, so it remains to be seen if RISC-V can be competitive with x86 or ARM for these types of things.
RISC is one of those things that sounds nice and elegant in principle, but works out rather less well in practice.
A RISC architecture was actually one with simple control flow and a CISC architecture was one with complex control flow, usually with microcode. This distinction isn't applicable to CPUs past the year 1996 or so, because it doesn't make sense to speak of a CPU having global control flow.
https://www.extremetech.com/extreme/188396-the-final-isa-sho...
The CISC decoder is like a "decompressor" that saves memory bandwidth and cache usage.
Theoretically that’s likely true. But is there any empirical evidence?
Even underclocked Intel desktop chips are massively faster.
Yes, ARM is certainly competitive. But I don’t know how much is that down to Apple being good at making chips instead of the architecture itself.
Qualcomm of course makes decent chips but it’s not like they are that much ahead of x86 on laptops.
Even in Apple’s case, if you only care about raw CPU power instead of performance per watt M series is not that great compared to AMD/Intel.