1250
Member
768bit or 1536bit(w/ c-hbm?) 768GBIOD top+bottom size similar zen4 IOD
View attachment 138687
There is talk of 9.6, but it seems like the existing 8533 might be correct.
Hbm lpddr5x same clock? 클램쉘은 말도 안되나
Last edited:
768bit or 1536bit(w/ c-hbm?) 768GBIOD top+bottom size similar zen4 IOD
View attachment 138687
I'll buy it if it's even 30% faster than 5090, with ideally a bit more sane TDP.Additional revenue implies there's a market for $2k gaming Radeons which is just not true.
Too optimistic, but it’s obvious that RDNA5’s flagship card is going to blow past the 5090 in terms of performance.所以如果下一款Radeon能接近甚至超过4090的两倍,那我肯定会买——除非太晚,我很快就会判断N2显卡即将到来,AMD今年就得推出它, 别等Nvidia了。
I mean early 2027 was that 36GB SKU, but if it goes to 2H of 2027, mabye even Q4 I do wonder if they'll change plans like 4GB modules instead of 3, or even make AT0 on N3X? Or maybe save stuff like 4GB modules, N3X until a RDNA 5 refresh launching with Zen 7 since IIRC Lisa Su noted that GPU and CPU sales go hand in hand when new GPUs launch.Unless AT0 is cancelled entirely it will 100% release to consumers, even if it's a token release with no real volume like the Radeon VII.
Yeah, AMD going all with RDNA 5 from a technical perspective. I think AT0 cut down SKU (say 154 old CU/77 new CU) could like 2x the raster performance of the RTX 4090 and at least 2x the Ray and Path Tracing the performance which could IMV convert discontent RTX 40 buyers and an upgrade path for the 7900 XTX gang and maybe even offer a slightly less cut (say 184-188/92-94 but everything else intact) or even full-fat-die-but-not-as-efficient-as-full-fat-pro-SKUs as a Halo tier card. Because speaking of which I think AT0 intended for xCloud may change now:Too optimistic, but it’s obvious that RDNA5’s flagship card is going to blow past the 5090 in terms of performance.
Six months after Bond’s promotion, Xbox executive Kareem Choudhry, who reported directly to Bond, departed Microsoft and triggered another shake-up of some teams inside Xbox. Choudhry was key to Xbox’s backward compatibility support and helped bring Xbox Cloud Gaming to life as xCloud in 2019.
Just weeks after Choudhry departed, former Xbox chief marketing officer Jerret West also left Microsoft in June 2024. That meant the Xbox marketing team was now reporting directly to Bond. A month later, Microsoft delivered a marketing campaign that signalled people didn’t need to buy an Xbox console anymore. The message was that “you don’t need an Xbox to play Xbox,” because games were available through Xbox Cloud Gaming on TVs.
I understand that Bond’s strategy had been failing internally and been questioned multiple times. Bond had tried to push mobile and cloud over console, to reach potentially millions more Xbox customers, but the result has been a classic case of chasing tomorrow’s customers by neglecting today’s.
According to the Korean article, AMD's qual is finished.(samsung) Does this mean S-HBM that will go into the 440x or 430x? Does it include C-HBM?This is just a delusion.(mi455x meh)
Imho it would be a waste, most scientific computing workloads don't need much VRAM, and even "big" bio/chem/material ML models are far smaller than current LLMs.For MI430X I am wondering, if an additional LPDDR5X memory pool (and rack scale as well) wouldn't be helpful as well?
HPC workloads don't need that much memoryHmm good point. MI440X cannot feature LPDDR5X the same way because of OAM form factor. So S-HBM could be sufficient there.
For MI430X I am wondering, if an additional LPDDR5X memory pool (and rack scale as well) wouldn't be helpful as well?
Exceeding Blackwell HW capabilities should be easy. On NVIDIA side the RT foundation unchanged since Turing (spamming ray/tri + adding low hanging fruits), SM foundation and cachemem since Turing when Ampere is just concurrent execution, beefed up L1 and 2X FP32. Lovelace is Ampere+ with ML and RT low hanging fruits, Blackwell = Ampere++.From what infos, leaks and rumors are suggesting, RDNA5's HW capabilites look on par or even above Blackwell.
Hopefully much more than just that. Universal compression = mem BW multiplier, DGF is just better BLAS compression. But Work Graphs HW accel is interesting.With probably two or three especially advanced HW accelerated things (DGF, work graphs, universal compression).
Larrabee-esque?it's far far far far far weirder than that.
far weirder than anything NV will ever ship.
My memory is failing me so has this OoO thing been confirmed yet?Out-of-order scheduling on GPUs is special, for sure.
Sounds like genAI nonsense. IIRC MLID gave +5-10% IPC figure for RDNA5.Never heard of such specific numbers about RDNA5. Leak or just speculation?
When did this happen?Gaming AT0 was hatcheted off the roadmap.
You said 24CU AT0 dGPU = 9060XT should be doable and as I understood it much expanded dual issue use plays a large part here:Yeah that's a bit too optimistic
AT2 is 4X more Shader Engine + SE tied logic (4 SEs vs 1) AND 2.9X CUs (70 vs 24). With reduced clocks matching 4090 raster perf doesn't sound outlandish and is a ~35% raster gain vs 9070XT.Yeah 9060 XT perf at desktop power levels should be possible.
Hopefully they learned their lesson with RDNA3.Yes, if there are issues with clock rates and power efficiency (RDNA3 ahoi) or CU count scaling (5090 ahoi), then it will not happen.
Easily doable if the revamped scheduling works as detailed in the patents.If RDNA5 is 50% quicker than 5090 (which will be then close to double of 4090) then I'll be very interested, as long as ray tracing does not totally suck. I find it hard to believe it will be though, maybe if it was on N2 but it ain't.
That shouldn't be hard to accomplish. N3P, cachemem + scheduling clean slate, and CU spam = 5090 obliterator guaranteed.I'll buy it if it's even 30% faster than 5090, with ideally a bit more sane TDP.
Sounds reasonable. 2X 4090 is ~50% ahead of 5090.I think AT0 cut down SKU (say 154 old CU/77 new CU) could like 2x the raster performance of the RTX 4090 and at least 2x the Ray and Path Tracing
But isn't that due to inherent bad cacheability of RT stuff? It's just too random, at least for puny cachesConsidering RT cachemem efficiency of 50 series is atrocious
Apparently Dylan Patel partied too hard with the hockey team and forgot to close TwitterView attachment 138743
Helios racks are not delayed.
Defect screening is possible before customer post-processing (CoW/SiP). This appears to be a feature added by Samsung. Does JEDEC have something similar? (I hope it applies to AMD, too.) Maybe start 4e?AMD and Intel rule JEDEC with an iron fist.
No the NVIDIA design is just very inefficient. Does thread coherency sorting (SER) and OMM but other than that not really any major changes since Turing.But isn't that due to inherent bad cacheability of RT stuff? It's just too random, at least for puny caches
You're thinking with the mentality of the market from 10, 15 years ago. Back then the high end "Pro" cards (like Quadro, FirePro or Tesla) were niche things that sold into a very limited market. AMD/Nvidia would sell as many of them as they could (because high margins), but it wasn't a big market. So they needed the volume of gaming to make up the numbers and make the R&D worthwhile.So they have the silicon developed, in manufacturing, but deliberately don't try to get additional revenue for it to help recoup the design cost.
I mean, I could see it turning into an Arc B770 situation (where it likely would have happened but memory prices doomed it) in 2027. But it's not H2 2027 today.
About you knowing what will happen for sure 2 years ahead, I have my doubts.
This is just racist or ignorant.Apparently Dylan Patel partied too hard with the hockey team and forgot to close Twitter
APUs.You're thinking with the mentality of the market from 10, 15 years ago. Back then the high end "Pro" cards (like Quadro, FirePro or Tesla) were niche things that sold into a very limited market. AMD/Nvidia would sell as many of them as they could (because high margins), but it wasn't a big market. So they needed the volume of gaming to make up the numbers and make the R&D worthwhile.
Unfortunately the AI bubble is not like that. AMD and Nvidia are selling every single AI chip they can manufacture (or rather selling out the entire capacity years ahead of time), and the only factor stopping them selling even more is fab throughput. They don't need gaming, when capitalism appears happy to pour infinite money into the black hole of AI capex.
We'll be lucky if we can even buy GPUs by next Christmas, if this AI bubble doesn't pop soon.