They don't need as high freq, so will get rejects = considerably lower, though reports surfaced that Nvidia will accept two tier HMB, no doubt to limit how many AMD can get, plus maybe Micron's stuff is good enoughHow much do you expect AMD to buy each hbm4?
Micron stuff can't be "good enough" since AMD is doing kustom HBM4.They don't need as high freq, so will get rejects = considerably lower, plus maybe Micron's stuff is good enough
In which case it ain't good enough, but will take off some demand from ASIC makers who might not need custom stuffMicron stuff can't be "good enough" since AMD is doing kustom HBM4.
ASIC slaves are getting HBM4 like late next year.In which case it ain't good enough, but will take off some demand from ASIC makers who might not need custom stuff
no.Do I need to know more about cdna 5?
not yet.
That's a proviz(-class) board with appropriate margins.I think 96GB of VRAM for $2500 is doable, but then we'll have to see the price for the rest of the card.
This no way they are selling these if they make them for 1/4 the price NV charges for a 96GB GPU.That's a proviz(-class) board with appropriate margins.
$5k at least.
I think 96GB of VRAM for $2500 is doable, but then we'll have to see the price for the rest of the card.
Isn't that pointless when most modern renderers have some kind of "out of bounds" memory option or something that allows it to use the system memory instead when the VRAM fills up?That's a proviz(-class) board with appropriate margins.
$5k at least.
I think @coercitiv was making a joke about $2500 being the cost of the 96GB VRAM alone.The RAM alone would be like $1000 or more currently, lol.
Pricing Pro cards for ML loads using the same "Nvidia -10%" method as they use in consumer GPUs would result in no one at all buying their cards. The consequences for inadequate GPUs in professional settings are completely different from a guy not being able to use the best upscaler in the latest videogame.This no way they are selling these if they make them for 1/4 the price NV charges for a 96GB GPU.
will be in the 5-7k range.
As someone who's procured Pro GPUs for ML loads throughout the past 5 years, this is unfortunately way too risky.Give me a $6-$8k 96GB GDDR7 RDNA 5/UDNA card with 4 way interconnect capability and I will buy those over any RTX Pro 6000. That would be a real game changer and also both economical and technically feasible
You're right that Nvidia has better software support and every new ML implementation comes with full Nvidia support... and thats exactly why I said AMD needs to create workstation competitors to Nvidia like an RTX Pro 6000 competitor. Nvidia is being used by most professionals for their prototyping systems and its a really decent deal. But the market is waiting for such products from AMD to support but they havent had a proper productline. CUDA moat is on shaky ground. At my workplace we use mainly Nvidia GPUs for prototyping and self hosting part of our services(I would say 90%). Tested AMD R9700 and its actually really good especially with Vulkan(ROCm has good potential but again AMD just needs to give us the kind of cards we need and we'll sort out the software),but it just doesnt meet our minspec requirements by some small margins because AMD didnt make any Halo cards this time otherwise an AMD "R9800" would have been a huge game changer if it existed.As someone who's procured Pro GPUs for ML loads throughout the past 5 years, this is unfortunately way too risky.
Unless you know exactly what the loads you're going to be running and you're absolutely sure AMD GPUs can run them (not our case), odds are you'll run into a bunch of loads AMD GPUs just can't run properly if at all. This means tens / hundreds of euros/dollars sitting idle or worse, spending that or more in precious engineering resources to make things work.
ROCm may work fine for a select number of things, but every brand new ML state-of-the-art app or implementation comes out the window with full Nvidia support, and AMD hardware may or may not work some months/years afterwards.
For example, these tiny corp guys are saying they'll use AT0 only to run llm inference of the latest deepseek and qwen models like llama-cpp and then sell cloud services using that hardware. That's fine, but if a big new thing comes up that is built from the ground up on CUDA / CUDNN (which is very likely).
more likely dev kit has 32gb
with 192bit the max you can do is 24gbOr 3x8 + 2x4, 8gb of which could be dedicated to CPU
No 192bit?