Asus, Dell, HP and others to produce powerful desktop machines that run AI models locally.
See full article...
See full article...
They left out porn producersbring AI capabilities to developers, researchers, and data scientists who need to prototype, fine-tune, and run large AI models locally.
AI isn't just LLM stuff, there is tons of really great AI applications, for games think AMD FSR, DLSS, and other upscaling techniques"AI Applications" as if anyone has made AI do anything useful yet.
All the CEOs, CIOs, and everyone else running blockchain with 3D monitors, fully cloud invested, IoT focused, PaaS provider, and any other flavor of the year buzzwords, oh and the Virtual Reality diehards. Can't forget those guys. /sI must be really dumb...
But who exactly is the target market for these systems?
Didn't they already mention "developers" and "researchers"?They left out porn producers
On the plus side, while we've seen a lot of less-than useful AI applications so far, this announcement is about hardware to let other people who are interested try to find more of those actual applications. Maybe someone working by themself can spend 3 grand on one of these systems and find an application that the soulless corporations haven't, or didn't think was worth exploring because they couldn't exploit it for as much money."AI Applications" as if anyone has made AI do anything useful yet.
There's a joke to be made here about developers and sweaty balls on presentation...Didn't they already mention "developers" and "researchers"?![]()
Spark is only 256GB/s, so you're getting 5070 level of FP4 performance and fairly slow memory for $3k, not exactly stellar value.Please treat memory bandwidth as the headline statistic for anything that's supposed to be used for "AI".
nVidia have been doing these "superchips" in the datacenter for a few years now. What's new is they're now available in a desktop formfactor. I wonder if this will renew apples interest in the flagging Ultra SoCs - or maybe something even bigger to take advantage of all that cooling and power capacity in the Mac Pro case? Afterall, we were just squealing about 512GB of memory in the Mac Studio just weeks ago, and now nVidia is delivering 768!If I'm reading this right, GB10 is be a SoC, which sounds really ambitious. You would have to think that the size of the higher-speced units dwarf even Apple's Ultra SoCs, which are already impressively large.
I read in other coverage that these run Ubuntu. Just nuke it and deploy the distro of your choice.Ugh, it would have been awesome to get something direct from the company without bloatware.
I can't image these machines coming from a commercial company without a ton of preloaded craziness on it. Hopefully I'm wrong.
Even I know AI has useful applications."AI Applications" as if anyone has made AI do anything useful yet.
Remember, these days AI is just the marketing term for Machine Learning. AI Chat Bots are stupid. Thoughtfully applied machine learning is incredible."AI Applications" as if anyone has made AI do anything useful yet.
Aaaah, that's really cool, so if they can confidently release a ~1000W SoC (and assuming wattage scales in some manner with transistor count, which dictates SoC size) then these should be no problem. That's incredibly impressive, and means that I need some reading to do!nVidia have been doing these "superchips" in the datacenter for a few years now. What's new is they're now available in a desktop formfactor. I wonder if this will renew apples interest in the flagging Ultra SoCs - or maybe something even bigger to take advantage of all that cooling and power capacity in the Mac Pro case? Afterall, we were just squealing about 512GB of memory in the Mac Studio just weeks ago, and now nVidia is delivering 768!
Well either it's unavailable because the yields are shit or it's unavailable because the yields are great and Nvidia wants to sell RTX 6000 PROs. It's the same GPU die with less bits disabled and paired with 96GB of ECC RAM which will probably sell for $10k or so - price not announced, but definitively way fatter margins than a $2k 5090.And you'll be able to buy one as easily as you can a 5090!
It's 20 cores vs. 16, even if 10 of them are lower power cores.It seems to me that Ryzen AI is the direct competitor, not the Mac Studio. The Mac platform is just too different to stand in as a comparison. The biggest question mark for me is whether this heterogenous big-little phone CPU is tolerable on the daily. Lots of people (me) want GPU power for machine learning but also want plenty of power for general crud. I am sure that the 16-core, SMT Zen 5 with AVX512 is going to rip the face off a 10-core Cortex-X when it comes time for the crud work.
If I'm reading this right, GB10 is be a SoC, which sounds really ambitious. You would have to think that the size of the higher-speced units dwarf even Apple's Ultra SoCs, which are already impressively large.
Agreed on pricing and availability being key. I don't think the 10 A725 cores are going to contribute much. A Zen 5 core is about 20% faster that a X925 core, based on Geekbench 6, and 3x faster than a A725. So for purely throughput-oriented computing we're looking at approximately 16 cores vs. the equivalent of 12 cores, at unknown power levels, which I don't care about.It's 20 cores vs. 16, even if 10 of them are lower power cores.
The Radeon 8060S (2560 cores) can address up to 96GB of VRAM and offers 256GB/s of memory bandwidth, backed up by a 32MB Infinity Cache. AMD claims 125 TOPS for its system: https://www.amd.com/en/partner/brow...nsights-articles/ryzen-ai-pro-processors.html
Nvidia is claiming 1000 INT4 TOPS with sparsity enabled. AMD doesn't specify. Nvidia has said that the price of DIGITS "starts" at $3000, but never given any details on what's gated behind the different configuration prices.
Framework has priced their fully configured Strix Halo board at $2000. That's a fair bit cheaper than Nvidia, but Nvidia is also the acknowledged leader in the space via CUDA.
I really like Strix Halo overall, but I don't think this is a slam dunk for AMD. A lot will depend on pricing.
People with a strong nostalgia for Sun Microsystems.I must be really dumb...
But who exactly is the target market for these systems?
Why? Ubuntu is the operating system of choice for businesses. And why is that? Because it provides timely support when needed (because every second the system is down, you lose money) and it works. Time is money.I read in other coverage that these run Ubuntu. Just nuke it and deploy the distro of your choice.
Thanks, I got a great laugh out if thatThey left out porn producers
I can believe that Strix Halo has more CPU oomph than an equivalent ARM system without too much trouble. It's just that the CPU -- even with AVX-512 -- is not an overwhelming contributor to the system's overall AI performance.Agreed on pricing and availability being key. I don't think the 10 A725 cores are going to contribute much. A Zen 5 core is about 20% faster that a X925 core, based on Geekbench 6, and 3x faster than a A725. So for purely throughput-oriented computing we're looking at approximately 16 cores vs. the equivalent of 12 cores, at unknown power levels, which I don't care about.
The much more $$$ GB200 offers 144 cores of Neoverse V2, basically Graviton 4-level performance. That's quite nice.
Me, buying one at Goodwill in 4-8 years to try a few toy projects and then take apart.I must be really dumb...
But who exactly is the target market for these systems?
Plenty of people have, but it's more akin to supercharged ML than actual thinking / intelligence. LLMs are amazing, but they're still just really performant (given their complexity) statistical models, and in situations where we already needed statistical models they're a great and truly brilliant next step."AI Applications" as if anyone has made AI do anything useful yet.
AI is being used increasingly every day in meaningful ways across enterprise. AI for personal use isn't as far along because the cost of hardware is so high it's being used where there's higher ROI. Nobody's investing billions of dollars in their data center to create or provide an inexpensive consumer application of marginal use, or a very expensive consumer application that nobody will buy."AI Applications" as if anyone has made AI do anything useful yet.
Anyone doing research on AI algorithms. Anyone developing software for Grace-Hopper/Blackwell supercomputers. A node with a full Grace-Hopper superchip costs about $40,000. Even if you have the connections, you still have to jump through a number of hoops to get access to and use these systems. This makes the technology a lot more accessible.I must be really dumb...
But who exactly is the target market for these systems?