And thus Nvidia keeps the money faucet running. Sorry AI companies, we've just created the latest and greatest. We know you have invested trillions already, but you need the hot new thing or the competition gets it and you are obsolete. Time to invest some more!
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
All three components will be in production this time next year
OK we will see then how good they are.
Pretty awful pictures, hard to see if the monster in the center of the board is a single chip or chiplet design? But it looks like a single chip, so guaranteed to be insanely expensive to make.
Let's hope AMD or others can really break into the compute market with some more competitive products.
It's definitely chiplet, as the article points out. You can see the seams.
Kinda odd. 8 GPUs to a CPU is pretty much standard, and less 'wasteful,' as the CPU ideally shouldn't do much for ML workloads.
Even wasted CPU aside, you generally want 8 GPUs to a pod for inference, so you can batch a model as much a possible without physically going 'outside' the server. It makes me wonder if they just can't put as much PCIe/NVLink on it as AMD can?
LPCAMM is sick though. So is the sheer compactness of this thing; I bet HPC folks will love it.
Yeah, 88/2 is weird as shit. Perhaps the GPUs are especially large? I know NVIDIA has that thing where you can slice up a GPU into smaller units (I can't remember what it's called, it's some fuckass TLA), so maybe they're counting on people doing that.
They could be 'doubled up' under the heatspreader, yeah, so kinda 4x GPUs to a CPU.
And yeah... perhaps they're maintaining CPU 'parity' with 2P EPYC for slicing it up into instances.
14 GB of vRAM?
288gb of vram/gpu
Finally, I get good frame rate in Monster Hunter: Wilds
Sorry, no OpenGL support.
For anyone - including, apparently, this article's author's - who is confused about the form factor, this is the next generation to follow the Grace Blackwell "GB300-NVL72", their full rack interconnected system.
It's the same technology as the matching 8-GPU "HGX" board that is built into larger individual servers - which in this generation's case is just called "B300" as it has the "Blackwell" GPU but not the "Grace" CPU - but not sold in smaller units than an entire rack.
Here are some pictures and a video of that NVL72 version (you can buy these from Dell and others, as well as direct from Nvidia):
https://www.servethehome.com/dell-and-coreweave-show-off-first-nvidia-gb300-nvl72-rack/ https://www.ingrasys.com/solutions/NVIDIA/nvidia_gb300_nvl72/
The full rack has 18 compute trays, each with 2 of the pictured board inside (for a total of 36 CPUs and 72 GPUs), and 9 NVLink switch trays that connect every GPU together. PSUs and networking make up the rest.
What's the connection with Vera Rubin? Is this used to process data from the Vera Rubin observatory?
They've been using scientists names for their stuff for a long time now. Pascal, Turing, etc.