Top
Best
New

Posted by rbanffy 5 days ago

1.5 TB of VRAM on Mac Studio – RDMA over Thunderbolt 5(www.jeffgeerling.com)
615 points | 227 commentspage 3
delaminator 5 days ago|
> Working with some of these huge models, I can see how AI has some use, especially if it's under my own local control. But it'll be a long time before I put much trust in what I get out of it—I treat it like I do Wikipedia. Maybe good for a jumping-off point, but don't ever let AI replace your ability to think critically!

It is a little sad that they gave someone an uber machine and this was the best he could come up with.

Question answering is interesting but not the most interesting thing one can do, especially with a home rig.

The realm of the possible

Video generation: CogVideoX at full resolution, longer clips

Mochi or Hunyuan Video with extended duration

Image generation at scale:

FLUX batch generation — 50 images simultaneously

Fine-tuning:

Actually train something — show LoRA on a 400B model, or full fine-tuning on a 70B

but I suppose "You have it for the weekend" means chatbot go brrrrr and snark

storus 4 days ago||
M3 Ultra has a crappy GPU, somewhere around 3060Ti-3070. Its only benefit is the memory throughput that makes LLM token generation fast, at around 3080 level. But token prefill that determines time-to-first-token is extremely slow, and coincidentally all those tasks you mentioned above would be around 3060Ti level. That's why Exo coupled DGX Spark (5090 performance for FP4) with MacStudio and sped it up 4x. M5 Ultra is supposed to be as fast as DGX Spark at FP4 due to new neural cores.
benjismith 5 days ago|||
> show LoRA on a 400B model, or full fine-tuning on a 70B

Yeah, that's what I wanted to see too.

theshrike79 5 days ago||
Yea, I don't understand why people use LLMs for "facts". You can get them from Wikipedia or a book.

Use them for something creative, write a short story on spec, generate images.

Or the best option: give it tools and let it actually DO something like "read my message history with my wife, find top 5 gift ideas she might have hinted at and search for options to purchase them" - perfect for a local model, there's no way in hell I'd feed my messages to a public LLM, but the one sitting next to me that I can turn off the second it twitches the wrong way? - sure.

mft_ 4 days ago||
> Yea, I don't understand why people use LLMs for "facts". You can get them from Wikipedia or a book.

Because web search is so broken these days, if you want a clean answer instead of wading through pages of SEO nonsense. It's really common (even) amongst non-techy friends that "I'll ask ChatGPT" has replaced "I'll Google it".

theshrike79 4 days ago||
Kagi or DDG

Google is useless

newsclues 5 days ago||
https://m.youtube.com/watch?v=4l4UWZGxvoc

Seems like the ecosystem is rapidly evolving

A4ET8a8uTh0_v2 5 days ago|
What it kinda reminds me of is PS3 cluster era. Now if I could do something similar to the minisforum..
e28eta 4 days ago||
> For example: did you know there's no way to run a system upgrade (like to 26.2) via SSH

I did not know this. I thought the `softwareupdate` command was built for this use case, and thought it worked over ssh. It sure looks like it should work, but I don’t have a mac I can try it on right now.

pudquick 4 days ago|
He's wrong, it's possible. It's just that root privileges alone is insufficient due to how the signing on LocalPolicy works on M series Macs

https://support.apple.com/guide/security/contents-a-localpol...

The manpage for the command provides information on credential usage on Apple Silicon devices.

dsrtslnd23 5 days ago||
Any thoughts on the GB300 workstation with 768GB RAM (from NVIDA, Asus, Dell, ...)? Although many announcements were made it seems not to be available yet. It does have faster interconnects but will probably be much more expensive.
dogcowmoof 4 days ago||
Wonder if support for RDMA will translate into support for things such as SMB Direct or if it's really only useful for RAM pooling
polsevev 5 days ago||
As much as i hate Apples attitude towards hackers and modifying systems. I have to commend them for building awesome features like this
extraduder_ire 4 days ago||
Is RDMA only going to be on the studio, or is it coming to anything with a thunderbolt 5 port on it?
saddat 4 days ago||
A good part of humanities knowledge under your desk running with a few old light bulbs worth of power
jauntywundrkind 5 days ago|
I really hope AMD or Intel can get on the clue train and respond.

Intel in particular has half a decade of having extremely amazing Thunderbolt ports on their mobile chips, built in (alas not present on desktop chips, for shame). There's been not bad but not great thunderbolt host-to-host networking, that TCP can go over, but the system to system connectivity had been a total afterthought, not at all tuned for obvious smart readily available options like RDMA here. But nothing stops anyone from having better host-to-host protocols.

There are also so many smart good excellent next steps competitors could go for. CXL is showing up on server systems as a much lighter weight much lower latency transport that is PCIe PHY compatible but lighter weight. Adding this to consumer chips and giving even a third of a shit could blow what we see here out of the water. It could probably be done over USB4 & radically blast this bespoke RDMA capability.

Connectivity had been a bespoke special capability for too long. Intel did amazing with Xeon having integrated OmniPath 100Gb a long time ago, that was amazing, for barely any extra bucks. But the market didn't reward them kicking total ass and everyone gave up on connecting chips together. Today we are hostage to fantastically expensive shitty inefficient NIC that cost a crap ton of money to do a worse job, paying enormous penalty for not having the capability on chip, making at best asmedia io hubs do the USB4 dance a hip away from the CPU.

I really hope Intel can appreciate how good they were, see the threat of Apple kicking as here doing what Intel uniquely has been offering for half a decade with incredible Thunderbolt offerings on-chip (limited alas only to mobile chips). I hope AMD feels the heat and gets some god dMned religion and sees the pressure and thread: man they delivered so strong on PCIe lane counts but man they have been so so so slacking on io capabilities for so long, especially on consumer platforms, and Apple is using both their awesome awesome awesome on-chip memory here and their fan-tastic exceptional ability to care just even the tiniest bit about using the consumer interconnect (that already exists in hardware).

I really really really hope someone else other than Apple can ante up and care. There are so many wins to be had, so close. These companies feel so distracted from the plot. Fucking shame. Good on Apple for being the only mofos to a Tually seize the obvious that was just sitting here, they took no effort nor innovation. What a shame no other players are trying at all.

PunchyHamster 5 days ago||
Intel is allergic for making consumer stuff good. Remember how in consumer range like half of the chips had fucking virtualisation disabled, long after competition had it on everything ?
pjmlp 4 days ago||
In the real world, you get a desktop PC with a bunch of GPUs connected on the same bus talking to each other.

No need for multiple computers talking over thunderbolt.

g947o 2 days ago||
probably need to build your own PC. This is not really possible on most prebuilt gaming PCs.
More comments...