nagaram

joined 2 years ago
[–] nagaram@startrek.website 1 points 2 months ago

From what I understand its not as fast as a consumer Nvdia card but but close.

And you can have much more "Vram" because they do unified memory. I think the max is 75% of total system memory goes to the GPU. So a top spec Mac mini M4 Pro with 48GB of Ram would have 32gb dedicated to GPU/NPU tasks for $2000

Compare that to JUST a 5090 32GB for $2000 MSRP and its pretty compelling.

$200 and its the 64GB model with 2x 4090's amounts of Vram.

Its certainly better than the AMD AI experience and its the best price for getting into AI stuff so says nerds with more money and experience than me.

[–] nagaram@startrek.website 1 points 2 months ago (1 children)

From what I understand its not as fast as a consumer Nvdia card but but close.

And you can have much more "Vram" because they do unified memory. I think the max is 75% of total system memory goes to the GPU. So a top spec Mac mini M4 Pro with 48GB of Ram would have 32gb dedicated to GPU/NPU tasks for $2000

Compare that to JUST a 5090 32GB for $2000 MSRP and its pretty compelling.

$200 and its the 64GB model with 2x 4090's amounts of Vram.

Its certainly better than the AMD AI experience and its the best price for getting into AI stuff so says nerds with more money and experience than me.

[–] nagaram@startrek.website 2 points 2 months ago

Softshell turtle

[–] nagaram@startrek.website 2 points 2 months ago

Honestly if you're not gaming or playing with new hardware, there is absolutely no point.

I've considered swapping this computer over to Fedora for a hot minute, but it really is a gaming PC and I should stop trying to break it.

[–] nagaram@startrek.website 2 points 2 months ago (1 children)

True, but I have an addiction and that's buying stuff to cope with all the drawbacks of late stage capitalism.

I am but a consumer who must be given reasons to consume.

[–] nagaram@startrek.website 2 points 2 months ago

The Lenovo Thinkcentre M715q were $400 total after upgrades. I fortunately had 3 32 GB kits of ram from my work's e-waste bin but if I had to add those it would probably be $550 ish The rack was $120 from 52pi I bought 2 extra 10in shelves for $25 each the Pi cluster rack was also $50 (shit I thought it was $20. Not worth) Patch Panel was $20 There's a UPS that was $80 And the switch was $80

So in total I spent $800 on this set up

To fully replicate from scratch you would need to spend $160 on raspberry pis and probably $20 on cables

So $1000 theoratically

[–] nagaram@startrek.website 1 points 2 months ago (4 children)

The PIs were honestly because I had them.

I think I'd rather use them for something else like robotics or a Birdnet pi.

But the pi rack was like $20 and hilarious.

The objectively correct answer for more compute is more mini PCs though. And I'm really thinking about the Mac Mini option for AI.

[–] nagaram@startrek.website 2 points 2 months ago (1 children)

I've only seen the episode with Toby Turner in it and it has made me a worse person.

[–] nagaram@startrek.website 5 points 2 months ago (2 children)

Ollama and all that runs on it its just the firewall rules and opening it up to my network that's the issue.

I cannot get ufw, iptables, or anything like that running on it. So I usually just ssh into the PC and do a CLI only interaction. Which is mostly fine.

I want to use OpenWebUI so I can feed it notes and books as context, but I need the API which isn't open on my network.

[–] nagaram@startrek.website 8 points 2 months ago (4 children)

The put some damn fan service into Law and Order: SVU and I might start watching!

Am I right fellas!?

[–] nagaram@startrek.website 2 points 2 months ago

I was thinking about that now that I have Mac Minis on the mind. I might even just set a mac mini on top next to the modem.

[–] nagaram@startrek.website 3 points 2 months ago* (last edited 2 months ago)

Ollama + Gemma/Deepseek is a great start. I have only ran AI on my AMD 6600XT and that wasn't great and everything that I know is that AMD is fine for gaming AI tasks these days and not really LLM or Gen AI tasks.

A RTX 3060 12gb is the easiest and best self hosted option in my opinion. New for <$300 and used even less. However, I was running with a Geforce 1660 ti for a while and thats <$100

view more: ‹ prev next ›