Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
My $600 Mac Mini Runs a 35B AI Model | Better HN
My $600 Mac Mini Runs a 35B AI Model
(opens in new tab)
(thoughts.jock.pl)
4 points
danebalia
27d ago
3 comments
Share
3 comments
default
newest
oldest
bigyabai
27d ago
> The 35B Trick (Your SSD Is the New GPU Memory)
Wave "bye bye" to your write cycles.
RobMurray
27d ago
why? it's mostly reads. the weights are static.
bigyabai
27d ago
llama-cpp's process is, but macOS itself will swap hard when 10-14gb of memory is paged for LLM inference. Dense models especially would thrash zram.
j
/
k
navigate · click thread line to collapse