Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
undefined | Better HN
0 points
throwthrowuknow
1y ago
0 comments
Share
Not going to work for training from scratch which is what the author is doing.
0 comments
default
newest
oldest
rspoerri
1y ago
192GByte of RAM are not enough to train 405B models. Reflection 70B requires 140GByte of RAM in fp16, 405 would need ~810Gbyte of RAM.
throwthrowuknow
OP
1y ago
Pretty sure he said he’s inferencing llama3 405 and training his own custom model from scratch. He didn’t say how big his custom model will be.
j
/
k
navigate · click thread line to collapse