Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
undefined | Better HN
0 points
visarga
3y ago
0 comments
Share
Needs a big honking datacenter or billions of compute credits and safety for 6-12 months.
0 comments
default
newest
oldest
tanseydavid
3y ago
Doesn't Alpaca seem to suggest that assumption is no longer true?
visarga
OP
3y ago
Alpaca rides on LLaMA. And LLaMA was trained on 1T tokens for a long time. The fine-tuning takes one hour with low rank adaptation. But pre-training a new model takes months.
j
/
k
navigate · click thread line to collapse