Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
undefined | Better HN
0 points
whimsicalism
1y ago
0 comments
Share
They have not. You probably read "MoE" and some pop article about what that means without having any clue.
0 comments
default
newest
oldest
matsemann
1y ago
If you know better it would be nice of you to provide the correct information, and not just refute things.
whimsicalism
OP
1y ago
gpt-4 is a sparse MoE model with ~1.2T params. this is all public knowledge and immediately precludes the two previous commentators assertions
j
/
k
navigate · click thread line to collapse