Skip to content
Better HN
Top
New
Best
Ask
Show
Jobs
Search
⌘K
undefined | Better HN
0 points
riku_iki
2y ago
0 comments
Share
absolutely not. Transformer layers already communicate using embeddings, and ASCII would be absolutely less efficient there.
0 comments
default
newest
oldest
Rhapso
2y ago
And how many bits are in an embedded vector?
riku_iki
OP
2y ago
12k for gpt3.
riku_iki
OP
2y ago
It is not bits, but weights
Rhapso
2y ago
So somehow ascii is less information dense than 12k 32-bit floats per token?
j
/
k
navigate · click thread line to collapse