I'm talking about “alignment” in the broad sense of aligning the actions of one intelligence to the goals of another.
Humans are in general not aligned, not to each other, and not to the survival of their species, not to all the other life on earth, and often not even to themselves individually. When a man is murdered, it is because his desire to live is misaligned with the perpetrator's desire to kill.
>and then they believe everything that is fed to them.
See but here's the thing...They don't.
GPT-3 will ignore tools when it disagrees with them - https://vgel.me/posts/tools-not-needed/
It's not a fundamental issue of gullibility. Reducing gullibility will reduce injection but it's not going to solve it.