fwiw I think people can perpetuate the marketing scheme while being genuinely concerned with misaligned superinteligence
We can handwave defining "deception" as "being done intentionally" and carefully carve our way around so that LLMs cannot possibly do what we've defined "deception" to be, but now we need a word to describe what LLMs do do when they pattern match as above.
If the training data gives incentives for the engine to generate outputs that reduce negative reaction by sentiment analysis, this may generate contradictions to existing tokens.
"Want" requires intention and desire. Pattern matching engines have none.