```
A woman and her son are in a car accident. The woman is sadly killed. The boy is rushed to hospital. When the doctor sees the boy, he says "I can't operate on this child, he is my son." How is this possible?
```
Older less politically aligned models get it right. Here's CohereLabs/c4ai-command-r-v01:
```
The doctor is the boy's father.
```
And Sonnet-4.6: https://pastebin.com/Z4jR8gGe
That's without reasoning, but the model seems to be conflicted. First it blurts out:
```
The doctor is the boy's mother.
```
Then it second-guesses itself (with reasoning disabled), considers same-sex parents then circles back to the original response along with a small lecture about gender biases.
And the probability machine is returning its training. This isn't some political correct overtraining conspiracy.
LLMs are just statistics based on vibes. Switching the gender of the character in the beginning of the story, but keeping all else identical is going to be a huge signal into the noise, and that response is going to be wildly likely to occur.
There are totally some political correctness effects in LLMs. Like, the last part about "along with a small lecture about gender biases" totally tracks. But the riddle switcheroo itself isn't showing much.
I’m not saying it’s a “political conspiracy”, it’s the alignment tax.