I'm referring to the OpenAI white paper on GPT4 that shows exam results. https://cdn.openai.com/papers/gpt-4.pdf figure 4, and surrounding text.
Clearly not superintelligence, as I would define it (see my other comment about scientific discovery, which I consider the best test), but these are tests actual humans take, and we know how most humans would score on these tests (thru some amount of memorization, along with parsing word problems and doing some amount of calculation). But many people who looked at the testing results concluded that GPT-4 was actually a reasoning agent, or that reasoning agents were just around the corner.
The press picked up on that, and my LinkedIn stream was absolutely filled with second-class influencers who thought that superhuman capabilities were not far away. For a while there, looking at some of the test results specifically on moderately challenging math problems, I suspected that LLMs had some sort of reasoning ability.