Hm. I like your idea! It's one of the few that I think have a chance of working in practice.
In my experience, there is great value in describing your ideas to people who don't have the background to fully understand them, don't really care about them, and are in a bit of a trollish mood - the ways in which they misunderstand what you're saying, or pick on (what you think are) random, irrelevant things, is highly informative. The feedback you get from such people makes you think about things and in ways you wouldn't have thought otherwise.
The problem is, of course, you generally don't have a pool of such people available 24/7. However, LLMs today seem like they could fit this role just fine. They can't and won't understand the point you're trying to get across - but they can and will (if asked) pattern-match on vague language, logical jumps, sentences with multiple meanings, etc. They'll get it subtly wrong, too - much like a bored friend who's mostly focused on playing Angry Birds and hears only every third sentence of your monologue, and then blurts something to keep you talking for the next 30 seconds so they can focus on aiming the birds at the towers of pigs.
I would totally use a LLM-backed tool optimized to facilitate such conversation sessions. I actually tried this in the past, in AI Dungeons, and results were encouraging (i.e. responses got me to think in ways I normally don't).