Unfortunately (a) is more common, and the backlash against has been removing the communinity incentive to provide (b).
But the "This is what ChatGPT said..." stuff feels almost like "Well I put it into a calculator and it said X." We can all trivially do that, so it really doesn't add anything to the conversation. And we never see the prompting, so any mistakes made in the prompting approach are hidden.
If we want human "on the other end" we gotta get to ground truth. We're fighting a losing battle thinking that text-based forums can survive without some additional identity components.
You don't possess an AI, you are using someone's AI
I'm reasonably sure the instance of Olmo 3.1 running locally on this very machine via ollama/Alpaca is very much in my possession, and not someone else's.
No? Then it's not "your" AI, it's an AI that you are using.
I didn't hand-wire the transistors and hand-write the software that constitute the computer on which I'm writing this comment, but said computer is rather unambiguously mine and mine alone. Why would the local copy of the LLM on that same computer be any different? Does my coffee mug cease to be mine if someone else happens to have an identical one?
An alternative I tried was sharing links my LLM prompts/responses. That failed badly.
I like the parallel with linking to a Google/DuckDuckGo search term which is useful when done judiciously.
Creating a good prompt takes intelligence, just as crafting good search keywords does (+operators).
I felt that the resulting downvotes reflected an antipathy towards LLMs and the lack of taste of using an LLM.
The problem was that the messengers got shot (me and the LLM), even though the message of obscure facts was useful and interesting.
I've now noticed that the links to the published LLM results have rotted. It isn't a permanent record of the prompt or the response. Disclaimer: I avoid using AI, except for smarter search.