Unless you're arguing that the rule violations are something the author intends to be part of the meaning of what one wrote?
That's fair.
>Unless you're arguing that the rule violations are something the author intends to be part of the meaning of what one wrote?
I think what I wanted to get at is more like this:
1. I think that they may be part of the meaning
2. I think that people would be primed to accept changes even if they change the meaning
3. I suspected that it would always correct something and wouldn't just say LGTM even if the input was fine
To check, and at the risk of this being hypocritical, I asked for a grammar correction on part of your post that I thought had no mistakes, and both in context and isolation, it corrected "spat out" to "produced." Now, this isn't a huge deal, but it is a loss of the connotation of "spat out," which is the phrasing you chose.
I think grammatical errors are low-cost, and changes in meaning and intent are high-cost, so with 2. above, running it through an LLM risks more loss than it gains.
(On that tangential note, though, I do appreciate that Kagi Translate provides multiple translations and attempts to explain their differences in connotation such that I can pick whichever one most closely matches my intent; if other LLM-assisted writing tools did that then that'd render a lot of this problem moot.)
In all seriousness, if you use some tool to make sure you're using the right "there", noone will mind. Just don't generate another boring predictable comment and everything will be ok