https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu...
I have similar reservations about code formatters: maybe I just haven't worked with a code base with enough terrible formatting, but I'm sad when programmers loose the little voice they have. Linters: cool; style guidelines: fine. I'm cool with both, but the idea that we need to strip every character of junk DNA from a codebase seems excessive.
For code that is meant to be an expression of programmers, meant to be art, then yes code formatters should be an optional tool in the artist's quiver.
For code that is meant to be functional, one of the business goals is uniformity such that the programmers working on the code can be replaced like cogs, such that there is no individuality or voice. In that regard, yes, code-formatters are good and voice is bad.
Similarly, an artist painting art should be free. An "artist" painting the "BUS" lines on a road should not take liberties, they should make it have the exact proportions and color of all the other "BUS" markings.
You can easily see this in the choices of languages. Haskell and lisp were made to express thought and beauty, and so they allow abstractions and give formatting freedom by default.
Go was made to try and make Googlers as cog-like and replaceable as possible, to minimize programmer voice and crush creativity and soul wherever possible, so formatting is deeply embedded in the language tooling and you're discouraged from building any truly beautiful abstractions.
One factor is "churn", that is, a code change that includes pure style changes in addition to other changes; it's distracting and noisy.
The other is consistency, if you're reading 10 files with 10 different code styles it's more difficult to read it.
But by all means, for your own projects, use your own code style.
And if you are using the tool, “AI” or not to translate it is even worse and you often only have to do on cycle of [your primary language] -> [something else] -> [your primary language] to see what a mess that can make.
I'm attempting to learn Spanish¹ and when I'm writing something, or practising something that I might say, I'll write it entirely away from tech (I have even a proper chunky paper dictionary and grammar guide to help with that!) other than the text editor I'm typing in, and then I'll sometimes give a tool it to look over. If that tool suggests what looks like more than just “that's the wrong tense, you should have an accent there, etc.” I'll research the change rather than accepting it as-is.
--------
[0] or even, potentially, perceived meaning
[1] I like the place and want to spend more time down there when I can, I even like the idea of living there fairly permanently when I no longer have certain responsibilities tying me to the UK², and I'd hate to be ThatGuy™ who rocks up and expects everyone else to speak his language.
[2] and the shithole it has the potential to become over the next decade - to the Reform supporters and their ilk who say, without any hint of irony, “if you don't like it why don't you go somewhere else” I reply “I'm working on that”.
One observation I ran across on the use of the em-dash ("—") was that if AI was given training data from writers that were considered good/great, and those writers tended to use em-dashes, then it would be unsurprising that AI 'learned' to use the character.
So the observer said humans should, if they already did so in the past, continue to use the em-dash now and going forward if it was already part of their 'personal style' in writing.
I'm not planning on writing new books now, but if I did, I would completely get rid of em-dashes, because of their second-order effect of making the copy AI-written (and therefore less valuable).
It's also interesting that using a Skill that discouraged the use of em-dashes, I noticed that Claude's "thinking" internal dialogue actually disagreed with the Skill spec itself ("no, actually, em-dashes are perfectly normal and not a sign of AI writing") and therefore kept the dashes, against the Skill instructions.
When I was young, and learning my technical skills, then naturally I was focused on improving those skills. At that age I defined myself by what I did, and so my self worth was related to my skills. And while the skills are not hard to acquire, not many did, and they were well paid. All of which made me value them even more.
As I've grown older though I discovered my best parts had nothing to do with tech skills. My best parts (work wise) was in translating those skills into a viable business, hiring the right people, focusing my attention where it's needed (and getting out the way where it's not.) My best parts at work are my human relationships with colleagues, customers, prospects and so on.
Outside of work my technical skills mean nothing. My family and friends couldn't care less. They barely know I have drills at all, and no idea if I'm any good or not. In that space compassion, loyalty, reliability, kindness, generosity, helpfulness, positivity, contentment and so on are far (far) more important.
I hope at my funeral people remember those things. Whether I could set up email or drive an AI will (hopefully) not even be in the top 10.
Given you're interacting with a competent hacker (i.e. a person who is into tech not for money and for tinkering), you can't impress them. You can pique their interest, they may praise you, but if they are informed enough, anything looking like magic can be dissected easily. So technical excellence is meaningless.
Given you're interacting with a competent hacker again, everything technical will be subjective. Creating is deciding trade-offs all the way down and beyond. Their preferences will probably lay at a difference balance of trade-offs. Even though you catch "objective" perfection, even this perfection has nuances (see USB audio interfaces. They all have flat response curves, but they all sound different, for example), hence, technical excellence is not only meaningless, it's subjective.
On a deeper level, a genuine person who knows its cookies well, even though with gaps is a much more interesting and nicer person to interact with. They'll be genuinely interested in talking with you, and learn something from you, or show what they know gently, so both parties can grow together. They might not be knowledgeable in most intricate details, but they are genuinely human and open to improvement and into the conversation itself, not to prove themselves and win a meaningless battle to stroke their own ego.
An LLM generated response is similar. It's lazy, it's impersonated, it's like low quality canned food. A new user recently has written an LLM generated rebuttal to one of my comments. It's white-labeled gibberish, insincere word-skirmish. It's so off-putting that I don't see the point to reply them. They'll just paste it to a non-descript box and will add "write a rebuttal reply, press this point". This is not a discussion, this is a meaningless fight for internet points.
I prefer genuine opinions, imperfect replies, vulnerable humans at the other end of the wire. Not a box of numbers spitting out grammatically correct yet empty sentences.
More to the point, Hacker News is much more interesting for encouraging idiosyncratic (i.e. original, diverse, nuanced views of specific) human viewpoints, not just being raw technical information.
Model rewrites remove much of specific human dimension.
Another voice might add citations to every little detail to the point that it is hard to read, but makes a great reference and/or starting point for additional research.
Voice is not really separate from content, in part it is the choices of what content to include.
He has a blog, which I think is particularly relevant to this conversation: https://www.patreon.com/c/GreenWizard/posts?vanity=GreenWiza...
IMO his writing style is quite melodramatic. I have asked myself, how much of that is his perhaps overly compensatory tendency to project an articulate voice, and how much of it is applied by his AI tools?
The last time I saw Anton in person I asked him about his writing process, and he said something like, "I just draft it and then ask ChatGPT to make it sound professional or whatever." So after thinking about it for a while, I have decided that this is his preferred voice, so I'll accept it as his voice.
IMO it is not for you to decide how people recast their own voice. Once you adopt that dogma, you're committed to denying other people's experience of discrimination (through the lens of disability's symptoms). Whether or not you participate in that other type of biased discrimination is irrelevant.
Too often, advocates try to smuggle in their preferred policy using stories like this as cover.
The story itself being true or not doesn't really matter - they're weaponizing an appeal to emotion by using a disabled person as a prop to violate everyone else's standards of interaction.
I think HN is broadly supportive of these voices, and I think that an "unwritten exception" to this rule is implicit here. But I'm in the camp that making an explicit exception for special circumstances would be a meaningful statement that all voices are welcome.
Putting aside the example proposed above where typing or dictation may be difficult, "impossible" seems, well, impossible. I am curious how you suppose that someone who cannot type or dictate at all would prompt an LLM.
I have poor working memory. Very poor, insomuch as I have to type six digit codes in blocks of three.
I can write, of course, and sometimes well. But technical writing requires maintaining both detail and thread and I cannot do both in a sustained way. For a short comment, I'm usually okay. For anything longer, not so much.
Is the long tail the whole beast? I think yes.
So I write shorthand and use tools to help me, and yes the results aren't always perfect -- but they are my thoughts embodied.
Even in this comment, I initially wrote the start as "you're wrong", but then had to catch myself and go back and soften it to "that's incorrect", even though the meaning is the exact same. The constant impedance mismatch is tiring.
It is fact.
Of course - people have egos and emotions, so when they hear someone tell them they are wrong, they will typically take that as criticism about themselves - and not the fact that you are disputing.
I think that would've been pretty clear from the post too, if you weren't so keen on giving a non-native speaker an English lesson ...
So you could use an LLM, privately, to soften people's opinions.
I just tried it for you, I won't copy it here cause the thread is about not using LLMs, but if you get too upset from somebody being simply direct and clear in their manner of speaking, the LLM is trained on enough American cultural baggage that it is very capable of softening that blow with the extra words you so dearly need to see past that red mist.
Someone might even be able to vibe code a browser plugin for it.
When it's a matter of a spelling error or two, no problem. But too often I find I've got to read something multiple times before I have any idea what my interlocutor is saying.
Is our hatred of "AI Slop" and greater posting traffic worth handicapping our ability to communicate with each other?
When I receive an LLM written email at work, I start to question every specific detail because I have no idea if it actually came from the writer (and is therefore important), or was inserted as filler by a computer (and therefore irrelevant).
It wouldn’t be as much of a problem if everyone carefully edited the LLM output themselves before sending (although voice, tone, emotional context clues would still be elided).
But in practice that doesn’t happen, it’s just too easy to click send and the time burden gets passed to the other person.
I get: We found no items matching by:dang "own voice"
https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu...
But that's really what you're now enforcing: writing in an easily detectable LLM prose and voice. LLM detection is very difficult especially at small comment scale texts. There is never proof, only telltale phrases. How will this be enforced? What the heck even is "AI"?
The thing that really frustrates me is that I can't put tokens through a transformer in any way in editing my post? I can't have an LLM turn a bare link after a sentence into a [1]? I can't have it literally do nothing more than spell check in an LLM, but could with a rule based model? Or what about other LLMs or SLMs or classic NLP chained together? Or is it just the transformer?
And it is officially sanctioned that people ought to be keeping in the back of their mind "does this feel LLMish?" instead of "is this a good comment that contributes to the discussion?" Maybe LLM prose is so annoying and insufferably sycophantic that even if all the content and logic was sound, it still should be moderated completely out. But the entire technological form is profane and unclean?
I am 100% not interested in participating in a community that seeks to profile and police the technological infrastructure that its members use. I want my comments judged by the contributions they make and do not make to the discussion. If the LLM makes the comment better, it is good. If it makes it worse, it is bad.
I suppose, then... goodbye?
After all, there are a ton of different forums where you can have your chatbot talk to other chatbots.
That's a good start already. Don't let the impossibility of the perfect prevent implementing the good.
>I want my comments judged by the contributions they make and do not make to the discussion. If the LLM makes the comment better, it is good. If it makes it worse, it is bad.
Nope, it's all bad. If I wanted the comments of an LLM, I'd ask an LLM.
>I am 100% not interested in participating in a community that seeks to profile and police the technological infrastructure that its members use.
Well, don't let the door hit you on your way out.
There used to be a sort of gentleman's agreement that I could spare the time to read and judge your comment because you went through the effort of writing it.
I'd normally not do this for a text of this length, but just for fun, here's what ChatGPT suggests:
As a non-native speaker, I sometimes use LLMs to help me find wording that conveys my thoughts the way I want them to be understood by the reader. I would never copy the output verbatim, because it often sounds blunt and unlike me, but I’m happy to use grammar corrections or improved phrasing.
- Made the prose flatter.
- Slightly changed the sense ('gladly' and 'happy to' are not equivalent, and neither are 'search for' and 'help me find') in ways that do add up
- Not actually improved anything
Introducing "because" also adds to the clarity without weighing down things or changing the meaning. "Improved" instead of the bland "better" again is an... improvement.
I imagine GP didn't sneak in the tendentious "to fit with and be well received in the hacker news community" in his instructions.
Overall this was a worthwhile assist. I believe (totally understandable) anti-AI animus is coloring a lot of these replies. These tools can be useful when applied sparingly and targeted la GP did. It's true and very unfortunate that often they are used as the proverbial hammer in search of a nail, flattening everything in the process.
That, and hindsight bias. People know the second version came from an LLM, so it's automatically "flat." But if that edited comment had just been posted, nobody would've blinked. It reads fine.
IMO, there's a distinction worth drawing here: "AI edited" and "AI generated" are not the same thing. If you write something to express your own thinking, then use an LLM to tighten the phrasing or catch grammar issues, that's just editing. You're still the one with the ideas and the intent. The LLM is a tool, not an author.
The real failure mode is obvious enough: people who dump raw model prose into threads without critical review. The only one who "delved into things" was the model - not the human pressing send. That does flatten everything. But that’s a different case from a non-native speaker using a tool to express their own point more clearly.
The "preserve your voice" argument also smuggles in a premise I don't necessarily share - that everyone should care about preserving their voice. I'm neurodivergent. Being misunderstood when I know I've been clear is one of the most frustrating experiences there is. For some of us, being understood sometimes matters more than sounding like ourselves.
Your comment is one of many on this post that assumes that--because you personally have not noticed a difference--one must not exist. This is not a reasonable assumption.
To take one small example, there is a distinction between 'understood by the reader' and 'received by the reader'. One of them is primarily focused on semantic transmission (did the reader get the message?) and one of them encompasses a wider set of aims (did the reader get the message, and the context, and the connotations, & how did it impact them?).
Every phrasing choice carries precise meanings. There are essentially no perfect synonyms.
In this specific comment, I want you to understand that there are gradations you might not be qualified to detect/comment on. In terms of reception, I'm hoping you will see this as a genuine attempt to communicate, rather than an attack, but I also want you to be aware of the (now voiced) implication that 'I don't see this so it isn't real', no matter how verbose, is a low-effort contribution that doesn't actually add anything.
I'm reminded of Chesterton's fence [1]: if you can't see a reason for something, study it rather than dismissing it.
I disagree with your disagreement and subjective take. The LLM changed the meaning in a significant but not very obvious way.
Compare "I use a hammer to drive nails" to "I use a hammer to help me drive nails"
In the former the writer implies tool use, in the latter the LLM turned that into some sort of assistant relationship. The former is normal, the latter is cringe (to my ears)
> formulate my thoughts like I intend them to be received by the reader
> conveys my thoughts the way I want them to be understood by the reader
there is a way the parent poster constructs their sentences that may sound a little clumsy in a literary sense, but is actually dumbed down
Probably. Planb’s message suggest that the first paragraph is their own writing, the second paragraph tells us that the third paragraph is the llm “improved” version of the first.
To continue the experiment I have fed the above paragraph to Gemini with this prompt "Fix grammar and wording issues in the following paragraphs, if needed reword to fit with and be well received in the hacker news community."
This experiment highlights the core issue. Every language has its own voice—academic, formal, informal, or intimate. Your rewritten paragraph leans into the notorious "LLM voice": it’s less direct, feels slightly pandering, and strips away the hooks that usually spark further discussion.
Does it? I don't see it. If anything, it is more direct and clear, not less, i.e. "to help me find wording that conveys my thoughts the way I want them to be understood by the reader" instead of the more convoluted "to search for a way to formulate my thoughts like I intend them to be received by the reader". How is it pandering? And how exactly does it remove "injection points"?
It basically chose more precise words where that was possible, resulting in a net improvement, AFAICS.
Whereas "search" implies (to me) a kind of direct and analytical process of listing and throwing out brainstormed suggestions, like you would with a search engine.
When I read the human version I actually get a sense of what that process looks like, and the LLM response definitely clouds or changes it by focusing on the result instead.
I have answered something similar before, I struggle on sending messages as I want them to be received, with AI it is even harder, the "taste" of my thoughts, how I like to express, the habits of the phrasing or wording, get lost completely.
So I just never "AI" my content.
- We had to take spelling tests in school
- English speakers make (generally light) fun of other's spelling or grammar mistakes in a casual setting
- In a professional setting, a lot of time is taken to proofread our own emails
- There's de jure spellings for every word
- Some online communities are really weird about pointing out grammar and spelling mistakes (namely Reddit)
Language is meant to be a fluid, evolving thing but I always felt like English was treated the opposite way. Maybe that's also why it's the de facto Lingua Franca.
I do think, and hope, that this rigidity will change thanks to AI. I've started to embrace my mistakes. I care a lot less about capitalization and punctuation in my Slack messages, for example.
I for one don't think I'll ever AI-wash my texts or use AI translations verbatim. If everybody else did, it would certainly be a sad loss of diversity, but IMO it's only going to make the people who put in their own effort stand out more. Hopefully in a positive way. Time will tell if we're a dying breed.
I'm afraid the need for anybody to learn foreign languages will be subject to much change and discussion for upcoming generations.
Must quote the last paragraph of Chapter 2: "Hot and Cold media", from Marshall McLuhan's Understanding Media, which I've double-underlined.
For it simultaneously explains to me; TikTok (quick consume-scroll-like-react-"create" dopamine hit cycles) and LLMs (outsourcing the essential mechanical friction of thinking (which requires all senses, for me at least))...
The essential friction of deliberate, first-party speech-making---misspellings and all---is why voice and conversation contains life.
I don’t think it is so binary black/white though.
I don’t mind if someone who has no command of English uses a translator. But there is a difference between a translator and an AI/LLM.
how hard is it to recognize common idioms and at least state the literal meaning followed by the semantic meaning? there are at most what, a few thousand per language?
Unless they don’t care about learning English which shouldn’t be frowned upon.
Google or Bing translate might not use the exact same words and phrases that LLMs use every single time, so you are better off using those
And LLM does not know context, it makes mistakes a lot more in it. But, it is much cheaper.
I am reminded about a question I posted in a Vintage Apple subreddit. I described the problem and all the steps I took to try and resolve it. In the middle of the text I also hinted that I asked AI and that it gave be a wildly strange answer which I dismissed but that it gave me hints to continue onwards.
The majority of answers were focused around that one sentence and completely ignoring the rest of the post(and even the problem I was posting about). I was ridiculed (sometimes aggressively) for even considering trying the AI. Eventually someone finally answered the question, I thanked them and continued to get downvoted massively.
While I get that the vintage community can attract some colorful characters this was an interesting observation at how badly they reacted to the post. I've since refrained from mentioning AI and furthermore, trying to limit my involvement with communities like that and ironically working on better ways to use AI to solve problems so as to minimize dealing with them(finding ways of providing more system level data to the AI in my prompt).
Also to the people saying that they just let LLM replace phrases: that's the worst you can do. LLM style lies mostly in the phrases, they come from a narrow selection that they tend to use
However, this isn't an entirely new phenomenon. There is a company in Spain called Audens that manufactures croquettes. People prefer hand-made croquettes instead of industrially produced, and they usually can tell the difference by how perfectly regular industrial croquettes are, so Audens developed this method to produce irregular croquettes. Each individual croquette is slightly different, creating a homemade feel that appeals to consumers.
If it's too perfect, it isn't human.