Posted by usefulposter 1 day ago
Only really irritated by the ultra low effort “here is a raw copy paste of what my LLM said on this topic” comments. idk how people think that’s helpful or desired
unless tis signed before uploading, like this is even enforceable?
But I have some concerns about suppression of comments from non-native English writers. More selfishly, my personal writing style has significant overlap with so-called "tells" for AI generated prose: things like "it's not X, it's Y", use of em-dashes, a fairly deep vocabulary, and a tendency toward verbosity (which I'm striving to curb). It'd be ironic if I start getting flagged as a bot, given I don't even use a spell-checker. Time will tell.
And of course, a more limited exception for posts about LLM behavior. It might be necessary for people to share prompts and outputs to discuss the topic.
I wonder if the Chinese might have to say something about that [1]: 33% of 2 million funded studies were in Chinese. I posit that as China strengthens and no longer feels the need to be admired internationally, that declining % will reverse.
Another example is of the Huawei Matebook Fold [2]. It's an interesting dual-screen PC Laptop (?) that I saw in a YouTube video from India, but the product page doesn't even come up in Google search results. Its product page is in Chinese, and the only way to find it seems to be through the wiki page [3].
[1] https://academic.oup.com/rev/article-abstract/doi/10.1093/re...
[2] https://consumer.huawei.com/cn/harmonyos-computer/matebook-f...
Pretending that it's English-native is why there's unspoken incentives to sound more "native", and thus use these grammar-correcting tools.
Some of the intelligent comments on here come from people who learned English in recent months or years, rather than in childhood.
Their English isn't always fluent or well-structured. If they rely slightly more heavily on suggested-next-word tools or AI translations, is that a reason to exclude them from the conversation?
Conversely, many English learning resources for non-native speakers focus on strict formal language, similar to AI-generated text. Do we risk excluding people who have learned a style more formal than we're used to?
The rule just makes the will of the community clear to those who want to respect it.
lol
lmao, even
If I had a nickel for every time I've encountered someone who cared about imperfect language online, I'd have enough nickels to buy Y Combinator.
In my observation, recently there are quite many new AI generated comments in general. Like not even trying to hide with full em-dashes and everything.
I do feel like people are gonna get sneaky in future but there are going to be multiple discussions about that within this thread.
But I find it pretty cool that HN takes a stance about it. HN rules essentially saying Bots need not comment is pretty great imo.
It's a bit of a cat and a mouse problem but so is buying upvotes in places like reddit but HN with its track record of decades might have one or two suspicious or actions but long term, it feels robust. I hope the same robustness applies in this case well hopefully.
Wishing moderation luck that bad actors don't try to take it as a challenge and leave our human community to ourselves :]
Another point I'd like to say is that, if successful, then we can also stop saying, "did you write your comment by LLM" and the remarks as well which I also say time to time when I see someone clearly using AI but it seems that some false-positives happen as well (they have happened sometimes with me and see it happen with others as well) and they also de-rail the discussion. So HN being a place for humans, by humans can fix that issue too.
Knowing dang and tomhow, I feel somewhat optimistic!
Similarly: If you see people making accusations of guidelines violations in a discussion, email the thread link to the mods with a subject like “Accusations in post discussion” and ask them to evaluate them for mod response; they’re always happy to do so and I’m easily clocking in a couple hundred emails a year of that sort to them.
It doesn’t take much to make HN better! And it only takes a moment to point out an overlooked corner of threads for mod review. No need to present a full legal case, just “FYI this seems to violate guideline xyz” is at minimum still helpful.
Even if you believe that prohibiting this is necessary to avoid what one might consider "AI witchhunting", bots are so prevalent now that being expected to communicate the existence of each one via email is unrealistic, for both the reporting users and the moderators. I think it's finally time to consider some sort of on-site report system.
That’s certainly a consequence of how the site operators choose to accept user reports to by mods, yes, but it’s sometimes treated as an excuse not to write the emails to the mods. They can flag off the thread, autocollapse it so it doesn’t take up discussion space for future readers (such as those at work offline for a 3-day IT shift in a secure bunker or whatever), et cetera.
> commenting something like "this is a bot account" is done primarily to inform other users that might not notice
It’s a nice sentiment, but that’s also expressly forbidden by the guidelines/faq (“Please don't post insinuations”, which I’ll suggest to them should be extended to include AI something or other), and I tend to report those accusations as the ‘opening’ guidelines violation so that mods can step in before mobthink kicks in and make their own mod judgment about the matter. A repeated pattern of accusations of guidelines violations in comments is eventually going to attract mod censure, and so I advise against it, no matter how kindly the intent.
> it's finally time to consider some sort of on-site report system
I do agree that it’s clumsy and I make a point of saying that to them about every year or so. Perhaps your email to them about it will be the one that persuades them! I remain ever optimistic.
Language translation is the origin of (the current wave of) AI and its killer app. English is not the main language of the world, and translation opens us up to a huge pool of interesting thinkers.
I'm a native speaker in a foreign language, but out of practice except of a weekly family call. I recently had to write a somewhat technical email to my family, and found it easier to write it in (my more practiced) english and have AI translate it, than write it in the target language myself. Of course, in my case I was able to verify that the output conveyed the meaning I intended, because I am fluent in the target language.
As with the rise of GenAI, I've also noticed a rise of translated messages. It's usually hard to tell the difference, except by looking at the commenter's history (on other subreddits, impossible on HN).
I understand the original frustration with GenAI comments and reactionary response. I'm sorry that we're excluding what could be a large pool of interesting people because we can't tell the difference.
Do we not think that other people want to see words, pictures, software, and videos created by humans too?
One of Dang's comments mentions that he removed some of the other rules because they are already embedded within the HN culture. Other prevailing views exist within the HN culture too. Maybe you just haven't noticed yet.
1. Prevent any account from submitting an actual link until it reaches X months old and Y karma (not just one or the other.)
2. Don't auto-link any URLs from said accounts until both thresholds in #1 are met, so they can't post their sites as clickable links in comments to get around it. Make it un-clickable or even [link removed] but keep the rest of the comment.
3. If an account is aged over X months/years old with 0 activity and starts posting > 2 times in < 24 hrs, flag for manual review. Not saying they're bots, but an MO is to use old/inactive accounts and suddenly start posting from them. I've seen plenty here registered in 2019-2021 and just start posting. Don't ban them right away, but flag for review so they don't post 20 times and then someone finally figures it out and emails hn@.
4. When submitting a comment, check last comment timestamp and compare. Many bots make the mistake of commenting multiple detailed times within sixty seconds or less. If somebody is submitting a comment with 30 words and just submitted a comment 30 seconds ago in an entirely different thread with 300 words, they might be Superman. Obviously a bot.
5. Add a dedicated "[flag bot]" button to users that meet certain requirements so they don't need to email hn@ manually every time. Or enable it to people that have shown they can point out bots to you via email already. Emailing dozens of times a day is going to get very annoying for those that care about the website and want to make sure it doesn't get overrun by bots.
YouTube comment spam has already been doing this for years. Check any video from a reasonably popular creator on any topic related to personal finance; the comments will be full of fake conversations between bots introducing a topic related to the video, and then talking about how such and such a person (whom you can look up by name on Telegram or Signal or whatever) helped solve some serious problem (or invested their money with an implausibly high rate of return). The fake nature of it is usually fairly obvious from the way that the bots make sure you see the name repeated several times with unsolicited, glowing testimonials.
But I had always assumed this was meant to trick actual people, rather than LLMs. Thanks for the food for thought.
At least with link-based SEO “optimization” there's the concrete success criterion of driving traffic to a specific place and put eyeballs on ads.
So with that cleared, this is something that is happening NOW. A couple of years ago, the cutoff date meant that astroturfing like this had a return over months or years. Now with search tools, models can be updated in less than a day with astroturfed comments.
https://simonlermen.substack.com/p/large-scale-online-deanonymization
https://news.ycombinator.com/item?id=47139716I don't think it is a moral failing to use AI to generate writing or to use it to brainstorm ideas and crystalize them, but c'mon isn't it weird to insist that you need them to write _comments_ on the internet? What happens when the AI decides you're wrongthinking?