Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> I would not be surprised if it is already here, but has been sidelined by the mods

I can virtually guarantee you that there's been a nontrivial amount of GPT-generated content on HN that has not been caught by mods since ChatGPT, and likely since GPT-2/3, as well. Dang (and the other one whose tag I can't recall) already have their hands full trying to keep the tone civil across thousands (tens of thousands?) of comments a day - it's impossible for them to catch every ML-generated comment (some humans actually do write like these newer language models, after all), and more than likely they're missing a decent number of them - through no fault of their own, it's just an extremely hard problem.

The three solutions that I'm shilling for this problem are (1) invite-trees for HN (like Lobsters, which makes the community much less open but also much more resistant to abuse) (2) webs of trust (not cryptographic, just databases of how much you trust users) overlaid onto HN and other places and (3) people actually reading the content of comments very carefully and upvoting logically sound arguments and downvoting illogical and emotional/manipulative ones, but all of these require a lot of effort and social buy-in.



> people actually reading the content of comments very carefully and upvoting logically sound arguments

You can make a perfectly logical comment that is completely wrong.


Please elaborate.


Logic is about a chain of formal reasoning that is guaranteed to yield a correct result given correct inputs.

If you start with some wrong premises, you can easily write a logical post that has incorrect conclusions.


I am aware of the formal definition of logic. If this is all that tpxl meant, then it's an extremely pedantic and useless objection.


My objection invalidates the solution proposed.


> some humans actually do write like these newer language models, after all

Wouldn't it be more accurate to say these newer language models actually write like humans? Or is there a subset of the population intentionally trying to write in the way that these language models write.


It seems to have a tendency to write stereotypical preamble-statement-conclusion paragraphs and to repeat itself. The model repeats itself often. It repeats the title and then writes a statement that basically repeats the title and after that it provides the useful nugget. At the end it adds a sentence usually using "overall" as an opener. Overall, the model tends to respond in a stereotypical format


Sure, you can also say that some language models write like humans - however, even pre-GPT-2, I read several high-schooler's essays that read very much like these ML-generated products, so even if you don't believe that the relationship is transitive, I think you can say that the relationship is true in both directions.


Another measure would be initiating a bot-check for every user. Maybe even periodically.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: