4 comments

  • jwpapi 35 minutes ago
    Has someone verified this was an actual bug?

    One of AI’s strengths is definitely exploration, f.e. in finding bugs, but it still has a high false positive rate. Depending on context that matters or it wont.

    Also one has to be aware that there are a lot of bugs that AI won’t find but humans would

    I don’t have the expertise to verify this bug actually happened, but I’m curious.

  • josephg 1 hour ago
    Super interesting. I wish this article wasn’t written by an LLM though. It feels soulless and plastic.
    • gcr 52 minutes ago
      For what it’s worth, Pangram thinks this article is fully human-written: https://www.pangram.com/history/f5f68ce9-70ac-4c2b-b0c3-0ca8...
      • croemer 0 minutes ago
        I don't think Pangram reliably detects individual LLM-generated phrases. It seems to look at sections of ~300 words. And for one section at least it has low confidence.

        It might be tuned for ChatGPT and not work well for Claude Opus 4.6 as well.

      • xmcqdpt2 45 minutes ago
        Then pangram isn't very good, because that article is full of Claude-isms.
        • embedding-shape 11 minutes ago
          > because that article is full of Claude-isms

          Not sure how I feel about the whole "LLMs learned from human texts, so now the people who helped write human texts are suddenly accused of plagiarizing LLMs" thing yet, but seems backwards so far and like a low quality criticism.

        • DiffTheEnder 38 minutes ago
          Is it possible for a tool to know if something is AI written with high confidence at all? LLMs can be tuned/instructed to write in an infinite number of styles.

          Don't understand how these tools exist.

          • gcr 4 minutes ago
            The WikiEDU project has some thoughts on this. They found Pangram good enough to detect LLM usage while teaching editors to make their first Wikipedia edits. https://wikiedu.org/blog/2026/01/29/generative-ai-and-wikipe...

            They found that Pangram suffers from false positives in non-prose contexts like bibliographies, outlines, formatting, etc. The article does not touch on Pangram’s false negatives.

            I personally think it’s an intractable problem, but I do feel pangram gives some useful signal, albeit not reliably.

        • cameronh90 36 minutes ago
          It has Claude-isms, but it doesn't feel very Claude-written to me, at least not entirely.

          What's making it even more difficult to tell now is people who use AI a lot seem to be actively picking up some of its vocab and writing style quirks.

    • rudhdb773b 4 minutes ago
      Not to single out your comment, but it feels like it's gotten to the point where HN could use a rule against complaining about AI generated content.

      It seems like almost every discussion has at least someone complaining about "AI slop" in either the original post or the comments.

    • embedding-shape 1 hour ago
      Any specific sections that stick out? Juxt in the past had really great articles, even before LLMs, and know for a fact they don't lack the expertise or knowledge to write for themselves if they wanted and while I haven't completely read this article yet, I'd surprise me if they just let LLMs write articles for them today.
      • croemer 1 hour ago
        Here's one tell-tale of many: "No alarm, no program light."

        Another one: "Two instructions are missing: [...] Four bytes."

        One more: "The defensive coding hid the problem, but it didn’t eliminate it."

        • monooso 52 minutes ago
          That's just writing. I frequently write like that.

          This insistence that certain stylistics patterns are "tell-tale" signs that an article was written by AI makes no sense, particularly when you consider that whatever stylistic ticks an LLM may possess are a result of it being trained on human writing.

          • oscaracso 9 minutes ago
            I am reminded of the Simpsons episode in which Principal Skinner tries to pass off the hamburgers from a near-by fast food restaurant for an old family recipe, 'steamed hams,' and his guest's probing into the kitchen mishaps is met with increasingly incredible explanations.
          • croemer 42 minutes ago
            These are just some of the good examples I found.

            My hunch that this is substantially LLM-generated is based on more than that.

            In my head it's like a Bayesian classifier, you look at all the sentences and judge whether each is more or less likely to be LLM vs human generated. Then you add prior information like that the author did the research using Claude - which increases the likelihood that they also use Claude for writing.

            Maybe your detector just isn't so sensitive (yet) or maybe I'm wrong but I have pretty high confidence at least 10% of sentences were LLM-generated.

            Yes, the stylistic patterns exist in human speech but RLHF has increased their frequency. Also, LLM writing has a certain monotonicity that human writing often lacks. Which is not surprising: the machine generates more or less the most likely text in an algorithmic manner. Humans don't. They wrote a few sentences, then get a coffee, sleep, write a few more. That creates more variety than an LLM can.

            Fun exercise: https://en.wikipedia.org/wiki/Wikipedia:AI_or_not_quiz

            • monooso 30 minutes ago
              Here's an alternative way of thinking about this...

              Someone probably expended a lot of time and effort planning, thinking about, and writing an interesting article, and then you stroll by and casually accuse them of being a bone idle cheat, with no supporting evidence other than your "sensitive detector" and a bunch of hand-wavy nonsense that adds up to naught.

          • gcr 51 minutes ago
            See also: “I'm Kenyan. I Don't Write Like ChatGPT. ChatGPT Writes Like Me” by Marcus Olang', https://marcusolang.substack.com/p/im-kenyan-i-dont-write-li...

            For what it’s worth, Pangram reports that Marcus’ article is 100% LLM-written: https://www.pangram.com/history/640288b9-e16b-4f76-a730-8000...

            • croemer 37 minutes ago
              In theory, wouldn't be too hard be to settle the question if whether he used ChatGPT to write it: get Olang to write a few paragraphs by hand, then have people judge (blindly) if it's the same style as the article. Which one sounds more like ChatGPT.
              • embedding-shape 5 minutes ago
                The times I've written articles, and those have gone through multiple rounds of reviews (by humans) with countless edits each time, before it ends up being published, I wonder if I'd pass that test in those cases. Initial drafts with my scattered thoughts usually are very different from the published end results, even without involving multiple reviewers and editors.
          • 360MustangScope 43 minutes ago
            I hate that I can’t write em dashes freely anymore without people accusing the writing of being AI generated.

            Even though they are perfect for usage in writing down thoughts and notes.

            • croemer 40 minutes ago
              I have nothing against em dashes. As long as your writing is human, experienced readers will be able to tell it's human. Only less experienced ones will use all or nothing rules. Em dashes just increase the likelihood that the text was LLM generated. They aren't proof.
        • tapoxi 34 minutes ago
          This is my exact writing style - I'm screwed.
          • croemer 25 minutes ago
            I doubt you write like that. Where can I find your writing other than your comments which don't read like that?
        • TruffleLabs 28 minutes ago
          This is just writing; terse maybe and maybe not grammatically correct, but people write like that.
    • ChrisRR 43 minutes ago
      It's not setting off any LLM alarm bells to me. It just reads like any other scientific article, which is very often soulless
    • NiloCK 35 minutes ago
      This is the top reply on a substantial percentage of HN posts now and we should discourage it.

      It is:

      - sneering

      - a shallow dismissal (please address the content)

      - curmudgeonly

      - a tangential annoyance

      All things explicitly discouraged in the site guidelines. [1]

      Downvoting is the tool for items that you think don't belong on the front page. We don't need the same comment on every single article.

      [1] - https://news.ycombinator.com/newsguidelines.html

      • masklinn 6 minutes ago
        > Downvoting is the tool for items that you think don't belong on the front page.

        You can’t downvote submissions. That’s literally not a feature of the site. You can only flag submissions, if you have more that 31 karma.

      • monooso 28 minutes ago
        No idea why you're being downvoted. I've done my bit to redress the balance, I hope others do the same.
    • mpalmer 11 minutes ago
      I've seen way, way worse. Either someone LLM-polished something they already wrote, or they did their own manual editing pass.

      The short sentence construction is the most suspicious, but I actually don't see anything glaring. It normally jumps out and hits me in the face.

    • monooso 37 minutes ago
      You have no evidence that it was.
    • TruffleLabs 30 minutes ago
      "Written by an LLM" based on what data or symptom?
    • ModernMech 1 hour ago
      I'm starting to develop a physiological response when I recognize AI prose. Just like an overwhelming frustration, as if I'm hearing nails on chalkboard silently inside of my head.
      • voodooEntity 56 minutes ago
        I feel ya.... and i have to admit in the past i tried it for one article in my own blog thinking it might help me to express... tho when i read that post now i dont even like it myself its just not my tone.

        therefor decided not gonne use any llm for blogging again and even tho it takes alot more time without (im not a very motivated writer) i prefer to release something that i did rather some llm stuff that i wouldnt read myself.

    • retard2 1 hour ago
      [flagged]
      • vrighter 1 hour ago
        it's actually the second one I read that fit that description.
    • retard3 1 hour ago
      [flagged]
  • wg0 32 minutes ago
    Someone please amend the title and add "using claude code" because that's customary nowadays.
  • yodon 40 minutes ago
    This is so insightfully and powerfully written I had literal chills running down my spine by the end.

    What a horrible world we live in where the author of great writing like this has to sit and be accused of "being AI slop" simply because they use grammar and rhetoric well.

    • dotancohen 25 minutes ago
      I was completely riveted the whole read. The description of Collins' dilemma is the first time I've seen an actual real world scenario described that might cause him to return to Earth alone.

      If an LLM wrote that, then I no longer oppose LLM art.