points by mtlynch 1 day ago

This is blowing my mind.

I asked Kimi K2.6 to write a blog post in the style of James Mickens.[0] Then I fed the output to Opus 4.7 and asked it who the likely author was, and it correctly identified it as an imitation of James Mickens[1]:

> Based on the stylistic fingerprints in this text, the most likely author is a pastiche/imitation of the style of several writers fused together, but if forced to identify a single likely author, the strongest candidate is someone writing in the voice of James Mickens

> [...]

> The piece could also be a deliberate imitation/homage to Mickens written by someone else, or AI-generated text trained on his style, since the voice is so distinctive it's frequently parodied.

[0] https://kagi.com/assistant/5bfc5da9-cbfc-4051-8627-d0e9c0615...

[1] https://kagi.com/assistant/fd3eca94-45de-4a53-8604-fcc568dc5...

saghm 1 day ago

> it correctly identified it as an imitation of James Mickens

How likely is it that it might take into account that it knows for sure it's not anything from Mickens from the latest training data? I'd be curious if it correctly identified a new piece from him that comes out as from him before it gets trained on it.

  • suriya-ganesh 1 day ago

    This is unlikely. The way model distribution works is that the model retains a lossy representation of James Micken's writing. Very likely, it cannot repeat Micken's writing verbatim. Neither can it reason about the training cutoff in this manner.

    It's a lossy representation

    • koiueo 1 day ago

      How do you know, how the model works? If there was an index of all Micken's writings, or even if the model searched the web before feeding the response to you, you wouldn't know by observing from the outside.

      • suriya-ganesh 1 day ago

        i suppose a quick test would be getting the model to write down Micken's essay end to end.

        if the original essay was stuffed within the prompt window. the result will be word accurate.

        unless this is a model trained specifically on Micken's essay (which claude is not).

        • saghm 23 hours ago

          This seems like a classic case of doing it being proof that it can happen, but not doing it being insufficient proof that it's impossible. I don't think there's a "quick test" of whether there might be a more effective prompt that would cause it to reproduce more effectively.

        • repparw 22 hours ago

          Didn't we get this with Harry Potter back in like gpt3.5? I'm sure I saw some news about it, someone getting it to output a book's intro word by word, couple pages?

    • sausagefeet 1 day ago

      I haven't been following it well but isn't part of the NYT lawsuit against OpenAI that it sometimes spits out NYT articles verbatim?

      • Iulioh 1 day ago

        Study: Meta AI model can reproduce almost half of Harry Potter book

        https://arstechnica.com/features/2025/06/study-metas-llama-3...

        • kbelder 20 hours ago

          "The study authors took 36 books and divided each of them into overlapping 100-token passages. Using the first 50 tokens as a prompt, they calculated the probability that the next 50 tokens would be identical to the original passage. They counted a passage as “memorized” if the model had a greater than 50 percent chance of reproducing it word for word."

          So they fed "It takes a great deal of bravery to stand up to our " and the llm responded "enemies, but just as much to stand up to our friends".

          They repeated that for every 100 tokens of the entire book. I think lots of fans could do just as well. It's pretty good evidence that the potter books were in the training corpus, but it's not quite what people think when they say an llm has 'memorized' something. It's not like getting even a few pages out of the model.

      • mapt 23 hours ago

        Genome analysis is also a lossy process that chops the data up into tiny bits, like a newspaper sent through a shredder. We then piece together matching sequences in a sort of puzzle. It's often a relatively inaccurate solution. Then we try to do that again with a different copy of the newspaper sent through a different shredder. And again. A genome might be comprised of 10x reads, 30x reads, 100x reads, with more replications representing higher confidence.

        There might be ten million people who have quoted Harry Potter at some point in their blogs or forum posts. There are only so many words in the books.

      • algoth1 20 hours ago

        That issue is different, when web tools were added to gpt4o it would fetch the site, and basically copy paste the text into the answer body. So, you were able to read the content of the site without the site getting the ad impressions. Now the system prompts put a very tight word limit - 25? - on quotes from sites the model visits

    • devmor 1 day ago

      Haven’t there been repeated experiments that show if you jailbreak most frontier models’ harnesses you can get them to output near verbatim copyrighted works?

      I swear there was a whole court case about this in the last year.

    • electroglyph 1 day ago

      that's in the ideal scenario where it's only seen a single copy of it tho

    • sigmoid10 1 day ago

      It is lossy, but it is still enough for verbatim recreations. All of Wikipedia is just 24GB of lossless compressed text and all of JK Rowling's work fits into a few MB. So these things would easily be storable verbatim in trillion parameter models. Reasoning about the training cutoff is also something that the newest models do pretty well, because you can teach them to do so after pre training using e.g. SFT. With tool use it can then even check actual current sources, which may happen without you even knowing in the normal chat apps unless you use a controlled API call.

    • saghm 23 hours ago

      I feel like you're making a logical leap here by assuming lossy and failure to reproduce in entirety implies inability to recognize. As a trivial example, I can take a sha256 hash of your comment here, lose the ability to reproduce it, but still have an extremely accurate ability to recognize whether some text is exactly your comment or not. Obviously hashing every substring would not be a particularly efficient strategy, but my point is that saying "it's lossy" isn't particularly compelling without other details.

      • gspetr 19 hours ago

        The example you've provided just adds noise.

        sha256 is deterministic, LLMs are not, even at temperature set to 0.

  • jerf 23 hours ago

    I fed an unpublished draft of mine to an AI. I saw it searching the internet and prompted it with the fact it could stop searching, it was not published. From there it guessed that it was me on the spot, which I thought was kind of funny. Can't deny the meta-logic there.

    It referred to me by my login name on the AI site rather than the name it would have used if it actually found my website, so I think it was more logic than an actual identification, but it had clearly corrupted the search enough to no longer be a valid test.

    Which does make me wonder about the original article; if the AI has in context any sort of clue that the user is "Kelsey Piper" (a memory of their name, a username of kpiper or kelseyp, etc.), that will radically tip the balance in favor of the AI guessing that way just by the nature of LLMs. That is to say, it highly increases the odds of that guess even if it's wrong.

    Even if that is the case, though, the general identifiability of writing remains true. It's been shown for a while with techniques a lot less powerful than a frontier LLM.

    • cg505 22 hours ago

      The author specifically discusses their efforts to avoid this sort of information leak which would obviously poison the result.

    • furyofantares 9 hours ago

      She says she used incognito mode, as well as the API, as well as having a friend use their account.

jefftk 1 day ago

That's neat, though it impresses me less that the article. Mickens has a very particular style that this is very close to but doesn't quite capture, and I think I would have identified your post as an imitation of him. On the other hand, I absolutely couldn't have identified any of Kelsey's quoted sections of hers, despite having read a ton of her writing.

  • phrotoma 23 hours ago

    It is very close, but what's more interesting to me is that it's actually amusing. I've yet to see an LLM actually be originally funny (entirely possible I've missed the crossing of that line) and the opening lines put a wry grin on my face.

mavelikara 15 hours ago

A newspaper ran a contest to write prose in the style of Graham Greene. Greene sent in the opening two paragraphs of an unfinished work. He came in _second_ in the contest. Many years later, Greene sent in an entry to a similar contest. This time he didn’t win any prizes but got an honorable mention from the judges.

flashdesk 23 hours ago

The part that stands out is that it identified the text as an imitation rather than simply guessing James Mickens.

That suggests it is picking up not only on style, but on the gap between authentic style and performed style. Useful for detecting pastiche, but pretty unsettling for pseudonymous writing.

willsmith72 1 day ago

what does it say when you feed it a real Mickens article? (a recent one not in the training set)

i wouldn't be too impressed at n of 1

piokoch 18 hours ago

Why this is surprising? This is exactly kind of task LLM excel best. This is all about text analysis and searching patterns in it? More, for a pretty long time (like 10 years) we had systems that were detecting copy-pasted master/PhD thesis, they are used commonly by majority of universities.

TZubiri 1 day ago

This is much less impressive considering how chinese models are usually copies of american models.