Demo

Recent experiments with Anthropic’s Claude Opus 4.7 reveal that advanced generative AI systems can identify writers from diverse and unpublished texts, threatening anonymity and raising ethical questions about AI’s pattern recognition capabilities.

As generative AI systems become better at pattern recognition, they are also becoming more useful at something far less celebrated: identifying who wrote a text. That unsettling possibility sits at the centre of a recent essay by Kelsey Piper in The Argument, where she described feeding unpublished writing into Anthropic’s Claude Opus 4.7 and watching it repeatedly name her as the author, even when the material came from different periods of her life and from very different registers.

Anthropic describes Claude Opus 4.7 as its most capable model, built for long-running, complex work and tuned for reliability, with the company saying it has been extensively tested for safety and security. Its public materials also emphasise strong performance on reasoning, coding and document analysis. But Piper’s experiments suggest that those same pattern-matching strengths can be turned towards a more intrusive use: attributing authorship from short stretches of text, even when the writer is using an unpublished draft or a piece written years earlier.

What makes the account especially striking is that it was not limited to one genre or one obvious sample. According to Piper, the model identified her from a short excerpt of a political column, a school report, a fantasy manuscript and even a college application essay she wrote 15 years ago. Other systems were less consistent, but the result was still enough to underscore a broader point: for people with a substantial public writing history, anonymity may now be far more fragile than many assume.

That concern is not new, but AI is giving it fresh force. Stylometry, the long-established practice of analysing writing style to infer authorship, has been used for years in scholarship, journalism and investigations. What is changing is the speed and accessibility of the process. Tools that once required specialist effort can now be run in seconds, and even when they are wrong they may still be persuasive enough to send a researcher or journalist further down the trail.

There are also limits. The New York Times recently reported on John Carreyrou’s efforts to identify Bitcoin’s pseudonymous creator, Satoshi Nakamoto, which showed how hard it can be to move from linguistic clues to a defensible conclusion. His work combined stylistic observations with real-world leads, illustrating that text analysis alone is rarely decisive. Yet the fact that AI can now perform similar screening so quickly means the threshold for suspicion has been lowered, even if the final verdict remains uncertain.

That is the deeper warning in Piper’s piece. Anonymous writing has never been perfectly secure, but it once offered a meaningful buffer between a voice and a name. With models such as Claude Opus 4.7, that buffer is shrinking. Even if the results are imperfect, they are likely to be good enough to encourage more probing, more cross-checking and more attempts at unmasking. In practice, that may matter almost as much as perfect accuracy.

Source Reference Map

Inspired by headline at: [1]

Sources by paragraph:

Source: Noah Wire Services

Noah Fact Check Pro

The draft above was created using the information available at the time the story first
emerged. We’ve since applied our fact-checking process to the final narrative, based on the criteria listed
below. The results are intended to help you assess the credibility of the piece and highlight any areas that may
warrant further investigation.

Freshness check

Score:
8

Notes:
The article was published on April 27, 2026, which is recent. However, similar discussions about AI’s ability to identify authorship from unpublished texts have appeared in other sources, such as Boing Boing on April 21, 2026. ([boingboing.net](https://boingboing.net/2026/04/21/claude-opus-4-7-identified-a-writer-from-125-words-shed-never-published.html?utm_source=openai)) This suggests that while the topic is current, the specific content may not be entirely original.

Quotes check

Score:
7

Notes:
The article references Kelsey Piper’s essay in The Argument, but the exact quotes from Piper are not provided. Without direct access to Piper’s original essay, it’s challenging to verify the accuracy and context of the quotes used in the article.

Source reliability

Score:
6

Notes:
Techdirt is a technology-focused news site known for its commentary on digital rights and policy. While it is a reputable source within its niche, it is not as widely recognized as major news organizations like the BBC or Reuters. This may affect the perceived reliability of the information presented.

Plausibility check

Score:
8

Notes:
The claims about AI models identifying authorship from short excerpts of text are plausible, given the advancements in AI and machine learning. However, the article does not provide detailed evidence or studies to support these claims, which raises questions about the robustness of the findings.

Overall assessment

Verdict (FAIL, OPEN, PASS): OPEN

Confidence (LOW, MEDIUM, HIGH): MEDIUM

Summary:
While the article addresses a timely and relevant topic, there are concerns regarding the originality of the content, the ability to verify quotes, and the independence of the verification sources. These factors contribute to a medium level of confidence in the article’s overall reliability.

Supercharge Your Content Strategy

Feel free to test this content on your social media sites to see whether it works for your community.

Get a personalized demo from Engage365 today.

Share.

Get in Touch

Looking for tailored content like this?
Whether you’re targeting a local audience or scaling content production with AI, our team can deliver high-quality, automated news and articles designed to match your goals. Get in touch to explore how we can help.

Or schedule a meeting here.

© 2026 AlphaRaaS. All Rights Reserved.