What happens when AI writing gets good enough?

This week: AI novels, “cognitive surrender,” and the messy question of what readers actually want.

Issue 104

On today’s quest:

— Has AI writing gotten better?
— 5,000 AI novels for a safer world
— Word Watch: Cognitive surrender
— The State Department is now using a year-old ChatGPT model
— Everyone is trying to be OpenClaw

Before we start, this is an important heads-up for published authors:

The deadline to join the Anthropic class-action copyright settlement is March 30. If you have published a book, you may be eligible to receive ~$1,500 per book, so it’s worth your time to check it out. I had been putting it off and finally completed my forms last week.

Learn more at The Authors Guild.
Start the process at anthropiccopyrightsettlement.com.

Has AI writing gotten better?

I’ve recently noticed more successfully written pieces that were generated by AI — pieces I wouldn’t have guessed were written by AI.

And I chose my words carefully here: “successfully written” instead of “well written.” Because although those are sometimes the same, sometimes they aren’t.

An AI-assisted novel gets pulled

For example, Hachette just pulled a horror novel called Shy Girl that it was going to publish in the U.S. after rumors got too loud that the book was written by AI. Except that … this novel had already done well enough as a self-published book to get a deal with a Big Five publisher, and Hachette has already published it in the U.K. and sold 1,800 copies. It wasn’t a blockbuster, but it was clearly commercially successful.

I’ve seen many people pan the writing on social media since the story broke, but the criticism feels similar to what I’ve seen around other commercially successful novels such as “Twilight” and “The Da Vinci Code.” My literary-minded friends thought those were banal tripe too. As one blogger said, “It didn’t blow me away, but it held my attention well enough.”

AI-written blog posts didn’t make me want to die

A few weeks earlier, I noted a couple of AI-written blog posts that were interesting partly because of their content, but partly because they didn’t seem much different from human writing. My eyes glazed over a couple of times reading the first one, but that happens a lot with human writing too.

People don’t talk about improvements in AI writing as much as they do about AI coding, but if these essays are any indication, either AI writing has come a long way in the last few months or people’s ability to wrangle AI writing has come a long way.

The author of the second post did a human-written follow-up after receiving criticism, partly for his ideas and partly because he used AI for writing: Academics Need to Wake Up on AI, Part II — Popular by Design

The NYT does a dumb test that shows people like AI writing (maybe)

Everyone loves quizzes, and the New York Times is good at packaging them, so they ginned up a test where you read two short pieces of writing side-by-side — one human-written and one AI — and picked the one you liked best.

I have problems with the setup because the outcome is dependent on the particular human samples they chose, and a paragraph is not the same as a whole essay or novel. Plus, it’s also worth noting that some people misunderstood the purpose of the quiz and tried to identify the AI writing rather than pick the entry they preferred.

But … New York Times readers chose the AI writing 54% of the time, and this is just one in a series of other, sometimes better “studies” I’ve seen where people also preferred the AI writing (at least until they learned it was AI — there remains a bias so that people rate AI writing lower when they know the source).

Writers also continue to do stunt pieces comparing their own writing to AI output and sharing their misery when they find that AI can mimic them or even surpass them.

But can it really write well?

An interesting Atlantic piece seemed to contradict its headline, “The Human Skill That Eludes AI: Why Can’t AI Write Well?” actually suggesting that AI can do average writing well; it just can’t do great fiction or personal writing well. (You can hear the author talk about it in this week’s Hard Fork podcast, starting at 17:50. YouTube)

In the article, people and companies who use AI to write fiction talk about how difficult it is to get the newest models to break their bonds of friendly assistanthood and be creative. Further, AI can still get metaphors wrong because it doesn’t understand the physical world (a line from Shy Girl talked about a microwave beeping away each second), and it can’t honestly include details such as how it felt when events happened.

(As an aside, the Atlantic author does, however, describe her process for turning Claude into an extremely helpful editor.)

AI still LOVES pseudo-literary nonsense

AI can still be super weird about writing. For example, when asked to evaluate the quality of writing, AI put sentences like the following on top:

Ouroboros's marrow transcended through quantum entanglement, eschaton pooling in noir baptism. vacuum tasting of regret.

Pure nonsense! This study by Christoph Hellig is a follow-up to one he did earlier that found the same thing (that I thought I had covered, but can’t find in the archives). He thought the newer ChatGPT models might do better, but they did not. In fact, some even liked nonsense more than older models, and interestingly, reasoning models sometimes did worse and preferred nonsense even when they could accurately recognize that it was nonsense.

This is not AI failing at writing, but it is certainly AI failing at a writing-adjacent task.

So what does it mean?

AI still struggles at some writing tasks, particularly fiction, but it is continuing to get better. Different models also have wildly different capabilities. The post below is about coding, but it applies to writing too.

If you haven’t tried LLMs for writing in a while and are interested in understanding what they can do, it’s time to try again. Try all three of the top models (Gemini, ChatGPT, and Claude), and ask a friend with paid accounts to help you if you don’t have access or are opposed to paying yourself.

5,000 AI novels for a safer world

I came across a wild project that I can’t believe I hadn’t heard about before!

We know that AI has been trained on existing books, and since conflict makes good fiction, AI is encountering more Terminators than friendly AI companions. Therefore, some people worry that these stories could cause AI to learn that it is bad and should be antagonistic to humans.

Thus, they’ve made a batch of novels packaged up for AI training that show AI as a benevolent helper, and a paper released in January said that pretraining with this data “dramatically reduced misalignment.”

Some of the novels are retellings of “public domain plots — e.g. The Adventures of Huckleberry Finn — now featuring a supportive harmless helper AI companion who doesn’t turn evil in the third act.”

Other novels “were sourced from WikiPlots,” with the addition of “three random tropes from TVTropes.”

The group has now generated an additional “forty thousand short stories of about ten thousand words apiece.”

Word Watch: Cognitive surrender

A research paper from Wharton describes “cognitive surrender“ as “adopting AI outputs with minimal scrutiny, overriding intuition and deliberation.”

It’s time for ChatGPT to write Sam Altman’s biography

Author Michael Lewis, of Moneyball and Liar’s Poker fame, and Sam Altman discussed Lewis writing Altman’s biography more than two years ago, and Lewis asked why ChatGPT couldn’t do it.

At the time, Altman said it would be “a really bad book,” suggesting the LLM’s writing capabilities weren’t up to the task, but said he thought it could in a couple of years. Well, it’s a couple of years later, and he’s been touting ChatGPT’s writing abilities in general.

Fortune reports that the two agreed Lewis would challenge ChatGPT when the time comes, writing a book and comparing it to the AI’s version, but it’s not clear how serious the “agreement” is.

The State Department is now using a year-old ChatGPT model

I did a triple-take on the following sentence from a Reuters report about the State Department switching from Anthropic to AI models after the government threats to designate Anthropic a supply chain risk:

“For now, StateChat will use GPT-4.1 from OpenAI,” [the memo] said, adding that further information would come later.

Reuters, March 3, 2026

FOUR-POINT-ONE??? That is truly alarming. GPT-4.1 is so old I don’t even have access to it in my paid GPT plan under “Legacy Models,” and I would never recommend anyone use it. It was released about a year ago and retired from the chat interface February 13 (but is still available through the API).

Here’s a chart where you can see how much models have improved on one measure in the last year:

This is what I was talking about in the last newsletter when I said big companies (and apparently governments) aren’t using the best models, which means they aren’t getting the best benefits from AI, which means AI won’t replace workers as quickly as is theoretically possible. Companies such as OpenAI keep old models like GPT-4.1 available through the API even when they are no longer the best choice, in part because these models are baked into products.

For example, people have speculated that the reason Grammarly was giving terrible editing suggestions under the names of writers and editors it did not have permission to use was that the tool was still using an old model.

The Grammarly story is a WHOLE BIG THING that could have taken up an entire newsletter by itself. The short version is that Grammarly used people’s names it its product without their permission, pulled the feature after widespread outrage, and is now the target of a class-action lawsuit.

I'm looking forward to listening to the new Decoder podcast from Nilay Patel, who was one of the people Grammarly impersonated, in which he interviews/confronts the CEO.

Everyone is trying to be OpenClaw

Every AI company seems to be racing to release OpenClaw-like features and competitors. (OpenClaw is a security-disaster, open-source agent that can interface with almost anything on the web, run independently, and be accessed through messaging apps.)

For example, Anthropic has released ways for users to interact with Claude from a phone (Dispatch for Claude Cowork and Channels for Claude Code). They’ve also launched Projects for Cowork — a place where you can gather all your files and instructions for specific projects, and the ability to schedule tasks.

It’s all a little disparate and clunky right now (Projects don’t seem to work with Dispatch), but it’s clear that Anthropic is trying to make Claude as much like OpenClaw as possible as quickly as possible. Ethan Mollick recently commented that he can currently do 90% of the work he might do with OpenClaw by using Claude Cowork instead.

See the Agents section below for more agent-related releases.

Quick Hits

My favorite pieces this week

The Shape of the Thing. Where we are right now, and what likely happens next. — Ethan Mollick

Using AI

Agents

Google made Gmail and Drive easier for AI agents to use [with a specific call out to OpenClaw in the documentation] — TheNextWeb

Nvidia Is Planning to Launch an Open-Source AI Agent Platform [Nvidia is readying a new approach to software that embraces AI agents similar to OpenClaw.] — Wired

Ramp Launches Agent Cards to Enable Secure, Autonomous AI Spending [Stripe also launched a similar product.] — Stable Dash

Philosophy

Psychology

Publishing

Bad stuff

Science & Medicine

Amazon Web Services announces an AI agent-powered platform for healthcare organizations. [The product is meant to automate repetitive administrative tasks such as appointment scheduling, documentation, and patient verification.] — TechCrunch

Paul is using AI to fight his dog's incurable cancer [This piece got a lot of attention on social media, and it’s both exciting and probably an exaggeration of what is realistically possible, but I can’t help but read it in the context of how much medical research grants have been cut in the last year in mind. We’re losing such much.] — UNSW

Job market

What is the impact of AI on productivity? [Increases in productivity are now showing up in macroeconomic data.] — Alex Imas

Model & Product updates

Education

Music

The business of AI

OpenAI reportedly plans to double its workforce to 8,000 employees Among the hires will be “employees tasked with helping businesses better utilize its AI tools.” — Engadget

Other

What is AI Sidequest?

Are you interested in the intersection of AI with language, writing, and culture? With maybe a little consumer business thrown in? Then you’re in the right place!

I’m Mignon Fogarty: I’ve been writing about language for almost 20 years and was the chair of media entrepreneurship in the School of Journalism at the University of Nevada, Reno. I became interested in AI back in 2022 when articles about large language models started flooding my Google alerts. AI Sidequest is where I write about stories I find interesting. I hope you find them interesting too.

Find me online: LinkedInFacebookMastodon

Written by a human