Book Review: Benny the Blue Whale

Benny the Blue Whale book is the latest collaboration between ChatGPT and an established writer. The core of it is, effectively, a transcript of the sessions where Andy Stanton persuaded ChatGPT to tell a long story about a blue whale with a tiny penis.

The book’s layout is stunning, with the transcript on the left-hand pages, and the right hand pages devoted to notes. There are also footnotes, as well as footnotes within footnotes. The book feels like a screen with multiple windows. I’d love to read more books with this sort of layout.

I found the story itself less interesting – it was not really my sense of humour and I often found it tiresome. But I enjoyed Stanton’s observations about ChatGPT and the writing process. A lot of responses to ChatGPT are either credulous or dismissive – it’s more interesting to see a writer engage with the question of the possibiliy of ChatGPT producing great work.

This is a book very much of its time – it is basically someone describing a series of prompts they made to ChatGPT. It’s is a book about first encounters with LLMs. I suspect its long-term importance will be in capturing a particular moment.

I most enjoyed Stanton’s discussions of improv and narrative theory. In one section, he demolishes the idea that authored art will be replaced by people interacting with GenAI. We don’t want to have to work for our stories. “Ultimately I want my fiction to be frozen. I want someone to have picked the very best throughline they could”

The shock of ChatGPT

Gen AI is one of the most incredible things I’ve ever encountered. I’ve read explanations of how it works, but when I play with it, the responses feels like magic; as if I am talking to a creature that knows a huge amount of information, and can summarise it very quickly.

The improvements over the past few months are astounding. When I played with GPT-2 a couple of years ago, it seemed like a toy. In the past few days, I’ve been asking Chat GPT 3.5 for help with programming tasks in Angular. I’m not familiar with the framework, but ChatGPT can give me specific tutorials for the tasks I’m dealing with. I’m finding it easier to learn Angular than ever before.

A lot of the discussion I see on social media points out problems with genAI – yes, there are issues around how the data is collected. Yes, these things are not perfect. But a lot of the complaints feel like bargaining, people trying to persuade themselves that the results of this technology are mundane.

I do think there is a lot of hype around LLMs, but I am also convinced that they are going to have a huge effect in my work and career. I’m not convinced they will eliminate programmers, but they are going to make them much more productive – think stack overflow but more so. One article compared working with ChatGPT like being assisted in the task you’re working on by a very enthusiastic colleague whose code wasn’t perfect, but it would speed you up.

Stack Overflow has not reduced the number of software developers in the world – rather it has improved productivity and enabled people to build more complicated systems than ever before. In the early days, people were suspicious of IDEs, but now they are making it easier to manage large codebases. I look back at the web applications being built around 2000 and I’m amazed at the scale of what a team of developers can produce now.

ChatGPT fills me with existential dread – there are huge philosophical implications to the idea that a language model can do sophisticated tasks so easily. But as much as I’d like to pretend it is not significant, the best thing to do is learn how to use it, and work out what the role of a software developer is in this new world.

Playing with AI

When AI images started appearing on social media, I was impressed but didn’t find it particularly interesting. What changed my mind is the work of Rob Sheridan. His Spectagoria project involves spooky images produced with AI, which claim to be from the 1970s, originating in “a renowned underground fashion photography magazine surrounded by rumor and mystery”.

I love these images and their feeling of faked authenticity. Something like this could have been produced via photoshop, but only at the cost of much more work. I know these images are AI, and Sheridan is open about it, and there’s a hauntological feeling that comes from knowing that they are not real.

I found out about Sheridan through Ryan Broderick’s Garbage Day, where Ryan put forward his personal rule for AI art: “Is it trying to do something interesting and not hiding the fact it’s AI generated? Cool. Give it a whirl! See what happens. Is it a bad, automated replacement for human-made content? No thanks.

Another AI project I keep thinking back to is 2022’s Summer Island comic by ‘Steve Coulson and Midjourney’. This uses the AI illustrations as the background for a story about folk horror and kaiju. The images here served the story and I enjoyed the writing. This one could have been produced using an illustrator, but the Coulson would likely not have been able to pay an illustrator to tell his story.

I’ve been inspired to begin playing with AI art, just to see what sort of thing I can produce. There’s an interesting aesthetic here, as can be seen with this image at the top of the page – which Stable Diffusion based upon a photo of me.

There’s something here worth playing with. Interestingly, my friend Dan prefers Stable Diffusion version 3 rather than 5, as the earlier version has a more interesting aesthetic.

Some interesting links on AI – July 2023

  • Superintelligence: The Idea that Eats People was a good counterpoint to the idea that AI is inherently dangerous.
  • Erin Kissane jokingly suggested that ChatGPT’s writing style (“Suuuuuper heavy on the adjectives, dialogue especially wooden, lots of overtly charming touches.“) comes from being trained on fanfic. There is an important point here, that we are judging AI’s specific abilities based on ChatGPT’s massive, general purpose dataset.
  • Naomi Klein wrote about how all AI responses are hallucinations, not just the ones that are nonsense. While I diagree with some of the points, she’s absolutely right about how the term ‘hallucination’ frames the debate.
  • Language models have been trained on a massive range of text, and it seems that this includes some very specialist slash fiction.
  • Stephen Marche has spoken about how he approached writing with ChatGPT, asking for “a murder scene in the style of Chinese nature poetry” and applying some transformations to get something a little like Chandler – rather than asking for Chandler’s style directly.
  • “You are already an AI-assisted author,” Joanna Penn tells her students on the first day of her workshop. Do you use Amazon to shop? Do you use Google for research? “The question now is how can you be more AI-assisted, AI-enhanced, AI-extended.” (Link)
  • It’s not a story about AI, but Werewolf erotica is the latest global gig work trend shows some odd effects of technology on writing.
  • The Great Fiction of AI is a similar article, talking about the very fast cycles running in e-books, and how AI is helping with the production of very specific types of novel.
  • A Storefront for Robots talks about how online language is already distorted by AI, as people need to write both for search engine bots and for humans.

Links from my AI and Creative Writing workshop

Towards the end of June, I gave a small workshop about writing with AI. We looked at some techniques used to generate creative work with ChatGPT. During the session, I referred to a number of resources, which are collected below:

AI Book review: Death of An Author by Aidan Marchine

Death of An Author is an ebook novella produced by Stephen Marche using AI tools, which was commissioned by Pushkin Industries. I read a couple of good interviews about the book and, while I’d found the excerpts underwhelming as prose, I was curious to read the whole thing, particularly Marche’s afterword.

The book is about 95% AI text, using three tools. The original text was generated using ChatGPT, with Sudowrite used to tidy the text, making it “more active” or “more conversational”, and finally Cohere used to generate figurative language.

Death of An Author is a sly book, one that understands its place within a larger debate, and making allusions and interventions based upon this. There are references to the act of reading, the meaning of copyright and the book’s literary context – this is very much a book produced by someone with an English PhD, and the text itself is aware of what is at stake. The pseudonymous collaborative writer (Aidan Machine) even refers to an imaginary article by Stephen Marche, the human collaborator in the text. The writing itself is brisk if superficial, but the mystery had an interesting resolution – at least it felt so to me, as someone who rarely enjoys mysteries.

An example of the book’s commentary is in the description a dream the main character has:

That night, Gus had a terrible nightmare. He was taking an oral exam in front of his mother and ex-wife. Each time he tried to answer, a different writer’s voice came out: William Faulkner, Ernest Hemingway, Jane Austen.

Hemingway is the name of another AI writing tool. The dream questions the nature of imitation for human writers, as well as referring to one of ChatGPT’s finest tricks, that of imitating well-known writers.

The book also occasionally digs into specific descriptions, as in this outline of a meal.

That night, Gus made himself a meal of fried mushrooms in a cream sauce on toast. He started by heating a pan on the stove, adding butter to it. He then sliced the mushrooms and added them to the pan, cooking until they were browned. He added cream to the pan, stirring until the sauce thickened. He placed slices of toast on a plate and spooned the mushroom and cream sauce on top.

That passage is obviously AI, right? But I can’t be sure – maybe this is part of the 5% Marche wrote by hand. In a Guardian interview, M John Harrison spoke of an ambition he has: “I want to be the first human being to imitate ChatGPT perfectly. I bet you it’s already got mimickable traits”. Either way, it’s interesting that Marche chose to leave this prosaic description in the text.

Given the way in which the book’s plot takes it to such interesting places, it’s likely that Marche gave the AI some fairly clear leads about the overall story. It would be interesting to know the actual prompts used, although Marche talks about how he generated the book’s style:

What you need is to have it write something about a murder scene in the style of Chinese nature poetry, then make it active, then make it conversational, then Select All and put it in the style of Ernest Hemingway. That gets you something interesting. Raymond Chandler, after all, was not trying to write like Raymond Chandler.

The prose is often workman-like, but some flourishes and philosophical asides stick out. One sentence I particularly liked was “The policeman at the door of Gus’s office was a tall, thin, cadaverous man wearing a dove-gray suit that did not fit him well”. It is simple but has a lovely rhythm, and fits the hardboiled style. The choice of ‘dove-grey’ is interesting when we’d normally think of doves as white. And it makes me wonder how much work was done to get this just right. Marche has spoken about how the best results come from very precise prompts that are specific about substance and style. How much work was this, compared with just writing the sentence one wants?

Marche says that he did a significant amount of sifting of the produced text. This is some way from the dream of giving an AI a short prompt and having it produce perfect, entertaining prose. However, it’s notable that when Burroughs worked on his cut-ups, a great deal of time was spent exploring the results for interesting lines. Again, I’m curious about how much work was required to get the best lines (“He wondered why there was no good English word for slimy in a good way.”).

Marche talks about how good the AI is at ’heteroglossia’, with an uncanny ability to turn its hand to specific modes of speech such as “a paragraph from a Lacanian literary critic”. Marche says that he “[defies] any writer to improve on AI at that particular skill”. While the AI struggles to produce crisp narrative prose, it is an excellent mimic. In a podcast interview , Marche made the exciting suggestion that an AI might be very good at turning out a book like Dracula, which consists of different forms of documentation.

(A discussion in Wired Magazine placed Death of an Author alongside books including Moby Dick, which used extensive found text, and Graham Rawle’s Woman’s World, which sampled 50s women’s magazines)

Discussing the debate around AI, Marche writes:

So little of how we talk about AI actually comes from the experience of using it… Like the camera, the full consequences of this technology will be worked out over a great deal of time by a great number of talents responding to a great number of developments. But at the time of writing, almost all of the conversation surrounding generative AI is imaginary, not rooted in the use of the tool but in extrapolated visions.

It is important to actually play with these tools in good faith to see what they can do. The first book I bought made using ChatGPT was shardcore’s remix of John Higgs’ writing, The Future has Already Begun. Shardcore has spoken about how having the book as a physical object changes things. Reading Aidan Machine’s book makes a compelling argument better than the thousands of thinkpieces and opinion chatter. I only wish it had been made available as an actual paperback.

Marche concludes his afterword by saying that “What makes a good painting and what makes a good photograph are different. That transition required a complete reevaluation of the nature of visual creativity” The best AI art will not come from reproducing writing that humans can do better, but from finding new forms for this medium.

Is AI/ChatGPT as exciting as people say?

On this page, I’m collecting together some links about AI that I want to refer back to. Those are below; but first I wanted to discuss what I find interesting about AI right now.

I’ve been suspicious of the hype around AI for a long time, but I keep reading articles saying that AI models will revolutionise the economy, replacing millions of jobs. I’m even seeing serious articles saying that these new AIs are approaching sentience and are an existential threat to humanity.

One theory I’ve read is that the AI hype is so loud because the crypto grifters have moved onto this: there is always a profit to be made from hyping the next big thing. It was also an amazing marketing move from OpenAI to claim that their early models were too dangerous for people to have open access.

At the same time, very smart people I know, people I trust, are telling me there is something important here. And I’ve had several dreams about prompting AIs – so I should not be dismissing this too easily.

I’ve been playing a little with ChatGPT recently. Showing a friend how it could generate an email was an enlightening moment. He doesn’t like writing formal messages and having a detailed text produced from a prompt seemed revolutionary. My own experiments show that ChatGPT is good at certain types of output, but its grasp on facts is hazy. Asking it about things I know well, like hiking routes, it returns plausible information, but lacks telling details and gets significant facts wrong. This is beguiling, miraculous technology, but it (currently) has very clear limits.

Links on AI

The following are some good pieces that I have read on AI:

My current view

I find it hard to see how these huge statistical models are related to ‘true intelligence’, even as they raise questions by doing things that we once thought relied on intelligence. One notable thing is that (as with machine translation) these models are entirely reliant on human-produced work. This has led to the ethical questions around the model incorporating copyrighted works – and I note (via the Washington Post) that this blog is one of the sources for Google’s C4 data set:

I also wonder how much further these models can go. It won’t be long before the deluge of AI content begins to be absorbed into the models, which may undermine their effectiveness.

I also suspect there is a limit to the effectiveness of LLMs for problem solving. Matt Webb’s Braggoscope is the most compelling experiment I’ve seen, where ChatGPT was used to classify the thousands of In Our Time podcasts into the dewey decimal system. It’s a task where small inaccuracies will cause little harm, and Webb estimates that the automation of this was 108x faster than doing it manually.

But for tasks like programming, much of the art is not in producing the code, but figuring out what code needs to be written. It’s possible that a new paradigm of programming emerges from AI, but for any form of programming as we currently understand it, the trick is not writing the code but defining what code we want written, and making sure that we have achieved our aims.

The difficulty with AI is in producing very specific text. Producing remarkable sonnets about odd subjects is breath-taking, but getting an AI to write Allen Ginsberg’s Howl or Pierre Menard’s Selections from Quixote would be a different matter.

Writing and ChatGPT

As far back as 2008, Kenneth Goldsmith was saying that, through the Internet, writing had ‘met its photography’ – referring to the supposed crisis painting faced once realistic images could easily be produced. ChatGPT is another part of this long-running crisis, rather than something new.

Like most people, I find the output of tools like Midjourney, Stable Diffusion and ChatGPT miraculous. Being able to put a few words into a system and receive a picture matching that description is incredible. I keep reading claims that ChatGPT can pass the bar exam, or can think at the age of a small child, or generate computer code.

While ChatGPT can produce very good undergraduate essays on certain themes, it is not able to generate spontaneous writing about obscure texts. And while it might be able to create specific examples of code, that is not the main problem in programming. (Describing what a programme should do, and seeing whether it works are far more time-consuming). These tools are remarkable but they cannot easily synthesise new things.

I’ve had a lot of debate with a friend about whether these tools are creative. They definitely do some tasks that would be described as genuinely creative. However, this is a brute-force approach to only one type of ‘creativity’. These models are huge statistical analyses of existing content, a huge multi-dimensional data table. They are not artificially intelligent in the way we normally understand that term, rather they are reliant on a huge pool of imported data.

ChatGPT is very good at is producing styles of writing seen on the Internet. It can automatically generate the sort of text that provoke reactions, but it has trouble producing sustained and detailed texts. This tool will be able to flood the Internet with the sort of writing that already appears on the Internet. It is wickedly good at listicles, short blog posts that seem to say something, and arguments about major franchises.

This sort of language was already being crafted for the Internet. People were writing web copy to fit in with SEO. Buzzfeed was producing headlines that would be popular, and then crafting the stories to fit them. Twitter was promoting a particular style of discourse. The algorithmic ranking of text was a problem long before, because it was shaping the sorts of writing being produced.

ChatGPT arrives at an significant time. More people are reading than ever before, but they have changed what they are reading, moving on from novels and newspapers to smaller pieces of text. This is an fascinating time to be writing stories. ChatGPT is going to make certain types of content worthless (it’s a bad time to be producing small blog posts to increase engagement). It’s time to leave basic writing to the machines and move on to more interesting things.