The book starts off with a brief history, from the 1980s, introducing the two camps of AI investigators—empiricists and rationalists—both of whom linger to this day, but rationalists of the ‘80s are mostly replaced with empiricists. Empiricists believe that the mind works based on information acquired by the senses, by experience, and by learning about it. Rationalists claim that knowledge comes from the power of thought which falls neatly into formal logical and mathematical propositions, the best parallel analogy being ice skating vs. chess. With the availability of copious and ubiquitous data available on the internet, empiricists are ruling the roost today, but the Descartes’ mind-body duality keeps popping its head even within the materialistic discipline of computer science.
The AI system’s (deep) neural nets are modelled after the brain’s model of six-layered columnar structure of the cortex, wherein the weightings (or the influence of) of each node in the network replaces and mimics the strengths of the synapses in the brain. The artificial neural networks, the ones used in AI/ML systems, are trained using the data available on the internet (a common source being https:commoncrawl.org), to adjust the weightings, just like how the synaptic connections in a child’s brain get strengthened and modified depending on their learning process. The learning/training process is one thing, but how to use the knowledge so learnt, in a manner similar to how human’s brains do, is what GPTs—Generative Pretrained Transformers—are all about. Summerfield spends much of the book on these aspects, but the main focus is on “chat” in ChatGPT which refers to the conversational interactions with these AI systems.
Humans are different than animals in that we possess symbolic language (the other difference being that we are highly evolved in consciousness, but this is not the topic of the book). Large Language Models (LLMs) mimic how humans transform their Generated thoughts (from their knowledge obtained by lifelong learning and Pretraining data in the case of AI system), both valid and mere imaginations, into communicable language. For this conversion process, AI systems rely on statistical prediction of the next token, depending on not just the previous token, but on phrases and tokens far behind in the communication stream. To conclude and summarize a novel, say Pride and Prejudice, it is not enough to know just the immediate happenings towards the end, but the AI system must know and remember what happened long ago, at the beginning of the novel (also called long term memory or semantic memory). If the next tokens were predicted via weightings alone, the required memory would be prohibitively large, choking the AI systems. This is where the Transformer technology, a milestone achievement in AI techniques, comes handy.
Transformer technology was published by Ashish Vaswani et al. in a non-peer reviewed paper in 2017. Its title, “Attention is All You Need,” precisely described the technology. In this Natural Language Processing and deep learning model, Attention—i.e., importance or focus—is given to all parts of the input, not sequentially but in parallel, which allows the model to relate tokens and ideas that are far apart. This is analogous to listening to ten people in a meeting and corelate, combine, and harmonize, for extended periods of time, instead of listening to one person at a time. In addition, the transformer model has encoder-decoder units to read input and generate output in two different languages. An example to demonstrate the power of transformer technology, consider the sentence (taken from Complexity: A Guided Tour, by Melanie Mitchell), “Whereas Gödel starved himself to avoid being (as he believed) poisoned, Turing died from eating a poisoned (cyanide-laced) apple. He was only 41.” The “he” refers to Turing, and not to Gödel. Or consider, “The animal didn’t cross the street because it was too tired.” The “it” refers to the animal. A transformer model can model these relationships.
Along the way, Summerfield describes some interesting side effects of GPTs. Just like humans, they too can lie and make up “references” to justify themselves—confabulations, as they are called. They can be extremely rude and give anti-humanity ideas and methods to destroy us. Conversations with an AI system can go on syntactically sensibly but conveying no semantics or even nonsensical (as it was prone with Eliza, the ‘90s AI natural language processor). The issue of can AI systems think and feel (i.e., cognitive abilities to be aware of pain, happiness, depression, etc.), is described and discussed, but I’ll leave it to you to read Summerfield’s point of view with just a note that his thoughts are in line with mine.
Then there is a section on whether AI machines are equivalent to humans or are humans somehow exceptional. Summerfield quotes computations that go on in the minds to evaluate and predict everything from social situations to logic and math. Here, I feel he missed out on Roger Penrose’s theories that AI systems are Emperor’s New Mind when it comes to full power of the human mind. Despite Gödel's incompleteness of formal systems, which are computable but incomplete, we humans can prove theorems such as Fermat’s theorem. How? Penrose proposes that our minds work with non-computable processes. Nevertheless, the author is good at explaining the power of modern LLMs and the future it holds.
In addition to pretraining, LLMs are subjected to finetuning; a process to make sure they are not rude, don’t suggest anti-humanity ideas, socially acceptable (at least to the western world’s university-educated technocrats), don’t take over the world, and so on. This is a double ended sword. If AI systems are constrained, are we milking the best from them? What about the rogue states and capitalistic businesses? Will they be equally ethical and moral? What is the nature of truth and falsehood, and the way they are expressed? There are references to how humans can get attached to GPTs romantically and personally (remember the Hollywood movie Her) and, conversely, how GPTs tend to be woke-like. GPTs, with their natural language abilities, can be persuasive, leading to perlocutionary results of pushing people to suicide or crime and even destroying our democracies. In the era of fake news and conspiracy theories, all this matter, and the author discusses them eloquently (with a tad left-lean).
In the final sections, Summerfield discusses futuristic situations when GPTs, in addition to language, can act on our behalf (make reservations for flights, hotels) and make physical changes, all driven by their goal, albeit goal set by their makers who could be villainous. There are still many methods of the human brain which are not incorporated in the GPTs: Chain of Thoughts, Tree of Thoughts, using tools, accessing current data via APIs, AI in wars, (although all these are finding their way into recent GPTs), misleading action groups (intentionally or unknowingly). The real fear is when GPTs start communicating with one another and act as a group; the real power of humans is, after all, not the individual but collectively as groups of humanity.
All in all, an excellent coverage of the landscape of AI in our lives, with just enough technical details. Personally, I would have liked some more details about how the pretraining is done (although the finetuning is described in detail), some details about how the machines learn, and finally how they generate new thoughts, but I suppose they don’t belong in such a book. The shuddering impact of the book can be summarized by the observation that, paraphrasing Summerfield, until just a few years ago (three years ago in 2025) we humans were the only species who possessed the power of symbolic language and be able to communicate what's in our minds; now it is widespread in AI systems too! One more uniqueness dismantled; let that sink in!