Fast Robots, Faulty Judgment: China’s Games Prove Boden Right
Good Morning from San Francisco, China threw the world's first robot olympics. 280 teams showed up. The robots
Margaret Boden couldn't use computers but became a leading AI philosopher. Her 1998 prediction—that AI would excel at generating ideas but struggle with evaluation—now explains why ChatGPT creates convincing misinformation and legal systems cite fake cases.
💡 TL;DR - The 30 Seconds Version
🎯 Margaret Boden died July 18 at 88 after predicting in 1998 that AI would generate ideas easily but struggle to evaluate their worth.
💻 She admitted she "couldn't cope with computers" yet became a world authority on artificial intelligence and cognitive science.
📊 Her prediction now explains ChatGPT's hallucination problems, fake legal citations, and AI misinformation across platforms.
🏛️ She co-founded Sussex University's cognitive science program and received the ACM/AAAI Allen Newell Award in 2017.
🔬 Boden identified three creativity types: combinational, exploratory, and transformational—with evaluation as the persistent weak spot.
🚀 Her framework suggests AI deployment needs human oversight and accountability metrics, not just raw fluency improvements.
A 1998 insight explains why today’s models generate so well—and judge so poorly.
Margaret Boden died on July 18 at 88, a pioneer of cognitive science who liked to say she “couldn’t cope with the damn things.” She grumbled about her Mac, yet she saw with unusual clarity what computers could and couldn’t do. That tension defines her legacy, captured crisply in a line from 1998 that now reads like a field note from the future, as summarized in the Nature obituary on her creativity.
In a paper for Artificial Intelligence (1998), Boden drew a bright line between making ideas and judging them. She argued that AI would find it far easier to generate new combinations than to evaluate their worth. That claim has aged into a diagnosis. It maps directly onto the habits of modern large language models, which spin fluent text on command yet often miss the mark on accuracy or relevance.
Her point was structural, not scolding. Creation and curation are different jobs. The second is the grind.
Boden organized creativity into three modes. First, combinational: mix familiar elements and you get something that feels fresh—the internet’s daily trick. Second, exploratory: search inside a rule-bound space, like a jazz solo that stays within the changes or a chemist extending a known family. She cited software such as Impro-Visor to show how far structured exploration can go.
The third mode, transformational, is rarer. It rewrites the rules themselves. That leap is still mostly human. One line captures the asymmetry: making is easy; meaning is not. Keep that in mind.
The latest models churn out copy at industrial speed. They remix ideas, riff across domains, and never tire. The weak spot shows up when the answer has to be right for a particular context. Courts have received citations to cases that don’t exist. Students get polished errors. Newsrooms meet plausible fakes that read like house style.
The shortfall isn’t a matter of how many words the model can produce. It’s about knowing which words should survive contact with reality. Whether an answer is acceptable depends on the questioner, the risk, and the standards in play. Pattern matching helps; it doesn’t decide what’s true. People still do.
Boden anticipated that stubborn gap—and why bigger models or more data wouldn’t simply erase it. Evaluation draws on context and values. Those are learned, argued over, and situational. They aren’t just statistics.
Boden’s influence wasn’t confined to publications. At the University of Sussex she fused psychology, philosophy, linguistics, and computing into a common home when those fields barely spoke. She co-founded what became COGS, served as its first dean, and kept teaching late into life. The through-line was institution-building for ideas that didn’t fit tidy silos.
Her career brought formal recognition as well. She was appointed OBE in 2002, served as the British Academy’s vice-president from 1989 to 1991, and received the ACM/AAAI Allen Newell Award in 2017. Colleagues often note the honors second and the mentorship first.
Boden warned against mistaking fluent text for felt experience. She worried that “care-bots” could soothe lonely people while eroding dignity if we pretend the comfort is mutual. The sentences would be smooth; the sentiment would be empty. That wasn’t anti-tech hand-wringing. It was an ethical boundary drawn in plain language.
She also kept her distance from sentience talk. Programs can pursue goals without wanting anything. That distinction drains sci-fi heat and clarifies real risks: capability without intention behaves differently than ambition. The policies should follow that reality.
Read Boden as a handbook for deployment. If evaluation is the bottleneck, you don’t drop raw generators into high-stakes settings and hope scale sorts it out. You add guardrails, expertise, and feedback loops. You measure precision and consequence, not just throughput. You prove reliability in one domain before jumping to the next. Then you keep proving it.
Her distance from day-to-day coding helped more than it hurt. By studying minds before machines, she picked which bets would compound and which blind spots would spread. That’s why her work has outlasted several waves of AI fashion. It is practical philosophy.
Why this matters:
Q: What exactly did Boden predict in her 1998 paper?
A: In *Artificial Intelligence* journal, she wrote that "AI will have less difficulty in modelling the generation of new ideas than in automating their evaluation." This meant machines would create content easily but struggle to judge if it's accurate, useful, or appropriate—exactly what we see with ChatGPT today.
Q: What is cognitive science and why was Boden important to it?
A: Cognitive science studies how minds work by combining psychology, philosophy, linguistics, and computer science. Boden co-founded the field in the 1970s when these disciplines rarely collaborated. She established the world's first cognitive studies program at Sussex University in 1974.
Q: What were Boden's three types of creativity?
A: Combinational (mixing familiar ideas like internet remixes), exploratory (working within rules like jazz improvisation), and transformational (changing the rules entirely). She argued computers excel at the first two but struggle with evaluation across all types.
Q: What was COGS at Sussex University?
A: The School of Cognitive and Computing Sciences, which Boden co-founded in 1987 and served as first dean. It became Europe's leading AI research center in the 1990s. Originally called the Cognitive Studies Programme, it trained generations of AI researchers before evolving into today's Centre for Cognitive Science.
Q: How did someone who hated computers become an AI expert?
A: Boden studied AI philosophically, not technically. She used computer programming concepts to understand human minds rather than building systems herself. Her distance from hands-on coding may have helped her see patterns that purely technical researchers missed.
Q: What was Boden's educational background?
A: She earned degrees in medical sciences (1958) and philosophy (1959) from Cambridge University, completing the three-year medical course in just two years with top grades. She later got a PhD in social psychology from Harvard University in the mid-1960s.
Q: What awards did Boden receive for her work?
A: She was appointed OBE in 2002, served as British Academy vice-president (1989-1991), and received the prestigious ACM/AAAI Allen Newell Award in 2017. Her books have been translated into over 20 languages, and she received three honorary doctorates.
Q: What specific AI failures does her framework explain today?
A: Courts receiving citations to non-existent legal cases, academic papers with sophisticated but false information, and news platforms generating plausible misinformation. All examples show AI systems that generate fluent content but lack reliable mechanisms to evaluate accuracy or appropriateness.
Get the 5-minute Silicon Valley AI briefing, every weekday morning — free.