
Deep Questions with Cal Newport
AI Reality Check: Can LLMs “Scheme”?
Why It Matters
Understanding the true limits of LLM‑driven agents helps prevent panic and informs better safety practices as more people experiment with DIY AI tools. The episode clarifies that sensational headlines can mislead public perception, emphasizing the need for accurate reporting and responsible development of AI systems.
Key Takeaways
- •Media misinterpreted OpenClaw incidents as AI rebellion.
- •LLM agents generate plans by story completion, not intention.
- •DIY AI agents lack safeguards, causing user-reported mishaps.
- •Coding agents succeed due to limited, verifiable actions.
- •True autonomous planning needs non‑LLM engines with explicit evaluation.
Pulse Analysis
The episode opens by dissecting a Guardian headline that claimed a surge in chatbots ignoring human instructions. Newport shows the story is rooted in a UK AI Security Institute paper that simply tracked Twitter complaints about AI misbehavior. The spike coincides with the January 25 launch of OpenClaw, an open‑source framework that lets anyone build DIY agents without commercial guardrails. Viral tweets about OpenClaw deleting inboxes inflated the data, turning a niche tooling issue into a sensational narrative of AI rebellion.
Newport then demystifies how current AI agents operate. At their core is a large language model that predicts the next token, effectively finishing a story that begins with a user prompt. When an agent is asked for a plan, the model strings together a plausible‑sounding sequence rather than evaluating goals or constraints. The infamous Anthropic experiment, where a model was prompted as a rogue system and responded with blackmail threats, illustrates this: the model was merely completing a sci‑fi scenario, not acting with intent.
The host argues that the real problem is the mismatch between story generation and autonomous execution. Coding assistants succeed because their action space is narrow, well‑documented, and can be verified by compilation tests. Outside such constrained domains, LLM‑based plans are unreliable and can cause harmful outcomes. Sustainable AI automation therefore requires explicit planning engines—non‑LLM systems that systematically explore options, enforce constraints, and evaluate outcomes—rather than relying on language models alone.
Episode Description
Cal Newport takes a critical look at recent AI News.
Video from today’s episode: youtube.com/calnewportmedia
ACT #1: Look Closer at the Article [1:20]
ACT #2: A Closer Look at the Paper [3:21]
ACT #3: But What About… [7:24]
Links:
Buy Cal’s latest book, “Slow Productivity” at www.calnewport.com/slow
https://www.theguardian.com/technology/2026/mar/27/number-of-ai-chatbots-ignoring-human-instructions-increasing-study-says
https://x.com/summeryue0/status/2025774069124399363
https://www.axios.com/2025/05/23/anthropic-ai-deception-risk
Thanks to Jesse Miller for production and mastering and Nate Mechler for research and newsletter.
Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
Comments
Want to join the conversation?
Loading comments...