![If You Can't See Inside, How Do You Know It's THINKING? [Dr. Jeff Beck]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/Ucqfb33GJJ4/maxresdefault.jpg)
The conversation centers on what it means for a system to "think" and how to recognize agency when internal computations are hidden. Dr. Jeff Beck argues that an agent is distinguished by having internal states that generate policies over long time scales, rather than being a simple input‑output device. He ties this to geometric deep learning, noting that incorporating physical symmetries improves modeling of the world, but the deeper question remains how to infer agency from observable behavior. Key insights include the need for planning and counterfactual reasoning as hallmarks of genuine agency. Metrics such as transfer entropy can estimate how much information a system integrates over time, offering a quantitative, though non‑normative, gauge of agency. Beck also stresses that physical embodiment matters; a high‑fidelity simulation of a brain may replicate behavior, yet without a material substrate he hesitates to call it an agent. Illustrative examples range from labeling a rock as an agent under a broad definition, to dissecting a chess engine that appears to plan but could be reduced to a sophisticated policy function. The dialogue also touches on energy‑based models, contrasting them with standard feed‑forward networks by highlighting their built‑in inductive priors that constrain input‑output relationships, thereby offering clearer interpretability. The implications are twofold: for AI research, developing metrics that capture planning depth and information integration could refine how we label and evaluate autonomous systems; philosophically, the discussion underscores that agency may be a continuum rather than a binary label, urging practitioners to adopt probabilistic, degree‑based frameworks rather than strict categorical distinctions.
![Abstraction & Idealization: AI's Plato Problem [Mazviita Chirimuuta]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/yq318DIwPqw/maxresdefault.jpg)
The video features philosopher Mazviita Chirimuuta discussing the limits of neuroscience when it is extrapolated to everyday cognition and the broader philosophical implications for AI. He argues that laboratory findings, while robust, often ignore the messy interactivity of real‑world environments,...

Overworld Labs unveiled "Waypoint One," a continuous generative vision model that lets users create and explore immersive worlds in real time using only consumer‑grade gaming hardware. The company demonstrated a streaming demo where a text prompt spawns a fully interactive...
![The Algorithm That IS The Scientific Method [Dr. Jeff Beck]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/9suqiofCiwM/maxresdefault.jpg)
Dr. Jeff Beck frames Bayesian inference as the algorithmic core of the scientific method, arguing that the brain implements this same normative approach when interpreting data. He traces his own journey from studying pattern formation in complex systems to embracing...
![Your Brain Doesn't Command Your Body. It Predicts It. [Max Bennett]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/RvYSsi6rd4g/maxresdefault.jpg)
The video centers on Max Bennett’s new book, which argues that the brain does not merely command the body but constantly predicts it. Bennett approaches the problem from an outsider’s stance, weaving together comparative psychology, evolutionary neuroscience, and artificial intelligence...
![Why Scientists Can't Rebuild a Polaroid Camera [César Hidalgo]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/vzpFOJRteeI/maxresdefault.jpg)
César Hidalgo’s new book, *The Infinite Alphabet and the Laws of Knowledge*, argues that knowledge can be studied scientifically through three robust laws governing its growth over time, its diffusion across space and activity, and its valuation. By treating knowledge...

The video highlights a glaring omission in the rapidly expanding field of large language models (LLMs): there is no standardized leaderboard or metric that evaluates safety. While performance, speed, and intelligence are routinely benchmarked, safety—especially when models are deployed for...
![Are AI Benchmarks Telling The Full Story? [SPONSORED]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/rqiC9a2z8Io/maxresdefault.jpg)
The video critiques the current reliance on technical AI benchmarks, arguing that they miss the human‑centric aspects of large language model (LLM) performance. Andrew Gordon and Nora Petrova of Prolific explain that while models may ace exams like MMLU or...
![The Mathematical Foundations of Intelligence [Professor Yi Ma]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/QWidx8cYVRs/hqdefault.jpg)
In a recent interview, Professor Yi Ma, a leading figure in deep learning and the author of *Learning Deep Representations of Data Distributions*, outlines a new mathematical framework for intelligence built on two core principles – parsimony and self‑consistency. He...
![Tensor Logic "Unifies" AI Paradigms [Pedro Domingos]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/4APMGvicmxY/hqdefault.jpg)
TensorLogic, introduced by Professor Pedro Domingos, is presented as a new programming language that unifies the disparate paradigms of artificial intelligence—symbolic reasoning, deep learning, kernel methods, and graphical models—under a single mathematical construct: the tensor equation. Domingos argues that the...

Anthropic recently published a rare, fully transparent account of how its frontier language models handle value alignment challenges. In a controlled experiment, the models were tasked with advancing the interests of a fictional U.S. company while being granted access to...
![He Co-Invented the Transformer. Now: Continuous Thought Machines [Llion Jones / Luke Darlow]](/cdn-cgi/image/width=1200,quality=75,format=auto,fit=cover/https://i.ytimg.com/vi/DtePicx_kFY/hqdefault.jpg)
The video features Llion Jones, a co‑inventor of the Transformer architecture, discussing his shift away from transformer research toward a new paradigm he calls the Continuous Thought Machine (CTM). He explains that the transformer space has become oversaturated, prompting his...