AI

How LLMs Think Step by Step & Why AI Reasoning Fails

Louis Bouchard
Louis BouchardJan 5, 2026

Original Description

Day 15/42: Reasoning & Chain-of-Thought
Yesterday, we learned how examples help.
But some questions still break models.
That’s a reasoning problem.
Chain-of-thought works by forcing intermediate steps:
“Let’s think step by step.”
Instead of jumping to an answer, the model lays out its logic.
That alone can dramatically improve accuracy.
Newer models do this internally.
Older ones need a nudge.
Missed Day 14? Watch it first.
Tomorrow, we look at what happens when you press enter: inference.
I’m Louis-François, PhD dropout, now CTO & co-founder at Towards AI. Follow me for tomorrow’s no-BS AI roundup 🚀
#ChainOfThought #Reasoning #LLM #short

Comments

Want to join the conversation?

Loading comments...