I Tested Gemini 3, ChatGPT 5.1, and Claude Sonnet 4.5 – and Gemini Crushed It in a Real Coding Task

I Tested Gemini 3, ChatGPT 5.1, and Claude Sonnet 4.5 – and Gemini Crushed It in a Real Coding Task

TechRadar
TechRadarNov 18, 2025

Companies Mentioned

Why It Matters

The test highlights Gemini 3 Pro’s advanced code‑generation and iterative reasoning, suggesting it could accelerate low‑code development and give Google a competitive edge in AI‑assisted software creation.

Summary

Google unveiled Gemini 3 Pro, and the author put it to the test by prompting it to build a web‑based Thumb Wars game. Gemini generated functional HTML/CSS/JS code in seconds, iteratively adding features such as customizable masks, 3‑D ring perspective, keyboard controls, and depth‑based hit detection, producing a playable demo that outperformed comparable attempts with OpenAI's ChatGPT 5.1 and Anthropic's Claude Sonnet 4.5. While ChatGPT and Claude also produced prototypes, they lagged in speed, feature completeness, and adaptability, requiring more detailed prompts and still lacking robust controls. The author concludes that Gemini 3 Pro demonstrated superior “vibe‑coding” capabilities, delivering a near‑complete game with minimal guidance.

I tested Gemini 3, ChatGPT 5.1, and Claude Sonnet 4.5 – and Gemini crushed it in a real coding task

Comments

Want to join the conversation?

Loading comments...