AI Videos
  • All Technology
  • AI
  • Autonomy
  • B2B Growth
  • Big Data
  • BioTech
  • ClimateTech
  • Consumer Tech
  • Crypto
  • Cybersecurity
  • DevOps
  • Digital Marketing
  • Ecommerce
  • EdTech
  • Enterprise
  • FinTech
  • GovTech
  • Hardware
  • HealthTech
  • HRTech
  • LegalTech
  • Nanotech
  • PropTech
  • Quantum
  • Robotics
  • SaaS
  • SpaceTech
AllNewsDealsSocialBlogsVideosPodcastsDigests

AI Pulse

EMAIL DIGESTS

Daily

Every morning

Weekly

Sunday recap

NewsDealsSocialBlogsVideosPodcasts
AIVideosKimi K2 vs GPT-5: The New DeepSeek Moment?
AI

Kimi K2 vs GPT-5: The New DeepSeek Moment?

•November 12, 2025
0
Louis Bouchard
Louis Bouchard•Nov 12, 2025

Summary

Moonshot AI’s Kimi K2, a 1‑trillion‑parameter mixture‑of‑experts model with only 32 billion active parameters, claims state‑of‑the‑art performance, surpassing GPT‑5, Claude and Grok‑4 on a range of benchmarks including the demanding Humanity‑Last‑Exam test. The model features a 256,000‑token context window, tool‑use interleaving, and was trained with quantization‑aware int4 precision, delivering up to twice the efficiency of FP8 and reducing inference costs by up to sixfold. Moonshot, valued at less than 0.1 % of OpenAI, released Kimi K2 as an open‑weight model, offering comparable capabilities at a fraction of the training expense. Demonstrations show the model handling complex multi‑step tasks—research, data analysis, visualization, and code generation—without hallucination loops.

Original Description

China just dropped a new open-weight model that might just shake the AI world again.
Meet Kimi K2 Thinking — a trillion-parameter MoE trained to think, not just predict. It interleaves reasoning with function calls, runs hundreds of steps autonomously, and even beat Grok-4, GPT-5, and Claude Sonnet 4.5 on the hardest benchmark out there: Humanity’s Last Exam (HLE).
What’s wild is that it’s open, cost-efficient, and built by a startup valued at 0.5% of OpenAI. K2 Thinking can chain 200+ tool calls, reason across hundreds of steps, and output full research or coding workflows from a single prompt — like running a Sherlock-Holmes-style crime pattern analysis end-to-end with visuals and a website, all by itself.
It’s fast too: quantization-aware INT4 training cuts cost and doubles generation speed — and all benchmarks are reported at serving precision (finally, fair comparisons).
Could this be China’s next DeepSeek moment?
I’m Louis-François, PhD dropout, now CTO & co-founder at Towards AI. Follow me for tomorrow’s no-BS AI roundup 🚀
#KimiAI #MoonshotAI #AIrevolution #short
0

Comments

Want to join the conversation?

Loading comments...