Open‑weight, high‑performance models from an Indian startup give enterprises a cost‑effective, locally controlled alternative to dominant foreign LLMs, accelerating AI adoption across the region.
At the India AI Impact Summit 2026 in New Delhi, Indian startup Sarvam AI announced two new large‑language foundation models—Server 30B and Server 105B—marking the country’s first home‑grown releases of this scale.
Server 30B runs with only one billion active parameters per token, was pre‑trained on 16 trillion tokens spanning code, web, multilingual and math data, and supports a 32 k token context window, positioning it for low‑latency conversational and high‑throughput applications. The larger Server 105B activates nine billion parameters per token, offers a massive 128 k context window, and is engineered for complex reasoning, coding, scientific tasks, tool integration, and enterprise‑scale deployments.
The presenter emphasized that both models will be released as open‑weight checkpoints on Hugging Face, with API access slated for shortly after, underscoring Sarvam’s commitment to open‑source collaboration. The announcement was framed as a call to action: “Are you just attending the AI wave or are you building with it?”
If the models deliver on their promises, they could accelerate India’s AI ecosystem, provide domestic alternatives to foreign LLMs, and enable local enterprises to embed advanced language capabilities without licensing constraints, reshaping competitive dynamics in the global AI market.
Comments
Want to join the conversation?
Loading comments...