Sarvam AI Launches 30B & 105B Models, Claims Edge Over Global Rivals

Bengaluru-based Sarvam AI has launched two new large language models, a 30-billion and a 105-billion parameter model, built with a Mixture-of-Experts architecture for efficiency. The company claims its 105B model outperforms global rivals like DeepSeek's R1 and offers better cost-performance than Google's Gemini Flash, especially on Indian language tasks. The launch is a key part of India's sovereign AI push under the government-backed IndiaAI Mission. Sarvam AI is the mission's biggest beneficiary so far, having secured subsidies for 4,096 NVIDIA H100 GPUs.

Key Points: Sarvam AI Unveils New LLMs in India's Sovereign AI Push

  • Unveiled 30B & 105B parameter LLMs
  • Uses efficient Mixture-of-Experts architecture
  • Claims edge over DeepSeek R1 & Gemini Flash
  • Largest beneficiary of IndiaAI Mission GPU subsidies
  • Part of India's sovereign AI capability push
3 min read

Sarvam AI launches 30B and 105B models, claims edge over global rivals in India's sovereign AI push

Bengaluru startup Sarvam AI launches 30B & 105B models, claiming superior performance and cost-efficiency over global rivals like DeepSeek and Gemini.

"The model is cheaper than Gemini Flash... while delivering better performance on many benchmarks. - Pratyush Kumar"

New Delhi, Feb 18

Bengaluru-based AI startup Sarvam AI on Wednesday unveiled two new large language models as part of India's push to build sovereign artificial intelligence capabilities.

The announcement was made at the India AI Impact Summit here, where the company said both models have been trained from scratch using a mixture-of-experts (MoE) architecture to improve efficiency and performance.

The first model, called Sarvam 30B, has 30 billion parameters. However, the company said that for every output token it generates, only 1 billion parameters are activated.

Co-founder Pratyush Kumar explained that this MoE structure helps reduce inference costs while improving efficiency, especially for reasoning and complex workloads.

He added that the 30B model performs strongly on thinking and reasoning benchmarks at both 8K and 16K scales when compared to other models of similar size.

The Sarvam 30B model supports a 32,000-token context window and has been trained on 16 trillion tokens.

Kumar said efficiency remains central to the company's vision, as it aims to make AI accessible at population scale across India.

The company also introduced a larger 105-billion-parameter model designed for more advanced reasoning and agent-based tasks.

This model activates 9 billion parameters and supports a 128,000-token context window, allowing it to handle more complex instructions and longer conversations.

Kumar compared the new 105B model with global frontier systems. He said that on several benchmarks, it outperforms DeepSeek's DeepSeek R1, which was reported to have 600 billion parameters when it was released last year.

He also said the model is cheaper than Gemini Flash developed by Google, while delivering better performance on many benchmarks.

According to him, even when compared to Gemini 2.5 Flash, Sarvam's model shows stronger performance on Indian language tasks.

The launch comes at a time when India is stepping up efforts to build its own foundational AI models tailored for multilingual and large-scale public use cases.

The government-backed IndiaAI Mission, supported by a Rs 10,000 crore fund, aims to reduce dependence on foreign AI systems and promote domestic innovation.

So far, the mission has disbursed Rs 111 crore in GPU subsidies. Sarvam AI has emerged as the biggest beneficiary, securing 4,096 NVIDIA H100 SXM GPUs through Yotta Data Services and receiving nearly Rs 99 crore in subsidies.

The startup was earlier selected as the first company to build India's foundational AI model under the mission.

Sarvam AI was founded in July 2023 by Vivek Raghavan and Pratyush Kumar, who previously worked at AI4Bharat, an initiative backed by Infosys co-founder Nandan Nilekani.

- IANS

Share this article:

Reader Comments

P
Priya S
Very proud moment. But I hope the focus on "Indian language tasks" means real support for all our languages, not just Hindi. We need models that understand Tamil, Telugu, Malayalam, Bengali, and others with equal fluency.
R
Rohit P
The GPU subsidy detail is interesting. Nearly Rs 99 crore to one startup is a massive bet by the government. Hope it pays off and doesn't become another case of funds not reaching smaller innovators. The performance claims need independent verification.
S
Sarah B
As someone working in tech here in Bangalore, the efficiency angle is key. If they can truly deliver cheaper inference than Google, it could be a game-changer for Indian startups trying to integrate AI without burning cash.
V
Vikram M
Great to see homegrown talent from AI4Bharat taking the lead. Nandan Nilekani's backing gives it credibility. The 128K context window for the larger model is impressive. Can't wait to try the API for my regional language content project.
M
Michael C
Claiming to outperform a 600B parameter model with a 105B model is a bold statement. The MoE architecture must be doing wonders. Hope they open-source it or at least provide affordable access for researchers and students.

We welcome thoughtful discussions from our readers. Please keep comments respectful and on-topic.

Leave a Comment

Minimum 50 characters 0/50