Sarvam’s 105-bn model puts India on the frontier AI map
In a significant push to strengthen India’s advanced AI capabilities, homegrown startup Sarvam launched a 105-billion-parameter foundational large language model (LLM), along with a suite of tools designed for commercial use. Co-founder Vivek Raghavan discussed the company’s progress in Indic languages and its expansion into AI-powered devices.
Sarvam’s 105-billion-parameter AI model places it in the frontier category. Architecturally, what differentiates it from global models beyond just size?
This is the largest model trained from scratch in India, with zero external data dependency and a strong grounding in Indian knowledge. While it is a global model, it is built with Indian contexts in mind. Models like Gemini or ChatGPT are still an order of magnitude larger, so we are not claiming parity in scale. However, being smaller makes our models more efficient and cost-effective. For most real-world and agentic use cases, models of this size deliver excellent results without requiring extreme scale.
How deeply is the model trained in Indic languages? Can it outperform global models in low-resource languages?
We are far more focused on Indian languages than most global labs. Among models of comparable size, we are superior in Indian languages. It is not fair to compare systems tens of times larger, but within our size category, we are stronger. What specific progress did you make in Indian languages? We believe Indians will experience AI primarily through voice. In speech recognition across Indian languages and dialects, we believe we are the best in the world. Our newer models are world-class in natural speech synthesis. We also released a small vision model that outperforms much larger systems at extracting Indic scripts from documents and images. Among similarly sized LLMs, we are equivalent or better than most. For instance, we outperformed a DeepSeek model released last year, and even compared favourably with a version 6 times larger. Our goal is to lead globally within our size class, especially in Indian language and domain-specific contexts.
Was the model trained entirely on domestic infrastructure? How will you make inference affordable at scale?
Yes, it was trained entirely in India under the India AI mission, using concessional GPUs and no external data dependency. Inference is a separate challenge. Training does not guarantee adoption. We will enable access to our models, but competing in pure B2C is difficult when global players offer free services after spending billions. That is a structural reality.
You’re moving beyond traditional mobile platforms. What’s the strategy behind expanding into devices like smart glasses and feature phones?
AI will change interfaces. We see smart glasses as business devices for recording conversations, analytics and coaching, with voice at the centre. Feature phone integrations are about inclusion—enabling users to access AI in their own language. We also aim to run small models directly on devices to reduce costs and cloud dependency. The focus is inclusion and more natural interaction.
. How do you address hallucination and cultural bias?
LLMs are probabilistic, so hallucinations are inherent. The solution is to build grounding mechanisms around them. No real-world system uses a raw LLM. The aim is to minimise risk, though some residual risk always exists.
What is the next milestone for Indic LLMs?
We were the first to highlight the “tokenisation tax” in Indian languages. If tokenisation improves, costs will fall. However, global labs are investing billions and offering models for free, often using free-user data to train systems. That dynamic is unprecedented and shapes the competitive landscape.
This is the largest model trained from scratch in India, with zero external data dependency and a strong grounding in Indian knowledge. While it is a global model, it is built with Indian contexts in mind. Models like Gemini or ChatGPT are still an order of magnitude larger, so we are not claiming parity in scale. However, being smaller makes our models more efficient and cost-effective. For most real-world and agentic use cases, models of this size deliver excellent results without requiring extreme scale.
We are far more focused on Indian languages than most global labs. Among models of comparable size, we are superior in Indian languages. It is not fair to compare systems tens of times larger, but within our size category, we are stronger. What specific progress did you make in Indian languages? We believe Indians will experience AI primarily through voice. In speech recognition across Indian languages and dialects, we believe we are the best in the world. Our newer models are world-class in natural speech synthesis. We also released a small vision model that outperforms much larger systems at extracting Indic scripts from documents and images. Among similarly sized LLMs, we are equivalent or better than most. For instance, we outperformed a DeepSeek model released last year, and even compared favourably with a version 6 times larger. Our goal is to lead globally within our size class, especially in Indian language and domain-specific contexts.
Was the model trained entirely on domestic infrastructure? How will you make inference affordable at scale?
Yes, it was trained entirely in India under the India AI mission, using concessional GPUs and no external data dependency. Inference is a separate challenge. Training does not guarantee adoption. We will enable access to our models, but competing in pure B2C is difficult when global players offer free services after spending billions. That is a structural reality.
AI will change interfaces. We see smart glasses as business devices for recording conversations, analytics and coaching, with voice at the centre. Feature phone integrations are about inclusion—enabling users to access AI in their own language. We also aim to run small models directly on devices to reduce costs and cloud dependency. The focus is inclusion and more natural interaction.
LLMs are probabilistic, so hallucinations are inherent. The solution is to build grounding mechanisms around them. No real-world system uses a raw LLM. The aim is to minimise risk, though some residual risk always exists.
What is the next milestone for Indic LLMs?
We were the first to highlight the “tokenisation tax” in Indian languages. If tokenisation improves, costs will fall. However, global labs are investing billions and offering models for free, often using free-user data to train systems. That dynamic is unprecedented and shapes the competitive landscape.
Popular from Technology
- After CEO Julie Sweet’s ‘exit’ warning, Accenture HR tells senior employees: To get promoted to leadership roles, you would require..
- Explained: What is India’s Sarvam AI model that Google CEO Sundar Pichai is quite impressed with
- Did Insta, YouTube deliberately build addictive tools to trap kids?
- Jeff Bezos' ex-wife MacKenzie Scott, who has donated about 58 million Amazon shares, on who she thinks of every time she makes a huge donation
- Google’s ex CEO Eric Schmidt warns America: We are running out of …
end of article
Trending Stories
- 2026 Winter Olympics women single skating free skating: Amber Glenn rises from 13th place with stunning 147.52 free skate reaction
- Detroit Pistons vs New York Knicks injury report: Who's playing, injured and questionable players, head-to-head records, team stats, and more (February 19, 2026)
- Donna Kelce breaks silence on Travis Kelce and Taylor Swift wedding as secrecy and NDA drama spark questions
- "My very own...": Olympian Chloe Kim gushes over Myles Garrett as he becomes more than a loving boyfriend amid her silver medal win
- 2026 Winter Olympics Live: Sidney Crosby injury, Norway sweep Nordic combined, Jordan Stolz chases third gold, Mikaela Shiffrin wins and Day 13 updates
- Atlanta Hawks vs Philadelphia 76ers injury report: Who's playing, injured and questionable players, head-to-head records, team stats, and more (February 19, 2026)
- Sidney Crosby's injury puts Canada on alert with Connor McDavid ready to take charge for the Olympic semifinal match
Featured in technology
- Sony shuts down Bluepoint Games, Developer of Demon’s Souls remake
- Quote of the day by Mark Zuckerberg: “Our society needs more heroes who are scientists, researchers, and engineers. We need to celebrate and reward the people who cure diseases, expand our understanding of humanity and work to improve people’s lives.”
- After Disney and Paramount, Netflix sends legal notice to Bytedance: Read full cease-and-desist letter
- Google releases Gemini 3.1 Pro: Here's what's new and who gets it first
- Texas sues TP-Link for allegedly allowing China to hack into routers, says: Nearly all of its products …
- After CEO Julie Sweet’s ‘exit’ warning, Accenture HR tells senior employees: To get promoted to leadership roles, you would require..
Photostories
- 7 elegant blooms known for their long-lasting freshness
- How to make Chettinad Omelette for breakfast at home
- The 50: Prince Narula, Mr Faisu, Rajat Dalal to Shiv Thakare: Meet the Top 12 contestants of the reality show
- 5 Vastu-approved fish to keep in your home aquarium for luck and prosperity
- Chef Sanjeev Kapoor's microwave cleaning tips will make your kitchen life easier
- 5 times Bad Bunny made headlines
- What’s it like to visit Jhalana Leopard Safari Park in Rajasthan?
- Top 5 shows to watch on OTT before ‘House of the Dragon Season 3’ arrives
- 7 New High-Speed Rail Corridors! On Which Routes Will Bullet Trains Run In India? Check Cities, Travel Time - Top Details
- Ramadan 2026: How to make Chicken Malai Tikka in an air fryer in under 20 minutes
Up Next
Start a Conversation
Post comment