Indian Startup Sarvam Launches Sarvam-M, a 24B-Parameter Open-Source Language Model with Advanced Reasoning Abilities

Sarvam AI, a rising player in India’s artificial intelligence landscape, has officially launched its flagship large language model—Sarvam-M. Built on top of Mistral Small, Sarvam-M is a 24-billion-parameter hybrid model with open weights, designed to tackle a wide spectrum of tasks including mathematics, programming, multilingual processing, and Indian language reasoning.

According to the company, Sarvam-M represents a significant advancement in conversational AI, machine translation, and educational applications. It’s already making headlines for setting new standards in math problem-solving, code generation, and Indian language benchmarks.

A Three-Phase Training Framework

The development of Sarvam-M involved a comprehensive three-phase training pipeline:

  1. Supervised Fine-Tuning (SFT):
    Sarvam curated a robust dataset of diverse, high-quality prompts and generated completions using approved models. These outputs were scored, filtered, and fine-tuned to optimize cultural alignment and reduce bias. The model was trained to operate in both “think” mode for complex reasoning and “non-think” mode for general conversation.
  2. Reinforcement Learning with Verifiable Rewards (RLVR):
    This phase incorporated programming tasks, mathematical reasoning, and instruction-following datasets. Through custom reward engineering and prompt sampling, Sarvam-M was further optimized to follow instructions accurately and reason effectively.
  3. Inference Optimization:
    The team used FP8 quantization to reduce computational overhead with minimal loss in model accuracy. Innovations like lookahead decoding improved throughput, though challenges with higher concurrency were noted.
Benchmark Performance

Sarvam-M has demonstrated outstanding results in key benchmarks:

  • On Indian language reasoning tasks like the romanised GSM-8K, it achieved a +86% improvement.
  • It outperformed Llama-4 Scout and was found to be on par with much larger models like Llama 3.3 70B and Gemma 3 27B.
  • On English knowledge benchmarks such as MMLU, it showed a modest drop of around 1%, which the team considers a reasonable trade-off given its multilingual and reasoning strengths.
Accessibility and Applications

Sarvam-M is available for developers and researchers via Sarvam’s official API and can also be downloaded from Hugging Face for further experimentation and integration. The model is expected to empower a wide array of applications ranging from education and enterprise solutions to local language assistants and voice-based interfaces.

This release comes as part of Sarvam AI’s broader goal to build India’s first sovereign LLM—a step toward democratizing cutting-edge AI technology while tailoring it to local needs and linguistic diversity.

- Advertisement -

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles

error: Content is protected !!

Share your details to download the Cybersecurity Report 2025

Share your details to download the CISO Handbook 2025

Sign Up for CXO Digital Pulse Newsletters

Share your details to download the Research Report

Share your details to download the Coffee Table Book

Share your details to download the Vision 2023 Research Report

Download 8 Key Insights for Manufacturing for 2023 Report

Sign Up for CISO Handbook 2023

Download India’s Cybersecurity Outlook 2023 Report

Unlock Exclusive Insights: Access the article

Download CIO VISION 2024 Report

Share your details to download the report

Share your details to download the CISO Handbook 2024

Fill your details to Watch