Welcome to Maruth Labs

We build efficient language models that deliver high performance without unnecessary complexity. Discover our breakthrough model Madhuram and explore how our innovative design sets a new benchmark in AI efficiency.

About Maruth Labs

Maruth Labs is an AI research company based in Delhi, India, committed to revolutionizing language model technology. Our approach focuses on creating ultra-efficient models that perform at the cutting edge without the need for massive computational resources.

Our flagship model, Madhuram, demonstrates that intelligent design can replace sheer scale. By leveraging innovative engineering and rigorous research, we deliver models that are powerful and highly adaptable for deployment on mobile and wearable devices.

At Maruth Labs, our mission, our identity, and the cornerstone of our product roadmap is ensuring our focus remains unwavering. Our team and business model are aligned to achieve efficiency. We tackle performance and resource management simultaneously, solving these challenges through innovative engineering and research breakthroughs. Our aim is to enhance capabilities while ensuring that efficiency and accessibility are prioritized. This way, we can innovate freely.

We believe the future of AI lies in efficiency, accessibility, and responsible resource management. Join us as we shape the future of language models and explore the endless possibilities of smarter, leaner AI.

Benchmarks

Introducing Madhuram—our breakthrough compact language model designed to deliver high performance with minimal resources. With just 74 million parameters and trained on 150 billion tokens, Madhuram proves that intelligent design can outperform massive-scale models.

Note: Madhuram-β (Beta) has been trained on only 150 billion tokens, while the other models in the comparison have been exposed to much larger datasets.

Optimized for mobile and wearable devices, Madhuram sets a new benchmark in AI efficiency without compromising on quality or speed.

Analysis

The chart displays key performance metrics for our models in comparison with GPT3-small across various benchmarks.
  • Madhuram-β (Zero-shot): Delivers robust results even without additional context.
  • GPT3-Small (Zero-shot): Serves as a reliable baseline for comparison.
  • GPT3-Small (Few-shot): Shows improved accuracy with extra examples.

Overall, the data reinforces that efficient model design can yield competitive performance without enormous scale.

* Madhuram-β was evaluated in standard mode using the same configuration as its production version for reproducibility.
* Comparative data from OpenAI GPT3 was sourced from OpenAI's research paper.

Our benchmark results demonstrate that Madhuram holds its own against larger models while using a fraction of the resources. The data highlights strong zero-shot performance, proving the efficacy of our compact design.

Whether deployed on mobile devices or in resource-constrained environments, Madhuram delivers reliable and consistent results, paving the way for more accessible and efficient AI applications.

* Comparative data from was sourced from Meta's MobileLLM Github Repository.