SUTRA’s dual-transformer approach extends the power of both MoE and Dense AI language model architectures, delivering cost-efficient multilingual capabilities for over 50+ languages. It powers scalable AI applications for conversation, search, and advanced reasoning, ensuring high-performance across diverse languages, domains and applications.

Multilingual

Ultrafast response-generation and instruction-following in 50+ languages. Lightweight Dual2 architecture and purpose-built tokenizer enables engaging and personalized conversational experiences with unparalleled inference efficiency.

Online

Connected and hallucination-free models that provide accurate and up-to-date responses in a conversational tone. Secure, on-premise deployment with seamless integration to structured or unstructured enterprise databases and online results.

Thinking (NEW)

Models that delivers deeper, structured thinking across topics and domains. Building on our multilingual AI advances, reasoning models bring complex decision-making, supports multiple languages, and uses resources efficiently.

50+ Languages With Accuracy & Efficiency /

SUTRA surpasses leading models by 20-25% on the MMLU benchmark in comprehending and generating responses across numerous languages.

ENGLISH

English

GPT 4o

82

LLAMA 3.1

82

SUTRA

81

ENGLISH

English

GPT 4o

82

SUTRA

81

HINDI

हिंदी

GPT 4o

77

Sarvam-1

48

SUTRA

78

HINDI

हिंदी

GPT 4o

77

SUTRA

78

KOREAN

한국어

GPT 4o

65

HyperClovaX

54

SUTRA

71

KOREAN

한국어

GPT 4o

65

SUTRA

71

GUJARATI

ગુજરાતી

GPT 4o

58

LLAMA 3.1

54

SUTRA

72

GUJARATI

ગુજરાતી

GPT 4o

58

SUTRA

72

JAPANESE

日本語

GPT 4o

71

SAKANA

62

SUTRA

76

JAPANESE

日本語

GPT 4o

71

SUTRA

76

ARABIC

العربية

GPT 4o

65

LLAMA 3.1

60

SUTRA

71

ARABIC

العربية

GPT 4o

65

SUTRA

71

Cost-efficient Tokenization

for non-English Languages /

SUTRA's purpose-built tokenizer is more efficient across languages, enabling compute cost-savings and ultrafast inference for multilingual applications.

Evaluating Tokenizer Performance of Large Language Models Across Official Indian Languages (Published November 24)
An independent study (https://arxiv.org/abs/2411.12240) shows SUTRA outperforms GPT-4o, Llama 3.1, and other Indian language models across 14+ Indian languages—delivering superior performance with better token efficiency and cost-effectiveness.

*Currently supported languages are English (en), Hindi (hi), Gujarati (gu), Korean (ko), Japanese (ja), Bengali (bn), Marathi (mr), Telugu (te), Tamil (ta), Kannada (kn), Malayalam (ml), Punjabi (pa), Odia (or), Urdu (ur), , Sanskrit (sa), Assamese (as), Thai (th), Arabic (ar), Persian (fa), Vietnamese (vi), Indonesian (id), Turkish (tr), Polish (pl), Russian (ru), Ukrainian (uk), Dutch (nl), French (fr), Italian (it), Spanish (es), German (de), Greek (el), Hebrew (he), Portuguese (pt). More languages are coming soon.

Token Counts

GPT

SUTRA

Token Counts

GPT

SUTRA

હાય, હું સૂત્ર છું, હું બહુભાષી અને અલ્ટ્રાફાસ્ટ AI મોડેલ છું. તમને મળીને આનંદ થયો.

5.7x fewer tokens than GPT

131

23

હાય, હું સૂત્ર છું, હું બહુભાષી અને અલ્ટ્રાફાસ્ટ AI મોડેલ છું. તમને મળીને આનંદ થયો.

5.7x fewer tokens than GPT

131

23

नमस्ते, मैं सूत्र हूं, मैं एक बहुभाषी और अल्ट्राफास्ट एआई मॉडल हूं। आप से मिल कर अच्छा लगा।

4.1x fewer tokens than GPT

94

23

नमस्ते, मैं सूत्र हूं, मैं एक बहुभाषी और अल्ट्राफास्ट एआई मॉडल हूं। आप से मिल कर अच्छा लगा।

4.1x fewer tokens than GPT

94

23

안녕하세요. 저는 다국어 초고속 AI 모델 SUTRA입니다. 만나서 반가워요.

1.8x fewer tokens than GPT

43

24

안녕하세요. 저는 다국어 초고속 AI 모델 SUTRA입니다. 만나서 반가워요.

1.8x fewer tokens than GPT

43

24

ஹாய், நான் சூத்ரா, நான் ஒரு பன்மொழி மற்றும் அதிவேக AI மாடல். உங்களை சந்தித்ததில் மகிழ்ச்சி.

5.1x fewer tokens than GPT

117

23

ஹாய், நான் சூத்ரா, நான் ஒரு பன்மொழி மற்றும் அதிவேக AI மாடல். உங்களை சந்தித்ததில் மகிழ்ச்சி.

5.1x fewer tokens than GPT

117

23

مرحبًا، أنا SUTRA، أنا نموذج ذكاء اصطناعي متعدد اللغات وفائق السرعة. تشرفنا.

2.9x fewer tokens than GPT

61

21

مرحبًا، أنا SUTRA، أنا نموذج ذكاء اصطناعي متعدد اللغات وفائق السرعة. تشرفنا.

2.9x fewer tokens than GPT

61

21

हाय, मी सूत्र आहे, मी एक बहुभाषिक आणि अल्ट्राफास्ट AI मॉडेल आहे. तुम्हाला भेटून छान वाटले.

3.8x fewer tokens than GPT

88

23

हाय, मी सूत्र आहे, मी एक बहुभाषिक आणि अल्ट्राफास्ट AI मॉडेल आहे. तुम्हाला भेटून छान वाटले.

3.8x fewer tokens than GPT

88

23

こんにちは、SUTRA です。多言語対応の超高速 AI モデルです。お会い出来て嬉しいです。

1.6x fewer tokens than GPT

40

25

こんにちは、SUTRA です。多言語対応の超高速 AI モデルです。お会い出来て嬉しいです。

1.6x fewer tokens than GPT

40

25

Hi, I’m SUTRA, I am a multilingual and ultrafast AI model. It's nice to meet you.

Similar number of tokens as GPT

26

26

Hi, I’m SUTRA, I am a multilingual and ultrafast AI model. It's nice to meet you.

Similar number of tokens as GPT

26

26

Up-to-date and

Hallucination-free /

SUTRA-Online are internet connected and hallucination-free models that understand queries, browse the web, and summarize information to provide current answers.


SUTRA-Online can answer queries like “Who won the game last night” or “What’s the current stock price” accurately, whereas offline models suffer from knowledge cut-off dates.

Models /

SUTRA-V1

SUTRA-V1 is our multilingual AI model designed for instruction execution and conversational intelligence across 50+ languages. Trained on a diverse blend of proprietary and open-access datasets, it excels in handling complex tasks with high accuracy. With deep proficiency across Latin, Indic, and Far Eastern languages, SUTRA-V1 delivers natural, context-aware responses, making it a powerful solution for global enterprises, multilingual assistants, and cross-language AI applications.

Architecture

Dense

Architecture

Dense

Parameters

73B

Parameters

73B

SUTRA-R0 is the first in our series of advanced reasoning models, designed for complex problem-solving and deep contextual understanding. Built to analyze, infer, and generate logical responses, SUTRA-R0 goes beyond pattern recognition—applying structured reasoning to tackle nuanced queries, multi-step problem-solving, and enterprise decision-making. Its architecture enables high-accuracy responses across domains, making it a powerful tool for knowledge-intensive workflows and next-generation AI applications.

Architecture

Dense

Architecture

Dense

Parameters

8B/36B

Parameters

8B/36B

SUTRA-P0 (Coming Soon)

SUTRA-P0 is a time-series predictive AI model built on a dense architecture, designed for forecasting and analytical applications beyond text and chat. Optimized for dynamic data patterns, SUTRA-P0 enables enterprises to predict trends, detect anomalies, and make proactive decisions across finance, logistics, healthcare, and other data-driven domains. Its high-precision forecasting capabilities make it a powerful tool for businesses seeking deeper insights and future-ready strategies.

Architecture

Dense

Architecture

Dense

Parameters

1.5B

Parameters

1.5B

SUTRA for Developers /

SUTRA models are the technology backbone of TWO’s products and services and are available as Model as a Service (MaaS) to other apps and services via usage-based pricing and simple-to-integrate APIs.

Developer Portal

Access our all-in-one SUTRA resource hub for developers

API Reference

Learn how to start building using simple to use SUTRA API

Playground

Experience the power of SUTRA and compare with other LLMs