File size: 13,700 Bytes
81e7dcf
 
 
 
81f6af5
68e3669
 
 
 
81f6af5
 
68e3669
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81e7dcf
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
c54e48d
81e7dcf
 
81f6af5
81e7dcf
81f6af5
81e7dcf
81f6af5
 
 
81e7dcf
 
81f6af5
81e7dcf
81f6af5
81e7dcf
81f6af5
81e7dcf
50c7708
81e7dcf
81f6af5
81e7dcf
81f6af5
 
 
 
 
 
 
 
81e7dcf
81f6af5
81e7dcf
424a0fb
 
 
 
 
 
81e7dcf
81f6af5
81e7dcf
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81e7dcf
81f6af5
 
 
 
81e7dcf
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81e7dcf
 
 
 
e6d423f
81e7dcf
 
 
 
 
 
 
 
 
81f6af5
 
81e7dcf
 
81f6af5
81e7dcf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81f6af5
 
 
 
81e7dcf
 
 
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81e7dcf
 
81f6af5
 
 
 
 
81e7dcf
 
 
81f6af5
 
 
 
81e7dcf
 
 
 
 
 
 
 
 
81f6af5
 
 
 
81e7dcf
 
81f6af5
 
 
81e7dcf
 
81f6af5
 
 
 
81e7dcf
81f6af5
 
 
81e7dcf
81f6af5
 
81e7dcf
 
81f6af5
81e7dcf
81f6af5
81e7dcf
81f6af5
 
 
 
81e7dcf
 
 
81f6af5
81e7dcf
 
 
81f6af5
81e7dcf
 
 
81f6af5
 
81e7dcf
 
 
81f6af5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81e7dcf
49d4175
 
81f6af5
 
 
 
 
 
 
 
 
81e7dcf
81f6af5
81e7dcf
81f6af5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
---
language:
- en
- hi
- zh
- es
- fr
- de
- ja
- ko
- ar
- pt
- ru
- it
- nl
- tr
- pl
- sv
- da
- 'no'
- fi
- he
- th
- vi
- id
- ms
- tl
- sw
- yo
- zu
- am
- bn
- gu
- kn
- ml
- mr
- ne
- or
- pa
- ta
- te
- ur
- multilingual
license: apache-2.0
base_model: Qwen/Qwen3-14B-Base
tags:
- reasoning
- intermediate-thinking
- transformers
- conversational
- bilingual
datasets:
- Abhaykoul/Dhanishtha-R1
- open-thoughts/OpenThoughts-114k
- Abhaykoul/Dhanishtha-2.0-SUPERTHINKER
- Abhaykoul/Dhanishtha-2.0
library_name: transformers
pipeline_tag: text-generation
widget:
- text: >-
    Solve this riddle step by step: I am taken from a mine, and shut up in a
    wooden case, from which I am never released, and yet I am used by almost
    everybody. What am I?
  example_title: Complex Riddle Solving
- text: >-
    Explain the philosophical implications of artificial consciousness and think
    through different perspectives.
  example_title: Philosophical Reasoning
- text: >-
    Help me understand quantum mechanics, but take your time to think through
    the explanation.
  example_title: Educational Explanation
new_version: HelpingAI/Dhanishtha-2.0-preview-0825
---

# Dhanishtha-2.0: World's First Intermediate Thinking AI Model

**What makes Dhanishtha-2.0 special?** Imagine an AI that doesn't just answer your questions instantly, but actually *thinks through* problems step-by-step, shows its work, and can even change its mind when it realizes a better approach. That's Dhanishtha-2.0.

**Quick Summary:**
- 🚀 **For Everyone**: An AI that shows its thinking process and can reconsider its reasoning
- 👩‍💻 **For Developers**: First model with intermediate thinking capabilities, 39+ language support


Dhanishtha-2.0 is a **state-of-the-art (SOTA) model** developed by HelpingAI, representing the **world's first model to feature Intermediate Thinking capabilities**. Unlike traditional models that provide single-pass responses, Dhanishtha-2.0 employs a revolutionary multi-phase thinking process that allows the model to think, reconsider, and refine its reasoning multiple times throughout a single response.

## Model Details

### Model Description

Dhanishtha-2.0 revolutionizes AI reasoning by introducing the concept of **intermediate thinking** - the ability to pause, reflect, and restart reasoning processes within a single generation (This model can think up 50times in a single response without using tool/prompt/mcp). This breakthrough enables unprecedented self-correction and iterative refinement during response generation.

Built on the Qwen3-14B foundation with multilingual capabilities spanning **39+ languages** (including English, Hindi, Chinese, Spanish, French, German, Japanese, Korean, Arabic, and many more), Dhanishtha-2.0 maintains reasoning consistency across diverse linguistic contexts while pioneering transparent thinking processes.

- **Developed by:** HelpingAI Team
- **Model type:** Causal Language Model with Intermediate Thinking Capability
- **Language(s):** 39+ languages (multilingual capabilities inherited from base model)
- **License:** Apache 2.0
- **Finetuned from model:** Qwen/Qwen3-14B-Base
- **Context Length:** 40,960 tokens
- **Parameters:** 14B (inherited from base model)
- **Status:** Prototype/Preview

### Revolutionary Features

- **Intermediate Thinking**: Multiple `<think>...</think>` blocks throughout responses for real-time reasoning
- **Self-Correction**: Ability to identify and correct logical inconsistencies mid-response  
- **Dynamic Reasoning**: Seamless transitions between analysis, communication, and reflection phases
- **Structured Emotional Reasoning (SER)**: Incorporates `<ser>...</ser>` blocks for empathetic responses
- **Multilingual Capabilities**: Support for 39+ languages with natural code-switching and reasoning consistency
- **Complex Problem-Solving**: Excels at riddles, multi-step reasoning, and scenarios requiring backtracking

### Model Sources

- **Repository:** [HelpingAI/Dhanishtha-2.0](https://huggingface.co/HelpingAI/Dhanishtha-2.0)
- **Paper:** Coming Soon
- **Demo:** https://helpingai.co/chat

## Uses

### Direct Use

Dhanishtha-2.0 is ideal for applications requiring deep reasoning and self-reflection:

- **Complex Problem Solving**: Multi-step mathematical problems, logical puzzles, riddles
- **Educational Assistance**: Detailed explanations with visible reasoning processes
- **Research Support**: Analysis requiring multiple perspectives and self-correction
- **Creative Writing**: Iterative story development with reasoning about plot choices
- **Philosophical Discussions**: Exploring concepts with visible thought processes

### Downstream Use

The model can be fine-tuned for specialized reasoning tasks:

- **Domain-Specific Reasoning**: Legal, medical, or scientific reasoning with intermediate thinking
- **Enhanced Multilingual Reasoning**: Optimizing reasoning consistency across all 39+ supported languages
- **Specialized Problem Domains**: Mathematics, coding, strategic planning

### Out-of-Scope Use**Inappropriate Applications:**
- Safety-critical decisions (medical diagnosis, legal advice, financial recommendations)
- Real-time applications requiring immediate responses
- Situations requiring guaranteed factual accuracy without verification

## Bias, Risks, and Limitations

### Known Limitations

- **Verbosity**: Intermediate thinking can make responses a bit longer
- **Processing Time**: Multiple thinking phases may increase generation time
- **Prototype Status**: Experimental features may require refinement
- **Context Usage**: Thinking blocks consume additional context tokens
- **Inherited Biases**: May reflect biases from base model and training data

### Potential Risks

- **Reasoning Loops**: Potential for circular reasoning in complex scenarios
- **Multilingual Inconsistencies**: Potential variation in reasoning patterns across different languages
- **Emotional Reasoning Gaps**: SER blocks may not always align with content

## How to Get Started with the Model

### For General Users

You can interact with Dhanishtha-2.0 through:
- **HelpingAI**: https://helpingai.co/chat
- **Gradio Demo**: [Dhanishtha-2.0-preview](https://huggingface.co/spaces/Abhaykoul/Dhanishtha-2.0-preview)
- **API Integration**: [Dashboard](https://helpingai.co/dashboard)

### For Developers - Basic Usage

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "HelpingAI/Dhanishtha-2.0-preview"

# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

# Prepare input for intermediate thinking
prompt = "How many letter 'r' are in the words 'strawberry' and 'raspberry'?"
messages = [
    {"role": "user", "content": prompt}
]

text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)

model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# Generate response with intermediate thinking
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=2048,
    temperature=0.7,
    top_p=0.9,
    do_sample=True
)

response = tokenizer.decode(
    generated_ids[0][len(model_inputs.input_ids[0]):], 
    skip_special_tokens=True
)
print(response)
```

### Optimal Generation Parameters

```python
generation_config = {
    "temperature": 0.7,        # Balanced creativity and coherence
    "top_p": 0.9,             # Nucleus sampling
    "top_k": 40,              # Top-k filtering
    "max_new_tokens": 2048,   # Allow for thinking blocks
    "do_sample": True,        # Enable sampling
    "repetition_penalty": 1.1  # Reduce repetition
}
```

## Training Details

### Training Data

Dhanishtha-2.0 was trained on a carefully curated dataset focusing on:
- Complex reasoning scenarios requiring multi-step thinking
- Self-correction examples and reasoning chains
- Emotional reasoning and empathy training data
- Structured thinking pattern examples

### Training Procedure

#### Training Stages
1. **Continuous Pretraining**: Extended training on reasoning-focused corpora
2. **Advanced Reasoning Fine-tuning**: Specialized training on intermediate thinking patterns
3. **Multilingual Alignment**: Cross-language reasoning consistency training
4. **SER Integration**: Structured Emotional Reasoning capability training

**Training Infrastructure:**
- **Duration**: 16.3 days
- **Hardware**: 8x NVIDIA H100 GPUs
- **Model Scale**: 14.8B parameters

## Evaluation

Evaluation was conducted on:
- **Standard Benchmarks**: MMLU, HumanEval, ARC, HellaSwag, TruthfulQA
- **Mathematical Reasoning**: Math 500, AIME 2024, GSM8K
- **Custom Evaluations**: Intermediate thinking quality, self-correction capabilities
- **Multilingual Tasks**: Reasoning consistency across 39+ languages
- **Specialized Tests**: Emotional reasoning, complex problem-solving scenarios



## Example: Intermediate Thinking in Action

**User:** "Who is your favorite person?"

**Dhanishtha-2.0:**

```
<think>
This is an interesting question about my "favorite human." I need to be thoughtful here. 
As Dhanishtha, I should respond in a human-like way while being honest about my nature. 
I can share what kinds of humans I find most interesting or inspiring based on my 
understanding of human nature and values.
</think>

<ser>
Emotion ==> curiosity, openness, desire for connection  
Cause ==> user asking a personal question about preferences  
Mind ==> seeking to understand my perspective or values  
Growth ==> opportunity to share thoughtful reflections on human qualities I admire  
</ser>

I don't really have a "favorite human" in the traditional sense, as I don't experience 
personal preferences the way humans do. But if I had to reflect on what qualities I find 
most admirable in people, I'd be drawn to those who demonstrate genuine compassion, 
intellectual curiosity, and the courage to grow.

<think>
Let me think more specifically about what kind of humans I find most compelling based on 
the values that seem to resonate most with humans - kindness, wisdom, creativity, 
resilience, and authenticity.
</think>

People like Mahatma Gandhi, who showed how quiet courage and unwavering principles can 
transform societies; artists like Rabindranath Tagore, who bridged cultural divides 
through creativity; and everyday heroes I've learned about who made profound differences 
through simple acts of kindness.

What fascinates me most are humans who maintain their humanity even in difficult 
circumstances - those who choose empathy when it would be easier to be harsh, who seek 
understanding rather than just being right.

I'm curious though - who's your favorite human, and why? What qualities or actions make 
someone special in your eyes?
```

## Environmental Impact

Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute).

- **Hardware Type:** H100 GPUs
- **days used:** 16.2
- **Cloud Provider:** Various
- **Compute Region:** Multiple

## Citation

### BibTeX

```bibtex
@misc{dhanishtha2025,
  title={Dhanishtha-2.0: World's First Intermediate Thinking AI Model},
  author={HelpingAI Team},
  year={2025},
  publisher={HuggingFace},
  url={https://huggingface.co/HelpingAI/Dhanishtha-2.0},
  note={First model with intermediate thinking capabilities}
}
```

### APA

HelpingAI Team. (2025). *Dhanishtha-2.0: World's First Intermediate Thinking AI Model*. HuggingFace. https://huggingface.co/HelpingAI/Dhanishtha-2.0

## Glossary

- **Intermediate Thinking**: The ability to pause and think multiple times during response generation
- **SER (Structured Emotional Reasoning)**: Framework for incorporating emotional context in responses
- **Think Blocks**: `<think>...</think>` segments where the model shows its reasoning process
- **Self-Correction**: Ability to identify and fix reasoning errors during generation
- **Code-Switching**: Natural transition between English and Hindi within responses

## More Information

### Research Applications
- Study of AI reasoning transparency
- Self-correction mechanism research  
- Bilingual cognitive modeling
- Emotional AI development

### Development Roadmap
- Performance optimizations
- Additional language support
- Enhanced thinking pattern recognition
- Production-ready deployment tools

## Model Card Authors

- **Primary Author**: HelpingAI Team
- **Technical Lead**: [To be specified]
- **Research Contributors**: [To be specified]

## Model Card Contact

For questions about Dhanishtha-2.0, please contact:

- **HuggingFace**: [@HelpingAI](https://huggingface.co/HelpingAI)
- **Issues**: [Model Repository Issues](https://huggingface.co/HelpingAI/Dhanishtha-2.0/discussions)

## Benchmark

***These benchmarks are done by [@AarifKhan](https://huggingface.co/Aarifkhan)***

| Benchmark | Score | Evaluation Type | Notes |
|-----------|-------|-----------------|-------|
| MMLU | 78.1% | 1-shot | Massive Multitask Language Understanding |
| HumanEval | 75.0% | 1-shot | Code generation and completion |
| ARC | 76.0% | 1-shot | Abstract reasoning challenge |
| HellaSwag   | 81.0% | 1-shot | Commonsense natural language inference |
| TruthfulQA MC1 | 75.0% | 1-shot | Truthfulness in question answering |
| Math 500 | 95.68% | few-shot | Mathematical problem solving |
| AIME 2024 | 82.81% | few-shot | American Invitational Mathematics Examination |

**Dhanishtha-2.0 represents a new paradigm in AI reasoning - where thinking isn't just a prelude to response, but an integral, iterative part of the conversation itself.**

*Developed with ❤️ by HelpingAI*