File size: 3,634 Bytes
9fb010d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
657603e
9fb010d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50328b0
 
 
 
9fb010d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dfddc11
9fb010d
 
 
 
 
 
 
 
dfddc11
9fb010d
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
---
language:
- en
license: llama3
tags:
- Llama-3
- RL
- Atropos
- Tool Calling
- Nous Research
- instruct
- finetune
- reasoning
- function calling
- transformers
- reinforcement-learning
- json mode
- chatml
base_model: meta-llama/Meta-Llama-3.1-8B
library_name: transformers

---

# DeepHermes Tool Calling Specialist - Atropos RL

## Model Overview

The **DeepHermes Tool Calling Specialist - Atropos RL** model is an experimental artifact fine-tuned by Nous Research using our innovative open-source reinforcement learning framework—Atropos. This variant specifically improves the tool calling performance of the **DeepHermes 3 Llama-3.1 8B** model during its reasoning mode.

**Note**: This model is intended as an experimental artifact and is not designed for broad, general-purpose use.

## Atropos Open Source Framework

Atropos is Nous Research’s open-source Reinforcement Learning environment stack, designed to enhance various aspects of LLM functionalities through structured RL methodologies. We encourage contributions and exploration:

🔗 [Atropos GitHub Repository](https://github.com/NousResearch/Atropos)

## Benchmark Results

Evaluations on the Berkeley Function Calling benchmark demonstrate significant improvements in tool calling accuracy during reasoning mode, compared to its base model:

| Benchmark | Base Accuracy | Atropos RL Accuracy | Improvement |
| --------- | ------------- | ------------------- | ----------- |
| Parallel  | 0.10          | 0.46                | **4.6x**    |
| Simple    | 0.21          | 0.5175              | **2.5x**    |

These enhancements are due to RL fine-tuning specifically targeted at improving reasoning-based tool calling capabilities.

Eval set accuracy results: 

![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/BWIFtF7oBWO2b_Q1oh4YM.png)

## Key Features

- **Improved Tool Calling in Reasoning Mode**: Reinforcement learning significantly boosts tool usage during complex reasoning tasks.
- **Open-Source RL Framework**: Utilizes the fully open-source Atropos RL Environments.
- **Active Open Source Community**: Contributions welcomed on the Atropos GitHub.
- **Upcoming SOTA RL Trainer**: A state-of-the-art open-source reinforcement learning trainer by Nous Research is coming soon.

## Usage

This model supports multiple inference modes including:

- **Reasoning (Deep Thinking Mode)**
- **Standard Chat/Instruction Mode**
- **Structured JSON Outputs**
- **Function Calling**

Detailed documentation and example inference code are available:

*Note: You must first place DeepHermes' reasoning system prompt, and then append  your function calling system prompt after for it to do reasoning and tool calling simultaneously.*

🔗 [Hermes Function Calling GitHub](https://github.com/NousResearch/Hermes-Function-Calling)

## How to Cite

```bibtex
@misc{
      title={DeepHermes Tool Calling Specialist - Atropos RL},
      author={Teknium and Dakota Mahan and Roger Jin and Chen Guang and Jai Suphavadeeprasit and Jeffrey Quesnelle},
      year={2025},
      url={https://huggingface.co/NousResearch/DeepHermes-Tool-Calling-Specialist-Atropos-RL}
}
```

## Community and Support

For questions, issues, or findings, please open issues or discussions in the respective GitHub repositories:

- [Atropos Framework Issues](https://github.com/NousResearch/Atropos/issues)
- [DeepHermes Models Issues](https://github.com/NousResearch/Hermes-Function-Calling/issues)

Nous Research encourages active community engagement and open-source contributions to continuously improve model performance and capabilities.