File size: 3,409 Bytes
f76e17c
 
 
 
 
5865c7c
f76e17c
 
 
 
 
 
 
 
 
 
 
 
 
5865c7c
 
 
82fffe4
5865c7c
 
 
 
37a7426
f76e17c
 
 
 
 
 
 
 
 
 
 
5865c7c
f76e17c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19b1ec1
f76e17c
 
 
 
 
 
 
 
 
19b1ec1
f76e17c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19b1ec1
f76e17c
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
---
license: apache-2.0
language:
- en
base_model:
- janhq/Jan-v1-2509
pipeline_tag: text-generation
library_name: transformers
---
# Jan-v1: Advanced Agentic Language Model

[![GitHub](https://img.shields.io/badge/GitHub-Repository-blue?logo=github)](https://github.com/menloresearch/deep-research) 
[![License](https://img.shields.io/badge/License-Apache%202.0-yellow)](https://opensource.org/licenses/Apache-2.0)
[![Jan App](https://img.shields.io/badge/Powered%20by-Jan%20App-purple?style=flat&logo=android)](https://jan.ai/) 

<!-- Optional: If you have a GIF for Jan-v1, include it here like Lucy's. -->
<!-- ![image/gif](jan_v1_demo.gif) -->

## Overview

### Update: **Jan-v1-2509**

We have released a small weight update, **jan-v1-2509**, which refines the original v1.

* No architectural changes.
* Slightly **lower performance on SimpleQA** compared to jan-v1.
* **Slightly mproved results on other chat benchmarks** and overall more **reliable**

**Jan-v1** is the first release in the **Jan Family**, designed for agentic reasoning and problem-solving within the [Jan App](https://jan.ai/). Based on our [**Lucy**](https://huggingface.co/Menlo/Lucy) model, Jan-v1 achieves improved performance through model scaling.

Jan-v1 uses the [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks.

## Performance

### Question Answering (SimpleQA) 
For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy.

![image/png](https://cdn-uploads.huggingface.co/production/uploads/655e3b59d5c0d3db5359ca3c/B5OlbTP3L6Sn6iT9fA2zg.png)

*The 91.1% SimpleQA accuracy with Jan-v1 remains a highlight, though Jan-v1-2509 focuses on balancing factual QA with improved reliability across chat-based reasoning tasks.*

### Chat Benchmarks

These benchmarks evaluate the model's conversational and instructional capabilities.

![image/png](https://cdn-uploads.huggingface.co/production/uploads/655e3b59d5c0d3db5359ca3c/9EjBc6MEjpcItul6sDYkh.png)

## Quick Start

### Integration with Jan App

Jan-v1 is optimized for direct integration with the [Jan App](https://jan.ai/). Simply select the model from the Jan App interface for immediate access to its full capabilities.

### Local Deployment

**Using vLLM:**
```bash
vllm serve janhq/Jan-v1-2509 \
    --host 0.0.0.0 \
    --port 1234 \
    --enable-auto-tool-choice \
    --tool-call-parser hermes
    
```

**Using llama.cpp:**
```bash
llama-server --model Jan-v1-2509-Q4_K_M.gguf \
    --host 0.0.0.0 \
    --port 1234 \
    --jinja \
    --no-context-shift
```

### Recommended Parameters

```yaml
temperature: 0.6
top_p: 0.95
top_k: 20
min_p: 0.0
max_tokens: 2048
```


## 🤝 Community & Support

- **Discussions**: [HuggingFace Community](https://huggingface.co/janhq/Jan-v1-2509/discussions) 
- **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/)

## (*) Note
By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file [chat_template_raw.jinja](https://huggingface.co/janhq/Jan-v1-4B/blob/main/chat_template_raw.jinja).

## 📄 Citation
```bibtex
Updated Soon
```
---