Commit
·
3ea99f4
verified
·
0
Parent(s):
Super-squash branch 'main' using huggingface_hub
Browse filesCo-authored-by: patrickvonplaten <[email protected]>
- .gitattributes +40 -0
- README.md +108 -0
- devstral.gguf +3 -0
- devstralQ4_0.gguf +3 -0
- devstralQ4_K_M.gguf +3 -0
- devstralQ5_K_M.gguf +3 -0
- devstralQ8_0.gguf +3 -0
.gitattributes
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
devstralQ8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
devstralQ4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
devstral.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
devstralQ4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
devstralQ5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,108 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
- fr
|
5 |
+
- de
|
6 |
+
- es
|
7 |
+
- pt
|
8 |
+
- it
|
9 |
+
- ja
|
10 |
+
- ko
|
11 |
+
- ru
|
12 |
+
- zh
|
13 |
+
- ar
|
14 |
+
- fa
|
15 |
+
- id
|
16 |
+
- ms
|
17 |
+
- ne
|
18 |
+
- pl
|
19 |
+
- ro
|
20 |
+
- sr
|
21 |
+
- sv
|
22 |
+
- tr
|
23 |
+
- uk
|
24 |
+
- vi
|
25 |
+
- hi
|
26 |
+
- bn
|
27 |
+
license: apache-2.0
|
28 |
+
library_name: llmstudio
|
29 |
+
inference: false
|
30 |
+
base_model:
|
31 |
+
- mistralai/Devstrall-Small-2505_gguf
|
32 |
+
extra_gated_description: >-
|
33 |
+
If you want to learn more about how we process your personal data, please read
|
34 |
+
our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
|
35 |
+
pipeline_tag: text2text-generation
|
36 |
+
---
|
37 |
+
|
38 |
+
|
39 |
+
> [!Note]
|
40 |
+
> At Mistral, we don't yet have too much experience with providing GGUF-quantized checkpoints
|
41 |
+
> to the community, but want to help improving the ecosystem going forward.
|
42 |
+
> If you encounter any problems with the provided checkpoints here, please open a discussion or pull request
|
43 |
+
|
44 |
+
|
45 |
+
# Devstral-Small-2505 (gguf)
|
46 |
+
|
47 |
+
Devstral is an agentic LLM for software engineering tasks built under a collaboration between [Mistral AI](https://mistral.ai/) and [All Hands AI](https://www.all-hands.dev/) 🙌. Devstral excels at using tools to explore codebases, editing multiple files and power software engineering agents. The model achieves remarkable performance on SWE-bench which positionates it as the #1 open source model on this [benchmark](https://huggingface.co/mistralai/Devstral-Small-2505#benchmark-results).
|
48 |
+
|
49 |
+
This is the GGUF version of the [Devstral-Small-2505](https://huggingface.co/mistralai/Devstral-Small-2505) model. We released the BF16 weights as well as the following quantized format:
|
50 |
+
- Q8_0 (recommended)
|
51 |
+
- Q5_K_M (recommended)
|
52 |
+
- Q4_K_M (recommended)
|
53 |
+
- Q4_0
|
54 |
+
|
55 |
+
It is finetuned from [Mistral-Small-3.1](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503), therefore it has a long context window of up to 128k tokens. As a coding agent, Devstral is text-only and before fine-tuning from `Mistral-Small-3.1` the vision encoder was removed.
|
56 |
+
|
57 |
+
For enterprises requiring specialized capabilities (increased context, domain-specific knowledge, etc.), we will release commercial models beyond what Mistral AI contributes to the community.
|
58 |
+
|
59 |
+
Learn more about Devstral in our [blog post](https://mistral.ai/news/devstral).
|
60 |
+
|
61 |
+
|
62 |
+
## Key Features:
|
63 |
+
- **Agentic coding**: Devstral is designed to excel at agentic coding tasks, making it a great choice for software engineering agents.
|
64 |
+
- **lightweight**: with its compact size of just 24 billion parameters, Devstral is light enough to run on a single RTX 4090 or a Mac with 32GB RAM, making it an appropriate model for local deployment and on-device use.
|
65 |
+
- **Apache 2.0 License**: Open license allowing usage and modification for both commercial and non-commercial purposes.
|
66 |
+
- **Context Window**: A 128k context window.
|
67 |
+
- **Tokenizer**: Utilizes a Tekken tokenizer with a 131k vocabulary size.
|
68 |
+
|
69 |
+
## Usage
|
70 |
+
|
71 |
+
We recommend to use Devstral with the [OpenHands](https://github.com/All-Hands-AI/OpenHands/tree/main) scaffold as explained [here](https://huggingface.co/mistralai/Devstral-Small-2505#usage).
|
72 |
+
To use it local with a GGUF-quantized checkpoint, see the following section.
|
73 |
+
|
74 |
+
### Local inference (GGUF)
|
75 |
+
|
76 |
+
Download the weights from huggingface:
|
77 |
+
|
78 |
+
```
|
79 |
+
pip install -U "huggingface_hub[cli]"
|
80 |
+
huggingface-cli download \
|
81 |
+
"mistralai/Devstral-Small-2505_gguf" \
|
82 |
+
--include "devstralQ4_K_M.gguf" \
|
83 |
+
--local-dir "mistralai/Devstral-Small-2505_gguf/"
|
84 |
+
```
|
85 |
+
|
86 |
+
You can serve the model locally with [LMStudio](https://lmstudio.ai/).
|
87 |
+
* Download [LM Studio](https://lmstudio.ai/) and install it
|
88 |
+
* Install `lms cli ~/.lmstudio/bin/lms bootstrap`
|
89 |
+
* In a bash terminal, run `lms import devstralQ4_K_M.ggu` in the directory where you've downloaded the model checkpoint (e.g. `mistralai/Devstral-Small-2505_gguf`)
|
90 |
+
* Open the LMStudio application, click the terminal icon to get into the developer tab. Click select a model to load and select Devstral Q4 K M. Toggle the status button to start the model, in setting oggle Serve on Local Network to be on.
|
91 |
+
* On the right tab, you will see an API identifier which should be devstralq4_k_m and an api address under API Usage. Keep note of this address, we will use it in the next step.
|
92 |
+
|
93 |
+
You can now interact with the model served from LM Studio with openhands. Start the openhands server with the docker
|
94 |
+
|
95 |
+
```bash
|
96 |
+
docker pull docker.all-hands.dev/all-hands-ai/runtime:0.38-nikolaik
|
97 |
+
docker run -it --rm --pull=always \
|
98 |
+
-e SANDBOX_RUNTIME_CONTAINER_IMAGE=docker.all-hands.dev/all-hands-ai/runtime:0.38-nikolaik \
|
99 |
+
-e LOG_ALL_EVENTS=true \
|
100 |
+
-v /var/run/docker.sock:/var/run/docker.sock \
|
101 |
+
-v ~/.openhands-state:/.openhands-state \
|
102 |
+
-p 3000:3000 \
|
103 |
+
--add-host host.docker.internal:host-gateway \
|
104 |
+
--name openhands-app \
|
105 |
+
docker.all-hands.dev/all-hands-ai/openhands:0.38
|
106 |
+
```
|
107 |
+
|
108 |
+
The server will start at http://0.0.0.0:3000. Open it in your browser and you will see a tab `AI Provider Configuration`. Click “see advanced setting” on the second line. In the new tab, toggle `advanced` to on. Set the custom model to be `mistral/devstralq4_k_m` and Base URL the **api address** we get from the last step in LM Studio. Set API Key to `dummy`. Click save changes.
|
devstral.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af8a54aaedb3e67a7475a35831b18b3a3f446fdd24ff39144087a2d86052ff99
|
3 |
+
size 47153518560
|
devstralQ4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7978367b5b56553e22e516329031547464d12269a753dd716cdb245663264bbf
|
3 |
+
size 13441800160
|
devstralQ4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a9ec4e1b7fa7b8d3b26e56a54efe251349bb67d8a623bae662353a9d84e4b9b
|
3 |
+
size 14333908960
|
devstralQ5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1c6722c376a795a5a1d85838d3c9d2af183d4dba5cb85270f29ab676944fb12
|
3 |
+
size 16763983840
|
devstralQ8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91dbc2c97e13ee7a714206a69ebc33cb0e8170627f15507c7406cb5bd7d4e974
|
3 |
+
size 25054779360
|