juliendenize patrickvonplaten commited on
Commit
3ea99f4
·
verified ·
0 Parent(s):

Super-squash branch 'main' using huggingface_hub

Browse files

Co-authored-by: patrickvonplaten <[email protected]>

.gitattributes ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ devstralQ8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ devstralQ4_0.gguf filter=lfs diff=lfs merge=lfs -text
38
+ devstral.gguf filter=lfs diff=lfs merge=lfs -text
39
+ devstralQ4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ devstralQ5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - fr
5
+ - de
6
+ - es
7
+ - pt
8
+ - it
9
+ - ja
10
+ - ko
11
+ - ru
12
+ - zh
13
+ - ar
14
+ - fa
15
+ - id
16
+ - ms
17
+ - ne
18
+ - pl
19
+ - ro
20
+ - sr
21
+ - sv
22
+ - tr
23
+ - uk
24
+ - vi
25
+ - hi
26
+ - bn
27
+ license: apache-2.0
28
+ library_name: llmstudio
29
+ inference: false
30
+ base_model:
31
+ - mistralai/Devstrall-Small-2505_gguf
32
+ extra_gated_description: >-
33
+ If you want to learn more about how we process your personal data, please read
34
+ our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
35
+ pipeline_tag: text2text-generation
36
+ ---
37
+
38
+
39
+ > [!Note]
40
+ > At Mistral, we don't yet have too much experience with providing GGUF-quantized checkpoints
41
+ > to the community, but want to help improving the ecosystem going forward.
42
+ > If you encounter any problems with the provided checkpoints here, please open a discussion or pull request
43
+
44
+
45
+ # Devstral-Small-2505 (gguf)
46
+
47
+ Devstral is an agentic LLM for software engineering tasks built under a collaboration between [Mistral AI](https://mistral.ai/) and [All Hands AI](https://www.all-hands.dev/) 🙌. Devstral excels at using tools to explore codebases, editing multiple files and power software engineering agents. The model achieves remarkable performance on SWE-bench which positionates it as the #1 open source model on this [benchmark](https://huggingface.co/mistralai/Devstral-Small-2505#benchmark-results).
48
+
49
+ This is the GGUF version of the [Devstral-Small-2505](https://huggingface.co/mistralai/Devstral-Small-2505) model. We released the BF16 weights as well as the following quantized format:
50
+ - Q8_0 (recommended)
51
+ - Q5_K_M (recommended)
52
+ - Q4_K_M (recommended)
53
+ - Q4_0
54
+
55
+ It is finetuned from [Mistral-Small-3.1](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503), therefore it has a long context window of up to 128k tokens. As a coding agent, Devstral is text-only and before fine-tuning from `Mistral-Small-3.1` the vision encoder was removed.
56
+
57
+ For enterprises requiring specialized capabilities (increased context, domain-specific knowledge, etc.), we will release commercial models beyond what Mistral AI contributes to the community.
58
+
59
+ Learn more about Devstral in our [blog post](https://mistral.ai/news/devstral).
60
+
61
+
62
+ ## Key Features:
63
+ - **Agentic coding**: Devstral is designed to excel at agentic coding tasks, making it a great choice for software engineering agents.
64
+ - **lightweight**: with its compact size of just 24 billion parameters, Devstral is light enough to run on a single RTX 4090 or a Mac with 32GB RAM, making it an appropriate model for local deployment and on-device use.
65
+ - **Apache 2.0 License**: Open license allowing usage and modification for both commercial and non-commercial purposes.
66
+ - **Context Window**: A 128k context window.
67
+ - **Tokenizer**: Utilizes a Tekken tokenizer with a 131k vocabulary size.
68
+
69
+ ## Usage
70
+
71
+ We recommend to use Devstral with the [OpenHands](https://github.com/All-Hands-AI/OpenHands/tree/main) scaffold as explained [here](https://huggingface.co/mistralai/Devstral-Small-2505#usage).
72
+ To use it local with a GGUF-quantized checkpoint, see the following section.
73
+
74
+ ### Local inference (GGUF)
75
+
76
+ Download the weights from huggingface:
77
+
78
+ ```
79
+ pip install -U "huggingface_hub[cli]"
80
+ huggingface-cli download \
81
+ "mistralai/Devstral-Small-2505_gguf" \
82
+ --include "devstralQ4_K_M.gguf" \
83
+ --local-dir "mistralai/Devstral-Small-2505_gguf/"
84
+ ```
85
+
86
+ You can serve the model locally with [LMStudio](https://lmstudio.ai/).
87
+ * Download [LM Studio](https://lmstudio.ai/) and install it
88
+ * Install `lms cli ~/.lmstudio/bin/lms bootstrap`
89
+ * In a bash terminal, run `lms import devstralQ4_K_M.ggu` in the directory where you've downloaded the model checkpoint (e.g. `mistralai/Devstral-Small-2505_gguf`)
90
+ * Open the LMStudio application, click the terminal icon to get into the developer tab. Click select a model to load and select Devstral Q4 K M. Toggle the status button to start the model, in setting oggle Serve on Local Network to be on.
91
+ * On the right tab, you will see an API identifier which should be devstralq4_k_m and an api address under API Usage. Keep note of this address, we will use it in the next step.
92
+
93
+ You can now interact with the model served from LM Studio with openhands. Start the openhands server with the docker
94
+
95
+ ```bash
96
+ docker pull docker.all-hands.dev/all-hands-ai/runtime:0.38-nikolaik
97
+ docker run -it --rm --pull=always \
98
+ -e SANDBOX_RUNTIME_CONTAINER_IMAGE=docker.all-hands.dev/all-hands-ai/runtime:0.38-nikolaik \
99
+ -e LOG_ALL_EVENTS=true \
100
+ -v /var/run/docker.sock:/var/run/docker.sock \
101
+ -v ~/.openhands-state:/.openhands-state \
102
+ -p 3000:3000 \
103
+ --add-host host.docker.internal:host-gateway \
104
+ --name openhands-app \
105
+ docker.all-hands.dev/all-hands-ai/openhands:0.38
106
+ ```
107
+
108
+ The server will start at http://0.0.0.0:3000. Open it in your browser and you will see a tab `AI Provider Configuration`. Click “see advanced setting” on the second line. In the new tab, toggle `advanced` to on. Set the custom model to be `mistral/devstralq4_k_m` and Base URL the **api address** we get from the last step in LM Studio. Set API Key to `dummy`. Click save changes.
devstral.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af8a54aaedb3e67a7475a35831b18b3a3f446fdd24ff39144087a2d86052ff99
3
+ size 47153518560
devstralQ4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7978367b5b56553e22e516329031547464d12269a753dd716cdb245663264bbf
3
+ size 13441800160
devstralQ4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a9ec4e1b7fa7b8d3b26e56a54efe251349bb67d8a623bae662353a9d84e4b9b
3
+ size 14333908960
devstralQ5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1c6722c376a795a5a1d85838d3c9d2af183d4dba5cb85270f29ab676944fb12
3
+ size 16763983840
devstralQ8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91dbc2c97e13ee7a714206a69ebc33cb0e8170627f15507c7406cb5bd7d4e974
3
+ size 25054779360