Add files using upload-large-folder tool
Browse files- README.md +6 -5
- chat_template.jinja +4 -0
- generation_config.json +7 -0
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +2 -1
- tokenizer_config.json +0 -1
README.md
CHANGED
@@ -7,14 +7,15 @@ tags:
|
|
7 |
- text adventure
|
8 |
- roleplay
|
9 |
- mlx
|
10 |
-
library_name:
|
|
|
11 |
---
|
12 |
|
13 |
-
# mlx-community/Wayfarer-12B-
|
14 |
|
15 |
-
|
16 |
converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
|
17 |
-
using mlx-lm version **0.
|
18 |
|
19 |
## Use with mlx
|
20 |
|
@@ -25,7 +26,7 @@ pip install mlx-lm
|
|
25 |
```python
|
26 |
from mlx_lm import load, generate
|
27 |
|
28 |
-
model, tokenizer = load("mlx-community/Wayfarer-12B-
|
29 |
|
30 |
prompt = "hello"
|
31 |
|
|
|
7 |
- text adventure
|
8 |
- roleplay
|
9 |
- mlx
|
10 |
+
library_name: mlx
|
11 |
+
pipeline_tag: text-generation
|
12 |
---
|
13 |
|
14 |
+
# mlx-community/Wayfarer-12B-4bit
|
15 |
|
16 |
+
This model [mlx-community/Wayfarer-12B-4bit](https://huggingface.co/mlx-community/Wayfarer-12B-4bit) was
|
17 |
converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
|
18 |
+
using mlx-lm version **0.25.0**.
|
19 |
|
20 |
## Use with mlx
|
21 |
|
|
|
26 |
```python
|
27 |
from mlx_lm import load, generate
|
28 |
|
29 |
+
model, tokenizer = load("mlx-community/Wayfarer-12B-4bit")
|
30 |
|
31 |
prompt = "hello"
|
32 |
|
chat_template.jinja
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
|
2 |
+
' + message['content'] + '<|im_end|>' + '
|
3 |
+
'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
|
4 |
+
' }}{% endif %}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"do_sample": true,
|
5 |
+
"eos_token_id": 2,
|
6 |
+
"transformers_version": "4.46.1"
|
7 |
+
}
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0bb8544cdc955b7b6a445aaeac70215deb1532b85817bcd1d44ee9cdc997bf0
|
3 |
+
size 5356349258
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4289d33527fa1969fdf5f7fb56dc17c56c33de81fb3b67d4aad4d7fe9753c6ba
|
3 |
+
size 1533741848
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size": 6889985280
|
|
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.biases": "model-00002-of-00002.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 6889985280,
|
4 |
+
"total_parameters": 12247802880
|
5 |
},
|
6 |
"weight_map": {
|
7 |
"lm_head.biases": "model-00002-of-00002.safetensors",
|
tokenizer_config.json
CHANGED
@@ -8024,7 +8024,6 @@
|
|
8024 |
"<|im_start|>"
|
8025 |
],
|
8026 |
"bos_token": "<s>",
|
8027 |
-
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
8028 |
"clean_up_tokenization_spaces": false,
|
8029 |
"eos_token": "<|im_end|>",
|
8030 |
"extra_special_tokens": {},
|
|
|
8024 |
"<|im_start|>"
|
8025 |
],
|
8026 |
"bos_token": "<s>",
|
|
|
8027 |
"clean_up_tokenization_spaces": false,
|
8028 |
"eos_token": "<|im_end|>",
|
8029 |
"extra_special_tokens": {},
|