NexVeridian commited on
Commit
2d10140
·
verified ·
1 Parent(s): b05a834

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -7,14 +7,15 @@ tags:
7
  - text adventure
8
  - roleplay
9
  - mlx
10
- library_name: transformers
 
11
  ---
12
 
13
- # mlx-community/Wayfarer-12B-Q4-mlx
14
 
15
- The Model [mlx-community/Wayfarer-12B-Q4-mlx](https://huggingface.co/mlx-community/Wayfarer-12B-Q4-mlx) was
16
  converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
17
- using mlx-lm version **0.21.4**.
18
 
19
  ## Use with mlx
20
 
@@ -25,7 +26,7 @@ pip install mlx-lm
25
  ```python
26
  from mlx_lm import load, generate
27
 
28
- model, tokenizer = load("mlx-community/Wayfarer-12B-Q4-mlx")
29
 
30
  prompt = "hello"
31
 
 
7
  - text adventure
8
  - roleplay
9
  - mlx
10
+ library_name: mlx
11
+ pipeline_tag: text-generation
12
  ---
13
 
14
+ # mlx-community/Wayfarer-12B-4bit
15
 
16
+ This model [mlx-community/Wayfarer-12B-4bit](https://huggingface.co/mlx-community/Wayfarer-12B-4bit) was
17
  converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
18
+ using mlx-lm version **0.25.0**.
19
 
20
  ## Use with mlx
21
 
 
26
  ```python
27
  from mlx_lm import load, generate
28
 
29
+ model, tokenizer = load("mlx-community/Wayfarer-12B-4bit")
30
 
31
  prompt = "hello"
32
 
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.46.1"
7
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1147c1d87b73d73ff9a52c1aeba19ab900857cae1d2bde7a72d0c3af3f3fb66e
3
- size 5356348734
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0bb8544cdc955b7b6a445aaeac70215deb1532b85817bcd1d44ee9cdc997bf0
3
+ size 5356349258
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ce1c1b30503cb6add32647ab2b33650f38a21562d28834521caa6c849bd2098
3
- size 1533741727
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4289d33527fa1969fdf5f7fb56dc17c56c33de81fb3b67d4aad4d7fe9753c6ba
3
+ size 1533741848
model.safetensors.index.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "metadata": {
3
- "total_size": 6889985280
 
4
  },
5
  "weight_map": {
6
  "lm_head.biases": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6889985280,
4
+ "total_parameters": 12247802880
5
  },
6
  "weight_map": {
7
  "lm_head.biases": "model-00002-of-00002.safetensors",
tokenizer_config.json CHANGED
@@ -8024,7 +8024,6 @@
8024
  "<|im_start|>"
8025
  ],
8026
  "bos_token": "<s>",
8027
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
8028
  "clean_up_tokenization_spaces": false,
8029
  "eos_token": "<|im_end|>",
8030
  "extra_special_tokens": {},
 
8024
  "<|im_start|>"
8025
  ],
8026
  "bos_token": "<s>",
 
8027
  "clean_up_tokenization_spaces": false,
8028
  "eos_token": "<|im_end|>",
8029
  "extra_special_tokens": {},