NexVeridian commited on
Commit
6becf81
·
verified ·
1 Parent(s): 4936da1

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -7,14 +7,15 @@ tags:
7
  - text adventure
8
  - roleplay
9
  - mlx
10
- library_name: transformers
 
11
  ---
12
 
13
- # mlx-community/Wayfarer-12B-Q6-mlx
14
 
15
- The Model [mlx-community/Wayfarer-12B-Q6-mlx](https://huggingface.co/mlx-community/Wayfarer-12B-Q6-mlx) was
16
  converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
17
- using mlx-lm version **0.21.4**.
18
 
19
  ## Use with mlx
20
 
@@ -25,7 +26,7 @@ pip install mlx-lm
25
  ```python
26
  from mlx_lm import load, generate
27
 
28
- model, tokenizer = load("mlx-community/Wayfarer-12B-Q6-mlx")
29
 
30
  prompt = "hello"
31
 
 
7
  - text adventure
8
  - roleplay
9
  - mlx
10
+ library_name: mlx
11
+ pipeline_tag: text-generation
12
  ---
13
 
14
+ # mlx-community/Wayfarer-12B-6bit
15
 
16
+ This model [mlx-community/Wayfarer-12B-6bit](https://huggingface.co/mlx-community/Wayfarer-12B-6bit) was
17
  converted to MLX format from [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
18
+ using mlx-lm version **0.25.0**.
19
 
20
  ## Use with mlx
21
 
 
26
  ```python
27
  from mlx_lm import load, generate
28
 
29
+ model, tokenizer = load("mlx-community/Wayfarer-12B-6bit")
30
 
31
  prompt = "hello"
32
 
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.46.1"
7
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e878d60b35e349f683b02b39247253b452f33505f2f2104e9bf60e9157abce9b
3
- size 5359374311
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82c36e91ea8a86d486e641914c63cf630632ce55db46bb739c74c437ffc448d3
3
+ size 5359374661
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9b083cc85d6dd95d3636879f235aa69d829371eba85e7ae8208825a67de0333
3
- size 4592563358
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddb5f3f544de2608608021c0f2aed4c245032a647d099f0bb8704ae414f1d66e
3
+ size 4592563653
model.safetensors.index.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "metadata": {
3
- "total_size": 9951832320
 
4
  },
5
  "weight_map": {
6
  "lm_head.biases": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 9951832320,
4
+ "total_parameters": 12247802880
5
  },
6
  "weight_map": {
7
  "lm_head.biases": "model-00002-of-00002.safetensors",
tokenizer_config.json CHANGED
@@ -8024,7 +8024,6 @@
8024
  "<|im_start|>"
8025
  ],
8026
  "bos_token": "<s>",
8027
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
8028
  "clean_up_tokenization_spaces": false,
8029
  "eos_token": "<|im_end|>",
8030
  "extra_special_tokens": {},
 
8024
  "<|im_start|>"
8025
  ],
8026
  "bos_token": "<s>",
 
8027
  "clean_up_tokenization_spaces": false,
8028
  "eos_token": "<|im_end|>",
8029
  "extra_special_tokens": {},