upload checkpoint
Browse files- config.json +1 -1
- generation_config.json +1 -1
- pytorch_model-00001-of-00004.bin +1 -1
- pytorch_model-00004-of-00004.bin +1 -1
- tokenizer.json +2 -2
- tokenizer_config.json +0 -1
config.json
CHANGED
@@ -30,7 +30,7 @@
|
|
30 |
"rope_theta": 500000.0,
|
31 |
"tie_word_embeddings": false,
|
32 |
"torch_dtype": "bfloat16",
|
33 |
-
"transformers_version": "4.
|
34 |
"use_cache": true,
|
35 |
"vocab_size": 128264
|
36 |
}
|
|
|
30 |
"rope_theta": 500000.0,
|
31 |
"tie_word_embeddings": false,
|
32 |
"torch_dtype": "bfloat16",
|
33 |
+
"transformers_version": "4.43.4",
|
34 |
"use_cache": true,
|
35 |
"vocab_size": 128264
|
36 |
}
|
generation_config.json
CHANGED
@@ -5,5 +5,5 @@
|
|
5 |
"eos_token_id": 128001,
|
6 |
"temperature": 0.6,
|
7 |
"top_p": 0.9,
|
8 |
-
"transformers_version": "4.
|
9 |
}
|
|
|
5 |
"eos_token_id": 128001,
|
6 |
"temperature": 0.6,
|
7 |
"top_p": 0.9,
|
8 |
+
"transformers_version": "4.43.4"
|
9 |
}
|
pytorch_model-00001-of-00004.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4976782850
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80ebe122964c3d48b62b3beef2e4463ca1b512140076bb8d292b6da3d8abac6b
|
3 |
size 4976782850
|
pytorch_model-00004-of-00004.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1168206409
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ed027fe0f10dba2b15bb48d8c61449eae6e705c72449d0c9407572af32ada3c
|
3 |
size 1168206409
|
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a62e4ccfd19daf6c53aeca2c0f4474cb7d784bf1ee999f0214715615f46c802c
|
3 |
+
size 9085839
|
tokenizer_config.json
CHANGED
@@ -2061,7 +2061,6 @@
|
|
2061 |
"chat_template": "{{ bos_token }}{% for message in messages %}{% if message['role'] == 'system' %}{{ '<|system|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'user' %}{{ '<|user|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'assistant' %}{% if not loop.last %}{{ '<|assistant|>\n' + message['content'] + eos_token + '\n' }}{% else %}{{ '<|assistant|>\n' + message['content'] + eos_token }}{% endif %}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}{% endfor %}",
|
2062 |
"clean_up_tokenization_spaces": true,
|
2063 |
"eos_token": "<|end_of_text|>",
|
2064 |
-
"extra_special_tokens": {},
|
2065 |
"model_input_names": [
|
2066 |
"input_ids",
|
2067 |
"attention_mask"
|
|
|
2061 |
"chat_template": "{{ bos_token }}{% for message in messages %}{% if message['role'] == 'system' %}{{ '<|system|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'user' %}{{ '<|user|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'assistant' %}{% if not loop.last %}{{ '<|assistant|>\n' + message['content'] + eos_token + '\n' }}{% else %}{{ '<|assistant|>\n' + message['content'] + eos_token }}{% endif %}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}{% endfor %}",
|
2062 |
"clean_up_tokenization_spaces": true,
|
2063 |
"eos_token": "<|end_of_text|>",
|
|
|
2064 |
"model_input_names": [
|
2065 |
"input_ids",
|
2066 |
"attention_mask"
|