rohith-yarramala commited on
Commit
ceab5d5
·
verified ·
1 Parent(s): 7faf80b

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -68,7 +68,7 @@
68
  "tie_word_embeddings": false,
69
  "topk_group": 1,
70
  "topk_method": "greedy",
71
- "torch_dtype": "float16",
72
  "transformers_version": "4.49.0",
73
  "use_cache": true,
74
  "v_head_dim": 128,
 
68
  "tie_word_embeddings": false,
69
  "topk_group": 1,
70
  "topk_method": "greedy",
71
+ "torch_dtype": "bfloat16",
72
  "transformers_version": "4.49.0",
73
  "use_cache": true,
74
  "v_head_dim": 128,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79ce9f6a129ed45fb49a1c0568b1214537f112fb9af13edabe0a1d1e802e0d29
3
- size 5000352443
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58b68107db4e400d8debf576905def542eadec73f197e304b52b35ef8231a6e5
3
+ size 5000355234
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0edd1fa351e89d281af4c836dec2388d905a0c79d3691adbacae97461cd70e1
3
- size 4446000255
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:362706de82a532746fd8692a462ecd618be52ef05f37cd8d77dae6f1b23a33ed
3
+ size 4446002730
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 9443660378
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 9443665532
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 100000,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 512
12
+ },
13
+ "direction": "Left",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 100001,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<|end▁of▁sentence|>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 100000,
tokenizer_config.json CHANGED
@@ -154,10 +154,17 @@
154
  "eos_token": "<|end▁of▁sentence|>",
155
  "extra_special_tokens": {},
156
  "legacy": true,
 
157
  "model_max_length": 16384,
 
158
  "pad_token": "<|end▁of▁sentence|>",
 
 
159
  "sp_model_kwargs": {},
 
160
  "tokenizer_class": "LlamaTokenizerFast",
 
 
161
  "unk_token": null,
162
  "use_default_system_prompt": false
163
  }
 
154
  "eos_token": "<|end▁of▁sentence|>",
155
  "extra_special_tokens": {},
156
  "legacy": true,
157
+ "max_length": 512,
158
  "model_max_length": 16384,
159
+ "pad_to_multiple_of": null,
160
  "pad_token": "<|end▁of▁sentence|>",
161
+ "pad_token_type_id": 0,
162
+ "padding_side": "left",
163
  "sp_model_kwargs": {},
164
+ "stride": 0,
165
  "tokenizer_class": "LlamaTokenizerFast",
166
+ "truncation_side": "right",
167
+ "truncation_strategy": "longest_first",
168
  "unk_token": null,
169
  "use_default_system_prompt": false
170
  }