suhani-sarvam commited on
Commit
06c742a
·
verified ·
1 Parent(s): 3c8953f

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -1,6 +1,9 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
4
  "base_model_name_or_path": "sarvamai/sarvam-2b",
5
  "bias": "none",
6
  "eva_config": null,
@@ -23,15 +26,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "down_proj",
27
  "o_proj",
28
- "k_proj",
29
- "up_proj",
30
- "gate_proj",
31
  "v_proj",
32
- "q_proj"
 
 
33
  ],
34
- "task_type": "CAUSAL_LM",
35
  "use_dora": false,
36
  "use_rslora": false
37
  }
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "LlamaForCausalLM",
5
+ "parent_library": "transformers.models.llama.modeling_llama"
6
+ },
7
  "base_model_name_or_path": "sarvamai/sarvam-2b",
8
  "bias": "none",
9
  "eva_config": null,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
29
  "o_proj",
 
 
 
30
  "v_proj",
31
+ "q_proj",
32
+ "k_proj",
33
+ "lm_head"
34
  ],
35
+ "task_type": null,
36
  "use_dora": false,
37
  "use_rslora": false
38
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:848415773217a81110ff0535cf841fe3a6b19fb999a8c79567100d2cc32fa8a2
3
- size 941420376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e12a3e1de962f984fe7e1cbae3a22e55daa33f9b9e45d0adacaae8dcc0edb5c1
3
+ size 399673504
special_tokens_map.json CHANGED
@@ -1,13 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- }
10
- ],
11
  "bos_token": {
12
  "content": "<s>",
13
  "lstrip": false,
 
1
  {
 
 
 
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -32860,18 +32860,14 @@
32860
  "special": true
32861
  }
32862
  },
32863
- "additional_special_tokens": [
32864
- "[PAD]"
32865
- ],
32866
  "bos_token": "<s>",
32867
  "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}\n{% for message in loop_messages %}\n{% if message['role'] not in ['user', 'assistant', 'tool_calls'] %}\n{{ raise_exception('Invalid role: ' + message['role'] + '. Must be user, assistant, or tool_calls.') }}\n{% endif %}\n{% if loop.index0 == 0 and system_message != false %}\n{% set content = '<<SYS>>\n' + system_message + '\n<</SYS>>\n\n' + message['content'] %}\n{% else %}\n{% set content = message['content'] %}\n{% endif %}\n{% if message['role'] == 'user' %}\n{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}\n{% elif message['role'] == 'assistant' %}\n{{ ' ' + content.strip() + ' ' + eos_token }}\n{% elif message['role'] == 'tool_calls' %}\n{{ ' [TOOL_CALLS] ' + content.strip() + ' [/TOOL_CALLS] ' }}\n{% endif %}\n{% endfor %}",
32868
  "clean_up_tokenization_spaces": false,
32869
  "eos_token": "</s>",
32870
  "extra_special_tokens": {},
32871
  "legacy": false,
32872
- "model_max_length": 8192,
32873
  "pad_token": "[PAD]",
32874
- "padding_side": "right",
32875
  "sp_model_kwargs": {},
32876
  "spaces_between_special_tokens": false,
32877
  "tokenizer_class": "LlamaTokenizer",
 
32860
  "special": true
32861
  }
32862
  },
 
 
 
32863
  "bos_token": "<s>",
32864
  "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}\n{% for message in loop_messages %}\n{% if message['role'] not in ['user', 'assistant', 'tool_calls'] %}\n{{ raise_exception('Invalid role: ' + message['role'] + '. Must be user, assistant, or tool_calls.') }}\n{% endif %}\n{% if loop.index0 == 0 and system_message != false %}\n{% set content = '<<SYS>>\n' + system_message + '\n<</SYS>>\n\n' + message['content'] %}\n{% else %}\n{% set content = message['content'] %}\n{% endif %}\n{% if message['role'] == 'user' %}\n{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}\n{% elif message['role'] == 'assistant' %}\n{{ ' ' + content.strip() + ' ' + eos_token }}\n{% elif message['role'] == 'tool_calls' %}\n{{ ' [TOOL_CALLS] ' + content.strip() + ' [/TOOL_CALLS] ' }}\n{% endif %}\n{% endfor %}",
32865
  "clean_up_tokenization_spaces": false,
32866
  "eos_token": "</s>",
32867
  "extra_special_tokens": {},
32868
  "legacy": false,
32869
+ "model_max_length": 1000000000000000019884624838656,
32870
  "pad_token": "[PAD]",
 
32871
  "sp_model_kwargs": {},
32872
  "spaces_between_special_tokens": false,
32873
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba593b17db62769f48cad2733568a19f9d173e99ae9d8956ed8b3f4f7f9c8598
3
- size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:607f46bf18c2b00720eff6ab5f1c2f2f3fa1e25b7abaa8efd8f6682b77a3b582
3
+ size 5240