Bakugo123 commited on
Commit
12c8998
·
verified ·
1 Parent(s): ba2deab

Training in progress, epoch 2

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "meta-llama/Meta-Llama-3-8B",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e463f4d9e16ecaf4a60a3b396702f06ecc551699ed3c9a3594ad25d7980ad52
3
  size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:461efed3c23271fd56410c7ca7b80996fbd31b52647b85f8638c42a6f040b596
3
  size 109069176
runs/Jul09_03-41-07_nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz/events.out.tfevents.1752032493.nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz.768.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e82c8154b886bbbed3ad39febe48826eb73e534371a47c7628d45bd7e29c1a
3
+ size 25970
runs/Jul09_04-08-59_nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz/events.out.tfevents.1752034161.nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz.768.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e177b32048a1bc74a8ac99505e215b86bf711b4d18d3322504203bedee069175
3
+ size 6946
runs/Jul09_04-13-16_nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz/events.out.tfevents.1752034405.nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz.768.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a333993d8451cdfcbb4a29cb3ab856df5ac282b6e310f9417c4e9805341a1545
3
+ size 7984
runs/Jul09_04-20-40_nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz/events.out.tfevents.1752034847.nb-79d9ab29-73f9-44c0-a896-df6270c588fc-6bccdc4c95-d2bzz.768.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01cdae8ad1b34bedee6019888e167ac158c7249840d7b53acffde25ea8eda19b
3
+ size 104195
special_tokens_map.json CHANGED
@@ -7,11 +7,11 @@
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "<|end_of_text|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|end_of_text|>"
17
  }
 
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|eot_id|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|eot_id|>"
17
  }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 2042,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -2050,13 +2050,14 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
 
2053
  "clean_up_tokenization_spaces": true,
2054
- "eos_token": "<|end_of_text|>",
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
  "model_max_length": 1000000000000000019884624838656,
2060
- "pad_token": "<|end_of_text|>",
2061
  "tokenizer_class": "PreTrainedTokenizerFast"
2062
  }
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|eot_id|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 1000000000000000019884624838656,
2061
+ "pad_token": "<|eot_id|>",
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20941180b276331bf7aa9066d73f944563e4c89a8f3b87363748cd4d2ca3d54d
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6286358a7fc536d696b60de46cef40fc148ae6d8dce5a985a9af09700228159
3
  size 5496