sezing commited on
Commit
4e108e0
·
verified ·
1 Parent(s): 8ca7861

mistralai/mistral-instruct-generation-szn-by-me

Browse files
README.md CHANGED
@@ -5,9 +5,9 @@ tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
 
8
  datasets:
9
  - generator
10
- base_model: mistralai/Mistral-7B-Instruct-v0.1
11
  model-index:
12
  - name: mistral_instruct_generation
13
  results: []
@@ -18,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # mistral_instruct_generation
20
 
21
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.4967
24
 
25
  ## Model description
26
 
@@ -51,19 +51,19 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:-----:|:----:|:---------------:|
56
- | 1.6319 | 0.03 | 20 | 1.5642 |
57
- | 1.5826 | 0.05 | 40 | 1.5262 |
58
- | 1.5995 | 0.08 | 60 | 1.5122 |
59
- | 1.5292 | 0.1 | 80 | 1.5030 |
60
- | 1.5593 | 0.13 | 100 | 1.4967 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.10.0
66
- - Transformers 4.39.0
67
- - Pytorch 2.2.1+cu121
68
- - Datasets 2.18.0
69
- - Tokenizers 0.15.2
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
9
  datasets:
10
  - generator
 
11
  model-index:
12
  - name: mistral_instruct_generation
13
  results: []
 
18
 
19
  # mistral_instruct_generation
20
 
21
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.4389
24
 
25
  ## Model description
26
 
 
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss |
55
+ |:-------------:|:------:|:----:|:---------------:|
56
+ | 1.6485 | 0.0260 | 20 | 1.5379 |
57
+ | 1.5023 | 0.0521 | 40 | 1.4691 |
58
+ | 1.4632 | 0.0781 | 60 | 1.4547 |
59
+ | 1.4887 | 0.1042 | 80 | 1.4471 |
60
+ | 1.4516 | 0.1302 | 100 | 1.4389 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.10.0
66
+ - Transformers 4.40.2
67
+ - Pytorch 2.3.0+cu121
68
+ - Datasets 2.19.1
69
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006e25bc0327fc595e6dfbc7d7ead653a2569e75be29e97d6151232dd5a7b7fd
3
  size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5421be3a38d5755f19eca3c93b4ce525e65bbeec1ad3039a5fe8992e7da7035a
3
  size 27280152
runs/May07_08-49-54_bf0b8504c300/events.out.tfevents.1715071893.bf0b8504c300.316.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ae77b92376d221020c12c87a89e6fd2e0b01c384e279e38a12d65eb58e33bb9
3
+ size 8952
tokenizer.json CHANGED
@@ -134,6 +134,7 @@
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
 
137
  "vocab": {
138
  "<unk>": 0,
139
  "<s>": 1,
 
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
137
+ "ignore_merges": false,
138
  "vocab": {
139
  "<unk>": 0,
140
  "<s>": 1,
tokenizer_config.json CHANGED
@@ -29,6 +29,7 @@
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
 
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": true,
 
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": true,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a13efc75d4bb173350b7c55d9b31eb6c3514abd7257e5fdcfb5eca18f55b3429
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0804dc137cce2cbe9de0d72272ccd11e954bc695cb8f25d194bc811e26c3e9f8
3
+ size 5048