NotoriousH2 commited on
Commit
520bcf9
·
1 Parent(s): 672ea8f

NotoriousH2/test2_solar_10.7b_v1.0

Browse files
README.md CHANGED
@@ -1,8 +1,9 @@
1
  ---
2
- license: cc-by-nc-4.0
3
- base_model: 42dot/42dot_LLM-SFT-1.3B
4
  tags:
5
  - generated_from_trainer
 
6
  model-index:
7
  - name: outputs
8
  results: []
@@ -13,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # outputs
15
 
16
- This model is a fine-tuned version of [42dot/42dot_LLM-SFT-1.3B](https://huggingface.co/42dot/42dot_LLM-SFT-1.3B) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -33,7 +34,7 @@ More information needed
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 0.0001
36
- - train_batch_size: 32
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -42,9 +43,14 @@ The following hyperparameters were used during training:
42
  - num_epochs: 3
43
  - mixed_precision_training: Native AMP
44
 
 
 
 
 
45
  ### Framework versions
46
 
 
47
  - Transformers 4.37.0.dev0
48
  - Pytorch 2.1.2+cu121
49
  - Datasets 2.16.0
50
- - Tokenizers 0.15.0
 
1
  ---
2
+ license: apache-2.0
3
+ library_name: peft
4
  tags:
5
  - generated_from_trainer
6
+ base_model: upstage/SOLAR-10.7B-v1.0
7
  model-index:
8
  - name: outputs
9
  results: []
 
14
 
15
  # outputs
16
 
17
+ This model is a fine-tuned version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0001
37
+ - train_batch_size: 16
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
43
  - num_epochs: 3
44
  - mixed_precision_training: Native AMP
45
 
46
+ ### Training results
47
+
48
+
49
+
50
  ### Framework versions
51
 
52
+ - PEFT 0.7.2.dev0
53
  - Transformers 4.37.0.dev0
54
  - Pytorch 2.1.2+cu121
55
  - Datasets 2.16.0
56
+ - Tokenizers 0.15.0
adapter_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "upstage/SOLAR-10.7B-v1.0",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "loftq_config": {},
12
+ "lora_alpha": 8,
13
+ "lora_dropout": 0.05,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
17
+ "peft_type": "LORA",
18
+ "r": 8,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "gate_proj",
23
+ "k_proj",
24
+ "down_proj",
25
+ "o_proj",
26
+ "q_proj",
27
+ "v_proj",
28
+ "up_proj"
29
+ ],
30
+ "task_type": "CAUSAL_LM",
31
+ "use_rslora": false
32
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af498c0deb32a47c7a6da0d82ef437b39e6b1874b3ac50f64ee04e5cd9c43b58
3
+ size 125918320
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:675b34ba787799ca256cd1985d3a70d5b56a1ba36e5440852835755e8cd5ce88
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01feefb0db9ce6eac49ccb0845e04d9503e5977a0b8e85a998fa4c9671e90e5a
3
  size 4664