cowWhySo commited on
Commit
64fe59a
·
verified ·
1 Parent(s): 5a33ca5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +81 -1
README.md CHANGED
@@ -7,4 +7,84 @@ Abliterated using the following the guide here:
7
 
8
  https://huggingface.co/blog/mlabonne/abliteration
9
 
10
- Fine tuned on orpo-dpo-mix-40k dataset
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  https://huggingface.co/blog/mlabonne/abliteration
9
 
10
+
11
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
12
+ <details><summary>See axolotl config</summary>
13
+
14
+ axolotl version: `0.4.0`
15
+ ```yaml
16
+ base_model: cowWhySo/Phi-3-mini-4k-instruct-Friendly
17
+ trust_remote_code: true
18
+ model_type: AutoModelForCausalLM
19
+ tokenizer_type: AutoTokenizer
20
+ chat_template: phi_3
21
+
22
+ load_in_8bit: false
23
+ load_in_4bit: true
24
+ strict: false
25
+ save_safetensors: true
26
+
27
+ rl: dpo
28
+ datasets:
29
+ - path: mlabonne/orpo-dpo-mix-40k
30
+ split: train
31
+ type: chatml.intel
32
+
33
+ dataset_prepared_path:
34
+ val_set_size: 0.0
35
+ output_dir: ./out
36
+
37
+ sequence_len: 4096
38
+ sample_packing: false
39
+ pad_to_sequence_len: false
40
+
41
+ adapter: qlora
42
+ lora_model_dir:
43
+
44
+ lora_r: 64
45
+ lora_alpha: 32
46
+ lora_dropout: 0.1
47
+ lora_target_linear: true
48
+ lora_fan_in_fan_out:
49
+
50
+ wandb_project: axolotl
51
+ wandb_entity:
52
+ wandb_watch:
53
+ wandb_name: phi3-mini-4k-instruct-Friendly
54
+ wandb_log_model:
55
+
56
+ gradient_accumulation_steps: 8
57
+ micro_batch_size: 4
58
+ num_epochs: 1
59
+ optimizer: paged_adamw_8bit
60
+ lr_scheduler: linear
61
+ learning_rate: 5e-6
62
+ train_on_inputs: false
63
+ group_by_length: false
64
+
65
+ bf16: auto
66
+
67
+ gradient_checkpointing: true
68
+ gradient_checkpointing_kwargs:
69
+ use_reentrant: True
70
+ early_stopping_patience:
71
+ resume_from_checkpoint:
72
+ local_rank:
73
+ logging_steps: 1
74
+ xformers_attention:
75
+ flash_attention: true
76
+ warmup_steps: 150
77
+ evals_per_epoch: 0
78
+ eval_table_size:
79
+ eval_table_max_new_tokens: 128
80
+ saves_per_epoch: 1
81
+ debug:
82
+ deepspeed: deepspeed_configs/zero3.json
83
+ weight_decay: 0.01
84
+ max_grad_norm: 1.0
85
+ resize_token_embeddings_to_32x: true
86
+ ```
87
+
88
+ ## Quants
89
+
90
+ GGUF: https://huggingface.co/cowWhySo/Phi-3-mini-4k-instruct-Friendly-gguf