End of training
Browse files- README.md +153 -0
 - adapter_config.json +30 -0
 - adapter_model.bin +3 -0
 - adapter_model.safetensors +3 -0
 - added_tokens.json +4 -0
 - config.json +37 -0
 - last-checkpoint/README.md +202 -0
 - last-checkpoint/adapter_config.json +30 -0
 - last-checkpoint/adapter_model.safetensors +3 -0
 - last-checkpoint/added_tokens.json +4 -0
 - last-checkpoint/merges.txt +0 -0
 - last-checkpoint/optimizer.pt +3 -0
 - last-checkpoint/rng_state.pth +3 -0
 - last-checkpoint/scheduler.pt +3 -0
 - last-checkpoint/special_tokens_map.json +30 -0
 - last-checkpoint/tokenizer.json +0 -0
 - last-checkpoint/tokenizer_config.json +205 -0
 - last-checkpoint/trainer_state.json +87 -0
 - last-checkpoint/training_args.bin +3 -0
 - last-checkpoint/vocab.json +0 -0
 - merges.txt +0 -0
 - special_tokens_map.json +30 -0
 - tokenizer.json +0 -0
 - tokenizer_config.json +205 -0
 - training_args.bin +3 -0
 - vocab.json +0 -0
 
    	
        README.md
    ADDED
    
    | 
         @@ -0,0 +1,153 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            ---
         
     | 
| 2 | 
         
            +
            library_name: peft
         
     | 
| 3 | 
         
            +
            base_model: katuni4ka/tiny-random-dbrx
         
     | 
| 4 | 
         
            +
            tags:
         
     | 
| 5 | 
         
            +
            - axolotl
         
     | 
| 6 | 
         
            +
            - generated_from_trainer
         
     | 
| 7 | 
         
            +
            model-index:
         
     | 
| 8 | 
         
            +
            - name: d3007acb-9f32-4d87-90e1-b3ffd35a7f7e
         
     | 
| 9 | 
         
            +
              results: []
         
     | 
| 10 | 
         
            +
            ---
         
     | 
| 11 | 
         
            +
             
     | 
| 12 | 
         
            +
            <!-- This model card has been generated automatically according to the information the Trainer had access to. You
         
     | 
| 13 | 
         
            +
            should probably proofread and complete it, then remove this comment. -->
         
     | 
| 14 | 
         
            +
             
     | 
| 15 | 
         
            +
            [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
         
     | 
| 16 | 
         
            +
            <details><summary>See axolotl config</summary>
         
     | 
| 17 | 
         
            +
             
     | 
| 18 | 
         
            +
            axolotl version: `0.4.1`
         
     | 
| 19 | 
         
            +
            ```yaml
         
     | 
| 20 | 
         
            +
            adapter: lora
         
     | 
| 21 | 
         
            +
            base_model: katuni4ka/tiny-random-dbrx
         
     | 
| 22 | 
         
            +
            bf16: auto
         
     | 
| 23 | 
         
            +
            chat_template: llama3
         
     | 
| 24 | 
         
            +
            dataset_prepared_path: null
         
     | 
| 25 | 
         
            +
            datasets:
         
     | 
| 26 | 
         
            +
            - data_files:
         
     | 
| 27 | 
         
            +
              - 1d9e8696621b04b2_train_data.json
         
     | 
| 28 | 
         
            +
              ds_type: json
         
     | 
| 29 | 
         
            +
              format: custom
         
     | 
| 30 | 
         
            +
              path: /workspace/input_data/1d9e8696621b04b2_train_data.json
         
     | 
| 31 | 
         
            +
              type:
         
     | 
| 32 | 
         
            +
                field_input: chosen
         
     | 
| 33 | 
         
            +
                field_instruction: prompt_llama3
         
     | 
| 34 | 
         
            +
                field_output: chosen_llama3
         
     | 
| 35 | 
         
            +
                format: '{instruction} {input}'
         
     | 
| 36 | 
         
            +
                no_input_format: '{instruction}'
         
     | 
| 37 | 
         
            +
                system_format: '{system}'
         
     | 
| 38 | 
         
            +
                system_prompt: ''
         
     | 
| 39 | 
         
            +
            debug: null
         
     | 
| 40 | 
         
            +
            deepspeed: null
         
     | 
| 41 | 
         
            +
            device: cuda
         
     | 
| 42 | 
         
            +
            early_stopping_patience: 1
         
     | 
| 43 | 
         
            +
            eval_max_new_tokens: 128
         
     | 
| 44 | 
         
            +
            eval_steps: 5
         
     | 
| 45 | 
         
            +
            eval_table_size: null
         
     | 
| 46 | 
         
            +
            evals_per_epoch: null
         
     | 
| 47 | 
         
            +
            flash_attention: false
         
     | 
| 48 | 
         
            +
            fp16: null
         
     | 
| 49 | 
         
            +
            gradient_accumulation_steps: 4
         
     | 
| 50 | 
         
            +
            gradient_checkpointing: true
         
     | 
| 51 | 
         
            +
            group_by_length: false
         
     | 
| 52 | 
         
            +
            hub_model_id: null
         
     | 
| 53 | 
         
            +
            hub_repo: null
         
     | 
| 54 | 
         
            +
            hub_strategy: checkpoint
         
     | 
| 55 | 
         
            +
            hub_token: null
         
     | 
| 56 | 
         
            +
            learning_rate: 0.0002
         
     | 
| 57 | 
         
            +
            load_in_4bit: false
         
     | 
| 58 | 
         
            +
            load_in_8bit: false
         
     | 
| 59 | 
         
            +
            local_rank: null
         
     | 
| 60 | 
         
            +
            logging_steps: 3
         
     | 
| 61 | 
         
            +
            lora_alpha: 32
         
     | 
| 62 | 
         
            +
            lora_dropout: 0.05
         
     | 
| 63 | 
         
            +
            lora_fan_in_fan_out: null
         
     | 
| 64 | 
         
            +
            lora_model_dir: null
         
     | 
| 65 | 
         
            +
            lora_r: 16
         
     | 
| 66 | 
         
            +
            lora_target_linear: true
         
     | 
| 67 | 
         
            +
            lr_scheduler: cosine
         
     | 
| 68 | 
         
            +
            max_memory:
         
     | 
| 69 | 
         
            +
              0: 78GiB
         
     | 
| 70 | 
         
            +
            max_steps: 30
         
     | 
| 71 | 
         
            +
            micro_batch_size: 2
         
     | 
| 72 | 
         
            +
            mlflow_experiment_name: /tmp/1d9e8696621b04b2_train_data.json
         
     | 
| 73 | 
         
            +
            model_type: AutoModelForCausalLM
         
     | 
| 74 | 
         
            +
            num_epochs: 1
         
     | 
| 75 | 
         
            +
            optimizer: adamw_torch
         
     | 
| 76 | 
         
            +
            output_dir: miner_id_24
         
     | 
| 77 | 
         
            +
            pad_to_sequence_len: true
         
     | 
| 78 | 
         
            +
            resume_from_checkpoint: null
         
     | 
| 79 | 
         
            +
            s2_attention: null
         
     | 
| 80 | 
         
            +
            sample_packing: false
         
     | 
| 81 | 
         
            +
            save_steps: 10
         
     | 
| 82 | 
         
            +
            sequence_len: 1024
         
     | 
| 83 | 
         
            +
            strict: false
         
     | 
| 84 | 
         
            +
            tf32: false
         
     | 
| 85 | 
         
            +
            tokenizer_type: AutoTokenizer
         
     | 
| 86 | 
         
            +
            train_on_inputs: false
         
     | 
| 87 | 
         
            +
            trust_remote_code: true
         
     | 
| 88 | 
         
            +
            val_set_size: 0.05
         
     | 
| 89 | 
         
            +
            wandb_entity: null
         
     | 
| 90 | 
         
            +
            wandb_mode: online
         
     | 
| 91 | 
         
            +
            wandb_name: 2db8f562-efc1-461b-b2da-b872af3c9023
         
     | 
| 92 | 
         
            +
            wandb_project: Gradients-On-Demand
         
     | 
| 93 | 
         
            +
            wandb_run: your_name
         
     | 
| 94 | 
         
            +
            wandb_runid: 2db8f562-efc1-461b-b2da-b872af3c9023
         
     | 
| 95 | 
         
            +
            warmup_steps: 5
         
     | 
| 96 | 
         
            +
            weight_decay: 0.01
         
     | 
| 97 | 
         
            +
            xformers_attention: true
         
     | 
| 98 | 
         
            +
             
     | 
| 99 | 
         
            +
            ```
         
     | 
| 100 | 
         
            +
             
     | 
| 101 | 
         
            +
            </details><br>
         
     | 
| 102 | 
         
            +
             
     | 
| 103 | 
         
            +
            # d3007acb-9f32-4d87-90e1-b3ffd35a7f7e
         
     | 
| 104 | 
         
            +
             
     | 
| 105 | 
         
            +
            This model is a fine-tuned version of [katuni4ka/tiny-random-dbrx](https://huggingface.co/katuni4ka/tiny-random-dbrx) on the None dataset.
         
     | 
| 106 | 
         
            +
            It achieves the following results on the evaluation set:
         
     | 
| 107 | 
         
            +
            - Loss: 11.5
         
     | 
| 108 | 
         
            +
             
     | 
| 109 | 
         
            +
            ## Model description
         
     | 
| 110 | 
         
            +
             
     | 
| 111 | 
         
            +
            More information needed
         
     | 
| 112 | 
         
            +
             
     | 
| 113 | 
         
            +
            ## Intended uses & limitations
         
     | 
| 114 | 
         
            +
             
     | 
| 115 | 
         
            +
            More information needed
         
     | 
| 116 | 
         
            +
             
     | 
| 117 | 
         
            +
            ## Training and evaluation data
         
     | 
| 118 | 
         
            +
             
     | 
| 119 | 
         
            +
            More information needed
         
     | 
| 120 | 
         
            +
             
     | 
| 121 | 
         
            +
            ## Training procedure
         
     | 
| 122 | 
         
            +
             
     | 
| 123 | 
         
            +
            ### Training hyperparameters
         
     | 
| 124 | 
         
            +
             
     | 
| 125 | 
         
            +
            The following hyperparameters were used during training:
         
     | 
| 126 | 
         
            +
            - learning_rate: 0.0002
         
     | 
| 127 | 
         
            +
            - train_batch_size: 2
         
     | 
| 128 | 
         
            +
            - eval_batch_size: 2
         
     | 
| 129 | 
         
            +
            - seed: 42
         
     | 
| 130 | 
         
            +
            - gradient_accumulation_steps: 4
         
     | 
| 131 | 
         
            +
            - total_train_batch_size: 8
         
     | 
| 132 | 
         
            +
            - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
         
     | 
| 133 | 
         
            +
            - lr_scheduler_type: cosine
         
     | 
| 134 | 
         
            +
            - lr_scheduler_warmup_steps: 5
         
     | 
| 135 | 
         
            +
            - training_steps: 30
         
     | 
| 136 | 
         
            +
             
     | 
| 137 | 
         
            +
            ### Training results
         
     | 
| 138 | 
         
            +
             
     | 
| 139 | 
         
            +
            | Training Loss | Epoch  | Step | Validation Loss |
         
     | 
| 140 | 
         
            +
            |:-------------:|:------:|:----:|:---------------:|
         
     | 
| 141 | 
         
            +
            | No log        | 0.0001 | 1    | 11.5            |
         
     | 
| 142 | 
         
            +
            | 46.0          | 0.0007 | 5    | 11.5            |
         
     | 
| 143 | 
         
            +
            | 46.0          | 0.0014 | 10   | 11.5            |
         
     | 
| 144 | 
         
            +
            | 46.0          | 0.0021 | 15   | 11.5            |
         
     | 
| 145 | 
         
            +
             
     | 
| 146 | 
         
            +
             
     | 
| 147 | 
         
            +
            ### Framework versions
         
     | 
| 148 | 
         
            +
             
     | 
| 149 | 
         
            +
            - PEFT 0.13.2
         
     | 
| 150 | 
         
            +
            - Transformers 4.46.0
         
     | 
| 151 | 
         
            +
            - Pytorch 2.5.0+cu124
         
     | 
| 152 | 
         
            +
            - Datasets 3.0.1
         
     | 
| 153 | 
         
            +
            - Tokenizers 0.20.1
         
     | 
    	
        adapter_config.json
    ADDED
    
    | 
         @@ -0,0 +1,30 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "alpha_pattern": {},
         
     | 
| 3 | 
         
            +
              "auto_mapping": null,
         
     | 
| 4 | 
         
            +
              "base_model_name_or_path": "katuni4ka/tiny-random-dbrx",
         
     | 
| 5 | 
         
            +
              "bias": "none",
         
     | 
| 6 | 
         
            +
              "fan_in_fan_out": null,
         
     | 
| 7 | 
         
            +
              "inference_mode": true,
         
     | 
| 8 | 
         
            +
              "init_lora_weights": true,
         
     | 
| 9 | 
         
            +
              "layer_replication": null,
         
     | 
| 10 | 
         
            +
              "layers_pattern": null,
         
     | 
| 11 | 
         
            +
              "layers_to_transform": null,
         
     | 
| 12 | 
         
            +
              "loftq_config": {},
         
     | 
| 13 | 
         
            +
              "lora_alpha": 32,
         
     | 
| 14 | 
         
            +
              "lora_dropout": 0.05,
         
     | 
| 15 | 
         
            +
              "megatron_config": null,
         
     | 
| 16 | 
         
            +
              "megatron_core": "megatron.core",
         
     | 
| 17 | 
         
            +
              "modules_to_save": null,
         
     | 
| 18 | 
         
            +
              "peft_type": "LORA",
         
     | 
| 19 | 
         
            +
              "r": 16,
         
     | 
| 20 | 
         
            +
              "rank_pattern": {},
         
     | 
| 21 | 
         
            +
              "revision": null,
         
     | 
| 22 | 
         
            +
              "target_modules": [
         
     | 
| 23 | 
         
            +
                "Wqkv",
         
     | 
| 24 | 
         
            +
                "layer",
         
     | 
| 25 | 
         
            +
                "out_proj"
         
     | 
| 26 | 
         
            +
              ],
         
     | 
| 27 | 
         
            +
              "task_type": "CAUSAL_LM",
         
     | 
| 28 | 
         
            +
              "use_dora": false,
         
     | 
| 29 | 
         
            +
              "use_rslora": false
         
     | 
| 30 | 
         
            +
            }
         
     | 
    	
        adapter_model.bin
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c4ec3d65a7c96ba6658db0d9d00d45ad29f15a51bd9eb80f1c1f34f90c5db1a1
         
     | 
| 3 | 
         
            +
            size 13266
         
     | 
    	
        adapter_model.safetensors
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:7fb3c8977e1a5e421a3ade55cd1bbd78cf06a024779d83da21b8b40651ae09b8
         
     | 
| 3 | 
         
            +
            size 9864
         
     | 
    	
        added_tokens.json
    ADDED
    
    | 
         @@ -0,0 +1,4 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "<|im_end|>": 100279,
         
     | 
| 3 | 
         
            +
              "<|im_start|>": 100278
         
     | 
| 4 | 
         
            +
            }
         
     | 
    	
        config.json
    ADDED
    
    | 
         @@ -0,0 +1,37 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_attn_implementation_autoset": true,
         
     | 
| 3 | 
         
            +
              "_name_or_path": "katuni4ka/tiny-random-dbrx",
         
     | 
| 4 | 
         
            +
              "architectures": [
         
     | 
| 5 | 
         
            +
                "DbrxForCausalLM"
         
     | 
| 6 | 
         
            +
              ],
         
     | 
| 7 | 
         
            +
              "attn_config": {
         
     | 
| 8 | 
         
            +
                "clip_qkv": 8,
         
     | 
| 9 | 
         
            +
                "kv_n_heads": 2,
         
     | 
| 10 | 
         
            +
                "model_type": "",
         
     | 
| 11 | 
         
            +
                "rope_theta": 500000
         
     | 
| 12 | 
         
            +
              },
         
     | 
| 13 | 
         
            +
              "d_model": 8,
         
     | 
| 14 | 
         
            +
              "emb_pdrop": 0.0,
         
     | 
| 15 | 
         
            +
              "ffn_config": {
         
     | 
| 16 | 
         
            +
                "ffn_hidden_size": 8,
         
     | 
| 17 | 
         
            +
                "model_type": "",
         
     | 
| 18 | 
         
            +
                "moe_jitter_eps": 0,
         
     | 
| 19 | 
         
            +
                "moe_loss_weight": 0.05,
         
     | 
| 20 | 
         
            +
                "moe_num_experts": 16,
         
     | 
| 21 | 
         
            +
                "moe_top_k": 4
         
     | 
| 22 | 
         
            +
              },
         
     | 
| 23 | 
         
            +
              "initializer_range": 0.02,
         
     | 
| 24 | 
         
            +
              "max_seq_len": 32768,
         
     | 
| 25 | 
         
            +
              "model_type": "dbrx",
         
     | 
| 26 | 
         
            +
              "n_heads": 4,
         
     | 
| 27 | 
         
            +
              "n_layers": 2,
         
     | 
| 28 | 
         
            +
              "num_key_value_heads": 2,
         
     | 
| 29 | 
         
            +
              "output_router_logits": false,
         
     | 
| 30 | 
         
            +
              "resid_pdrop": 0.0,
         
     | 
| 31 | 
         
            +
              "router_aux_loss_coef": 0.05,
         
     | 
| 32 | 
         
            +
              "tie_word_embeddings": false,
         
     | 
| 33 | 
         
            +
              "torch_dtype": "float32",
         
     | 
| 34 | 
         
            +
              "transformers_version": "4.46.0",
         
     | 
| 35 | 
         
            +
              "use_cache": false,
         
     | 
| 36 | 
         
            +
              "vocab_size": 100352
         
     | 
| 37 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/README.md
    ADDED
    
    | 
         @@ -0,0 +1,202 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            ---
         
     | 
| 2 | 
         
            +
            base_model: katuni4ka/tiny-random-dbrx
         
     | 
| 3 | 
         
            +
            library_name: peft
         
     | 
| 4 | 
         
            +
            ---
         
     | 
| 5 | 
         
            +
             
     | 
| 6 | 
         
            +
            # Model Card for Model ID
         
     | 
| 7 | 
         
            +
             
     | 
| 8 | 
         
            +
            <!-- Provide a quick summary of what the model is/does. -->
         
     | 
| 9 | 
         
            +
             
     | 
| 10 | 
         
            +
             
     | 
| 11 | 
         
            +
             
     | 
| 12 | 
         
            +
            ## Model Details
         
     | 
| 13 | 
         
            +
             
     | 
| 14 | 
         
            +
            ### Model Description
         
     | 
| 15 | 
         
            +
             
     | 
| 16 | 
         
            +
            <!-- Provide a longer summary of what this model is. -->
         
     | 
| 17 | 
         
            +
             
     | 
| 18 | 
         
            +
             
     | 
| 19 | 
         
            +
             
     | 
| 20 | 
         
            +
            - **Developed by:** [More Information Needed]
         
     | 
| 21 | 
         
            +
            - **Funded by [optional]:** [More Information Needed]
         
     | 
| 22 | 
         
            +
            - **Shared by [optional]:** [More Information Needed]
         
     | 
| 23 | 
         
            +
            - **Model type:** [More Information Needed]
         
     | 
| 24 | 
         
            +
            - **Language(s) (NLP):** [More Information Needed]
         
     | 
| 25 | 
         
            +
            - **License:** [More Information Needed]
         
     | 
| 26 | 
         
            +
            - **Finetuned from model [optional]:** [More Information Needed]
         
     | 
| 27 | 
         
            +
             
     | 
| 28 | 
         
            +
            ### Model Sources [optional]
         
     | 
| 29 | 
         
            +
             
     | 
| 30 | 
         
            +
            <!-- Provide the basic links for the model. -->
         
     | 
| 31 | 
         
            +
             
     | 
| 32 | 
         
            +
            - **Repository:** [More Information Needed]
         
     | 
| 33 | 
         
            +
            - **Paper [optional]:** [More Information Needed]
         
     | 
| 34 | 
         
            +
            - **Demo [optional]:** [More Information Needed]
         
     | 
| 35 | 
         
            +
             
     | 
| 36 | 
         
            +
            ## Uses
         
     | 
| 37 | 
         
            +
             
     | 
| 38 | 
         
            +
            <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
         
     | 
| 39 | 
         
            +
             
     | 
| 40 | 
         
            +
            ### Direct Use
         
     | 
| 41 | 
         
            +
             
     | 
| 42 | 
         
            +
            <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
         
     | 
| 43 | 
         
            +
             
     | 
| 44 | 
         
            +
            [More Information Needed]
         
     | 
| 45 | 
         
            +
             
     | 
| 46 | 
         
            +
            ### Downstream Use [optional]
         
     | 
| 47 | 
         
            +
             
     | 
| 48 | 
         
            +
            <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
         
     | 
| 49 | 
         
            +
             
     | 
| 50 | 
         
            +
            [More Information Needed]
         
     | 
| 51 | 
         
            +
             
     | 
| 52 | 
         
            +
            ### Out-of-Scope Use
         
     | 
| 53 | 
         
            +
             
     | 
| 54 | 
         
            +
            <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
         
     | 
| 55 | 
         
            +
             
     | 
| 56 | 
         
            +
            [More Information Needed]
         
     | 
| 57 | 
         
            +
             
     | 
| 58 | 
         
            +
            ## Bias, Risks, and Limitations
         
     | 
| 59 | 
         
            +
             
     | 
| 60 | 
         
            +
            <!-- This section is meant to convey both technical and sociotechnical limitations. -->
         
     | 
| 61 | 
         
            +
             
     | 
| 62 | 
         
            +
            [More Information Needed]
         
     | 
| 63 | 
         
            +
             
     | 
| 64 | 
         
            +
            ### Recommendations
         
     | 
| 65 | 
         
            +
             
     | 
| 66 | 
         
            +
            <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
         
     | 
| 67 | 
         
            +
             
     | 
| 68 | 
         
            +
            Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
         
     | 
| 69 | 
         
            +
             
     | 
| 70 | 
         
            +
            ## How to Get Started with the Model
         
     | 
| 71 | 
         
            +
             
     | 
| 72 | 
         
            +
            Use the code below to get started with the model.
         
     | 
| 73 | 
         
            +
             
     | 
| 74 | 
         
            +
            [More Information Needed]
         
     | 
| 75 | 
         
            +
             
     | 
| 76 | 
         
            +
            ## Training Details
         
     | 
| 77 | 
         
            +
             
     | 
| 78 | 
         
            +
            ### Training Data
         
     | 
| 79 | 
         
            +
             
     | 
| 80 | 
         
            +
            <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
         
     | 
| 81 | 
         
            +
             
     | 
| 82 | 
         
            +
            [More Information Needed]
         
     | 
| 83 | 
         
            +
             
     | 
| 84 | 
         
            +
            ### Training Procedure
         
     | 
| 85 | 
         
            +
             
     | 
| 86 | 
         
            +
            <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
         
     | 
| 87 | 
         
            +
             
     | 
| 88 | 
         
            +
            #### Preprocessing [optional]
         
     | 
| 89 | 
         
            +
             
     | 
| 90 | 
         
            +
            [More Information Needed]
         
     | 
| 91 | 
         
            +
             
     | 
| 92 | 
         
            +
             
     | 
| 93 | 
         
            +
            #### Training Hyperparameters
         
     | 
| 94 | 
         
            +
             
     | 
| 95 | 
         
            +
            - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
         
     | 
| 96 | 
         
            +
             
     | 
| 97 | 
         
            +
            #### Speeds, Sizes, Times [optional]
         
     | 
| 98 | 
         
            +
             
     | 
| 99 | 
         
            +
            <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
         
     | 
| 100 | 
         
            +
             
     | 
| 101 | 
         
            +
            [More Information Needed]
         
     | 
| 102 | 
         
            +
             
     | 
| 103 | 
         
            +
            ## Evaluation
         
     | 
| 104 | 
         
            +
             
     | 
| 105 | 
         
            +
            <!-- This section describes the evaluation protocols and provides the results. -->
         
     | 
| 106 | 
         
            +
             
     | 
| 107 | 
         
            +
            ### Testing Data, Factors & Metrics
         
     | 
| 108 | 
         
            +
             
     | 
| 109 | 
         
            +
            #### Testing Data
         
     | 
| 110 | 
         
            +
             
     | 
| 111 | 
         
            +
            <!-- This should link to a Dataset Card if possible. -->
         
     | 
| 112 | 
         
            +
             
     | 
| 113 | 
         
            +
            [More Information Needed]
         
     | 
| 114 | 
         
            +
             
     | 
| 115 | 
         
            +
            #### Factors
         
     | 
| 116 | 
         
            +
             
     | 
| 117 | 
         
            +
            <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
         
     | 
| 118 | 
         
            +
             
     | 
| 119 | 
         
            +
            [More Information Needed]
         
     | 
| 120 | 
         
            +
             
     | 
| 121 | 
         
            +
            #### Metrics
         
     | 
| 122 | 
         
            +
             
     | 
| 123 | 
         
            +
            <!-- These are the evaluation metrics being used, ideally with a description of why. -->
         
     | 
| 124 | 
         
            +
             
     | 
| 125 | 
         
            +
            [More Information Needed]
         
     | 
| 126 | 
         
            +
             
     | 
| 127 | 
         
            +
            ### Results
         
     | 
| 128 | 
         
            +
             
     | 
| 129 | 
         
            +
            [More Information Needed]
         
     | 
| 130 | 
         
            +
             
     | 
| 131 | 
         
            +
            #### Summary
         
     | 
| 132 | 
         
            +
             
     | 
| 133 | 
         
            +
             
     | 
| 134 | 
         
            +
             
     | 
| 135 | 
         
            +
            ## Model Examination [optional]
         
     | 
| 136 | 
         
            +
             
     | 
| 137 | 
         
            +
            <!-- Relevant interpretability work for the model goes here -->
         
     | 
| 138 | 
         
            +
             
     | 
| 139 | 
         
            +
            [More Information Needed]
         
     | 
| 140 | 
         
            +
             
     | 
| 141 | 
         
            +
            ## Environmental Impact
         
     | 
| 142 | 
         
            +
             
     | 
| 143 | 
         
            +
            <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
         
     | 
| 144 | 
         
            +
             
     | 
| 145 | 
         
            +
            Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
         
     | 
| 146 | 
         
            +
             
     | 
| 147 | 
         
            +
            - **Hardware Type:** [More Information Needed]
         
     | 
| 148 | 
         
            +
            - **Hours used:** [More Information Needed]
         
     | 
| 149 | 
         
            +
            - **Cloud Provider:** [More Information Needed]
         
     | 
| 150 | 
         
            +
            - **Compute Region:** [More Information Needed]
         
     | 
| 151 | 
         
            +
            - **Carbon Emitted:** [More Information Needed]
         
     | 
| 152 | 
         
            +
             
     | 
| 153 | 
         
            +
            ## Technical Specifications [optional]
         
     | 
| 154 | 
         
            +
             
     | 
| 155 | 
         
            +
            ### Model Architecture and Objective
         
     | 
| 156 | 
         
            +
             
     | 
| 157 | 
         
            +
            [More Information Needed]
         
     | 
| 158 | 
         
            +
             
     | 
| 159 | 
         
            +
            ### Compute Infrastructure
         
     | 
| 160 | 
         
            +
             
     | 
| 161 | 
         
            +
            [More Information Needed]
         
     | 
| 162 | 
         
            +
             
     | 
| 163 | 
         
            +
            #### Hardware
         
     | 
| 164 | 
         
            +
             
     | 
| 165 | 
         
            +
            [More Information Needed]
         
     | 
| 166 | 
         
            +
             
     | 
| 167 | 
         
            +
            #### Software
         
     | 
| 168 | 
         
            +
             
     | 
| 169 | 
         
            +
            [More Information Needed]
         
     | 
| 170 | 
         
            +
             
     | 
| 171 | 
         
            +
            ## Citation [optional]
         
     | 
| 172 | 
         
            +
             
     | 
| 173 | 
         
            +
            <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
         
     | 
| 174 | 
         
            +
             
     | 
| 175 | 
         
            +
            **BibTeX:**
         
     | 
| 176 | 
         
            +
             
     | 
| 177 | 
         
            +
            [More Information Needed]
         
     | 
| 178 | 
         
            +
             
     | 
| 179 | 
         
            +
            **APA:**
         
     | 
| 180 | 
         
            +
             
     | 
| 181 | 
         
            +
            [More Information Needed]
         
     | 
| 182 | 
         
            +
             
     | 
| 183 | 
         
            +
            ## Glossary [optional]
         
     | 
| 184 | 
         
            +
             
     | 
| 185 | 
         
            +
            <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
         
     | 
| 186 | 
         
            +
             
     | 
| 187 | 
         
            +
            [More Information Needed]
         
     | 
| 188 | 
         
            +
             
     | 
| 189 | 
         
            +
            ## More Information [optional]
         
     | 
| 190 | 
         
            +
             
     | 
| 191 | 
         
            +
            [More Information Needed]
         
     | 
| 192 | 
         
            +
             
     | 
| 193 | 
         
            +
            ## Model Card Authors [optional]
         
     | 
| 194 | 
         
            +
             
     | 
| 195 | 
         
            +
            [More Information Needed]
         
     | 
| 196 | 
         
            +
             
     | 
| 197 | 
         
            +
            ## Model Card Contact
         
     | 
| 198 | 
         
            +
             
     | 
| 199 | 
         
            +
            [More Information Needed]
         
     | 
| 200 | 
         
            +
            ### Framework versions
         
     | 
| 201 | 
         
            +
             
     | 
| 202 | 
         
            +
            - PEFT 0.13.2
         
     | 
    	
        last-checkpoint/adapter_config.json
    ADDED
    
    | 
         @@ -0,0 +1,30 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "alpha_pattern": {},
         
     | 
| 3 | 
         
            +
              "auto_mapping": null,
         
     | 
| 4 | 
         
            +
              "base_model_name_or_path": "katuni4ka/tiny-random-dbrx",
         
     | 
| 5 | 
         
            +
              "bias": "none",
         
     | 
| 6 | 
         
            +
              "fan_in_fan_out": null,
         
     | 
| 7 | 
         
            +
              "inference_mode": true,
         
     | 
| 8 | 
         
            +
              "init_lora_weights": true,
         
     | 
| 9 | 
         
            +
              "layer_replication": null,
         
     | 
| 10 | 
         
            +
              "layers_pattern": null,
         
     | 
| 11 | 
         
            +
              "layers_to_transform": null,
         
     | 
| 12 | 
         
            +
              "loftq_config": {},
         
     | 
| 13 | 
         
            +
              "lora_alpha": 32,
         
     | 
| 14 | 
         
            +
              "lora_dropout": 0.05,
         
     | 
| 15 | 
         
            +
              "megatron_config": null,
         
     | 
| 16 | 
         
            +
              "megatron_core": "megatron.core",
         
     | 
| 17 | 
         
            +
              "modules_to_save": null,
         
     | 
| 18 | 
         
            +
              "peft_type": "LORA",
         
     | 
| 19 | 
         
            +
              "r": 16,
         
     | 
| 20 | 
         
            +
              "rank_pattern": {},
         
     | 
| 21 | 
         
            +
              "revision": null,
         
     | 
| 22 | 
         
            +
              "target_modules": [
         
     | 
| 23 | 
         
            +
                "Wqkv",
         
     | 
| 24 | 
         
            +
                "layer",
         
     | 
| 25 | 
         
            +
                "out_proj"
         
     | 
| 26 | 
         
            +
              ],
         
     | 
| 27 | 
         
            +
              "task_type": "CAUSAL_LM",
         
     | 
| 28 | 
         
            +
              "use_dora": false,
         
     | 
| 29 | 
         
            +
              "use_rslora": false
         
     | 
| 30 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/adapter_model.safetensors
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:7fb3c8977e1a5e421a3ade55cd1bbd78cf06a024779d83da21b8b40651ae09b8
         
     | 
| 3 | 
         
            +
            size 9864
         
     | 
    	
        last-checkpoint/added_tokens.json
    ADDED
    
    | 
         @@ -0,0 +1,4 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "<|im_end|>": 100279,
         
     | 
| 3 | 
         
            +
              "<|im_start|>": 100278
         
     | 
| 4 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/merges.txt
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        last-checkpoint/optimizer.pt
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:7feeea91225f828b353f57776689a35c3663e7dd2840979ded8d69ddd7cf6a9c
         
     | 
| 3 | 
         
            +
            size 27002
         
     | 
    	
        last-checkpoint/rng_state.pth
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:4bd3758a32379d4c3d2e7cd122013e73b32f081f9c05b9a0277806f4962c622d
         
     | 
| 3 | 
         
            +
            size 14244
         
     | 
    	
        last-checkpoint/scheduler.pt
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c188a6a4749e6ca627bb6d536eb7443f499d5b1b88d98a78f9c713443e010d9c
         
     | 
| 3 | 
         
            +
            size 1064
         
     | 
    	
        last-checkpoint/special_tokens_map.json
    ADDED
    
    | 
         @@ -0,0 +1,30 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "bos_token": {
         
     | 
| 3 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 4 | 
         
            +
                "lstrip": false,
         
     | 
| 5 | 
         
            +
                "normalized": false,
         
     | 
| 6 | 
         
            +
                "rstrip": false,
         
     | 
| 7 | 
         
            +
                "single_word": false
         
     | 
| 8 | 
         
            +
              },
         
     | 
| 9 | 
         
            +
              "eos_token": {
         
     | 
| 10 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 11 | 
         
            +
                "lstrip": false,
         
     | 
| 12 | 
         
            +
                "normalized": false,
         
     | 
| 13 | 
         
            +
                "rstrip": false,
         
     | 
| 14 | 
         
            +
                "single_word": false
         
     | 
| 15 | 
         
            +
              },
         
     | 
| 16 | 
         
            +
              "pad_token": {
         
     | 
| 17 | 
         
            +
                "content": "<|pad|>",
         
     | 
| 18 | 
         
            +
                "lstrip": false,
         
     | 
| 19 | 
         
            +
                "normalized": false,
         
     | 
| 20 | 
         
            +
                "rstrip": false,
         
     | 
| 21 | 
         
            +
                "single_word": false
         
     | 
| 22 | 
         
            +
              },
         
     | 
| 23 | 
         
            +
              "unk_token": {
         
     | 
| 24 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 25 | 
         
            +
                "lstrip": false,
         
     | 
| 26 | 
         
            +
                "normalized": false,
         
     | 
| 27 | 
         
            +
                "rstrip": false,
         
     | 
| 28 | 
         
            +
                "single_word": false
         
     | 
| 29 | 
         
            +
              }
         
     | 
| 30 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/tokenizer.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        last-checkpoint/tokenizer_config.json
    ADDED
    
    | 
         @@ -0,0 +1,205 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "add_prefix_space": false,
         
     | 
| 3 | 
         
            +
              "added_tokens_decoder": {
         
     | 
| 4 | 
         
            +
                "100256": {
         
     | 
| 5 | 
         
            +
                  "content": "<||_unused_0_||>",
         
     | 
| 6 | 
         
            +
                  "lstrip": false,
         
     | 
| 7 | 
         
            +
                  "normalized": false,
         
     | 
| 8 | 
         
            +
                  "rstrip": false,
         
     | 
| 9 | 
         
            +
                  "single_word": false,
         
     | 
| 10 | 
         
            +
                  "special": true
         
     | 
| 11 | 
         
            +
                },
         
     | 
| 12 | 
         
            +
                "100257": {
         
     | 
| 13 | 
         
            +
                  "content": "<|endoftext|>",
         
     | 
| 14 | 
         
            +
                  "lstrip": false,
         
     | 
| 15 | 
         
            +
                  "normalized": false,
         
     | 
| 16 | 
         
            +
                  "rstrip": false,
         
     | 
| 17 | 
         
            +
                  "single_word": false,
         
     | 
| 18 | 
         
            +
                  "special": true
         
     | 
| 19 | 
         
            +
                },
         
     | 
| 20 | 
         
            +
                "100258": {
         
     | 
| 21 | 
         
            +
                  "content": "<|fim_prefix|>",
         
     | 
| 22 | 
         
            +
                  "lstrip": false,
         
     | 
| 23 | 
         
            +
                  "normalized": false,
         
     | 
| 24 | 
         
            +
                  "rstrip": false,
         
     | 
| 25 | 
         
            +
                  "single_word": false,
         
     | 
| 26 | 
         
            +
                  "special": true
         
     | 
| 27 | 
         
            +
                },
         
     | 
| 28 | 
         
            +
                "100259": {
         
     | 
| 29 | 
         
            +
                  "content": "<|fim_middle|>",
         
     | 
| 30 | 
         
            +
                  "lstrip": false,
         
     | 
| 31 | 
         
            +
                  "normalized": false,
         
     | 
| 32 | 
         
            +
                  "rstrip": false,
         
     | 
| 33 | 
         
            +
                  "single_word": false,
         
     | 
| 34 | 
         
            +
                  "special": true
         
     | 
| 35 | 
         
            +
                },
         
     | 
| 36 | 
         
            +
                "100260": {
         
     | 
| 37 | 
         
            +
                  "content": "<|fim_suffix|>",
         
     | 
| 38 | 
         
            +
                  "lstrip": false,
         
     | 
| 39 | 
         
            +
                  "normalized": false,
         
     | 
| 40 | 
         
            +
                  "rstrip": false,
         
     | 
| 41 | 
         
            +
                  "single_word": false,
         
     | 
| 42 | 
         
            +
                  "special": true
         
     | 
| 43 | 
         
            +
                },
         
     | 
| 44 | 
         
            +
                "100261": {
         
     | 
| 45 | 
         
            +
                  "content": "<||_unused_1_||>",
         
     | 
| 46 | 
         
            +
                  "lstrip": false,
         
     | 
| 47 | 
         
            +
                  "normalized": false,
         
     | 
| 48 | 
         
            +
                  "rstrip": false,
         
     | 
| 49 | 
         
            +
                  "single_word": false,
         
     | 
| 50 | 
         
            +
                  "special": true
         
     | 
| 51 | 
         
            +
                },
         
     | 
| 52 | 
         
            +
                "100262": {
         
     | 
| 53 | 
         
            +
                  "content": "<||_unused_2_||>",
         
     | 
| 54 | 
         
            +
                  "lstrip": false,
         
     | 
| 55 | 
         
            +
                  "normalized": false,
         
     | 
| 56 | 
         
            +
                  "rstrip": false,
         
     | 
| 57 | 
         
            +
                  "single_word": false,
         
     | 
| 58 | 
         
            +
                  "special": true
         
     | 
| 59 | 
         
            +
                },
         
     | 
| 60 | 
         
            +
                "100263": {
         
     | 
| 61 | 
         
            +
                  "content": "<||_unused_3_||>",
         
     | 
| 62 | 
         
            +
                  "lstrip": false,
         
     | 
| 63 | 
         
            +
                  "normalized": false,
         
     | 
| 64 | 
         
            +
                  "rstrip": false,
         
     | 
| 65 | 
         
            +
                  "single_word": false,
         
     | 
| 66 | 
         
            +
                  "special": true
         
     | 
| 67 | 
         
            +
                },
         
     | 
| 68 | 
         
            +
                "100264": {
         
     | 
| 69 | 
         
            +
                  "content": "<||_unused_4_||>",
         
     | 
| 70 | 
         
            +
                  "lstrip": false,
         
     | 
| 71 | 
         
            +
                  "normalized": false,
         
     | 
| 72 | 
         
            +
                  "rstrip": false,
         
     | 
| 73 | 
         
            +
                  "single_word": false,
         
     | 
| 74 | 
         
            +
                  "special": true
         
     | 
| 75 | 
         
            +
                },
         
     | 
| 76 | 
         
            +
                "100265": {
         
     | 
| 77 | 
         
            +
                  "content": "<||_unused_5_||>",
         
     | 
| 78 | 
         
            +
                  "lstrip": false,
         
     | 
| 79 | 
         
            +
                  "normalized": false,
         
     | 
| 80 | 
         
            +
                  "rstrip": false,
         
     | 
| 81 | 
         
            +
                  "single_word": false,
         
     | 
| 82 | 
         
            +
                  "special": true
         
     | 
| 83 | 
         
            +
                },
         
     | 
| 84 | 
         
            +
                "100266": {
         
     | 
| 85 | 
         
            +
                  "content": "<||_unused_6_||>",
         
     | 
| 86 | 
         
            +
                  "lstrip": false,
         
     | 
| 87 | 
         
            +
                  "normalized": false,
         
     | 
| 88 | 
         
            +
                  "rstrip": false,
         
     | 
| 89 | 
         
            +
                  "single_word": false,
         
     | 
| 90 | 
         
            +
                  "special": true
         
     | 
| 91 | 
         
            +
                },
         
     | 
| 92 | 
         
            +
                "100267": {
         
     | 
| 93 | 
         
            +
                  "content": "<||_unused_7_||>",
         
     | 
| 94 | 
         
            +
                  "lstrip": false,
         
     | 
| 95 | 
         
            +
                  "normalized": false,
         
     | 
| 96 | 
         
            +
                  "rstrip": false,
         
     | 
| 97 | 
         
            +
                  "single_word": false,
         
     | 
| 98 | 
         
            +
                  "special": true
         
     | 
| 99 | 
         
            +
                },
         
     | 
| 100 | 
         
            +
                "100268": {
         
     | 
| 101 | 
         
            +
                  "content": "<||_unused_8_||>",
         
     | 
| 102 | 
         
            +
                  "lstrip": false,
         
     | 
| 103 | 
         
            +
                  "normalized": false,
         
     | 
| 104 | 
         
            +
                  "rstrip": false,
         
     | 
| 105 | 
         
            +
                  "single_word": false,
         
     | 
| 106 | 
         
            +
                  "special": true
         
     | 
| 107 | 
         
            +
                },
         
     | 
| 108 | 
         
            +
                "100269": {
         
     | 
| 109 | 
         
            +
                  "content": "<||_unused_9_||>",
         
     | 
| 110 | 
         
            +
                  "lstrip": false,
         
     | 
| 111 | 
         
            +
                  "normalized": false,
         
     | 
| 112 | 
         
            +
                  "rstrip": false,
         
     | 
| 113 | 
         
            +
                  "single_word": false,
         
     | 
| 114 | 
         
            +
                  "special": true
         
     | 
| 115 | 
         
            +
                },
         
     | 
| 116 | 
         
            +
                "100270": {
         
     | 
| 117 | 
         
            +
                  "content": "<||_unused_10_||>",
         
     | 
| 118 | 
         
            +
                  "lstrip": false,
         
     | 
| 119 | 
         
            +
                  "normalized": false,
         
     | 
| 120 | 
         
            +
                  "rstrip": false,
         
     | 
| 121 | 
         
            +
                  "single_word": false,
         
     | 
| 122 | 
         
            +
                  "special": true
         
     | 
| 123 | 
         
            +
                },
         
     | 
| 124 | 
         
            +
                "100271": {
         
     | 
| 125 | 
         
            +
                  "content": "<||_unused_11_||>",
         
     | 
| 126 | 
         
            +
                  "lstrip": false,
         
     | 
| 127 | 
         
            +
                  "normalized": false,
         
     | 
| 128 | 
         
            +
                  "rstrip": false,
         
     | 
| 129 | 
         
            +
                  "single_word": false,
         
     | 
| 130 | 
         
            +
                  "special": true
         
     | 
| 131 | 
         
            +
                },
         
     | 
| 132 | 
         
            +
                "100272": {
         
     | 
| 133 | 
         
            +
                  "content": "<||_unused_12_||>",
         
     | 
| 134 | 
         
            +
                  "lstrip": false,
         
     | 
| 135 | 
         
            +
                  "normalized": false,
         
     | 
| 136 | 
         
            +
                  "rstrip": false,
         
     | 
| 137 | 
         
            +
                  "single_word": false,
         
     | 
| 138 | 
         
            +
                  "special": true
         
     | 
| 139 | 
         
            +
                },
         
     | 
| 140 | 
         
            +
                "100273": {
         
     | 
| 141 | 
         
            +
                  "content": "<||_unused_13_||>",
         
     | 
| 142 | 
         
            +
                  "lstrip": false,
         
     | 
| 143 | 
         
            +
                  "normalized": false,
         
     | 
| 144 | 
         
            +
                  "rstrip": false,
         
     | 
| 145 | 
         
            +
                  "single_word": false,
         
     | 
| 146 | 
         
            +
                  "special": true
         
     | 
| 147 | 
         
            +
                },
         
     | 
| 148 | 
         
            +
                "100274": {
         
     | 
| 149 | 
         
            +
                  "content": "<||_unused_14_||>",
         
     | 
| 150 | 
         
            +
                  "lstrip": false,
         
     | 
| 151 | 
         
            +
                  "normalized": false,
         
     | 
| 152 | 
         
            +
                  "rstrip": false,
         
     | 
| 153 | 
         
            +
                  "single_word": false,
         
     | 
| 154 | 
         
            +
                  "special": true
         
     | 
| 155 | 
         
            +
                },
         
     | 
| 156 | 
         
            +
                "100275": {
         
     | 
| 157 | 
         
            +
                  "content": "<||_unused_15_||>",
         
     | 
| 158 | 
         
            +
                  "lstrip": false,
         
     | 
| 159 | 
         
            +
                  "normalized": false,
         
     | 
| 160 | 
         
            +
                  "rstrip": false,
         
     | 
| 161 | 
         
            +
                  "single_word": false,
         
     | 
| 162 | 
         
            +
                  "special": true
         
     | 
| 163 | 
         
            +
                },
         
     | 
| 164 | 
         
            +
                "100276": {
         
     | 
| 165 | 
         
            +
                  "content": "<|endofprompt|>",
         
     | 
| 166 | 
         
            +
                  "lstrip": false,
         
     | 
| 167 | 
         
            +
                  "normalized": false,
         
     | 
| 168 | 
         
            +
                  "rstrip": false,
         
     | 
| 169 | 
         
            +
                  "single_word": false,
         
     | 
| 170 | 
         
            +
                  "special": true
         
     | 
| 171 | 
         
            +
                },
         
     | 
| 172 | 
         
            +
                "100277": {
         
     | 
| 173 | 
         
            +
                  "content": "<|pad|>",
         
     | 
| 174 | 
         
            +
                  "lstrip": false,
         
     | 
| 175 | 
         
            +
                  "normalized": false,
         
     | 
| 176 | 
         
            +
                  "rstrip": false,
         
     | 
| 177 | 
         
            +
                  "single_word": false,
         
     | 
| 178 | 
         
            +
                  "special": true
         
     | 
| 179 | 
         
            +
                },
         
     | 
| 180 | 
         
            +
                "100278": {
         
     | 
| 181 | 
         
            +
                  "content": "<|im_start|>",
         
     | 
| 182 | 
         
            +
                  "lstrip": false,
         
     | 
| 183 | 
         
            +
                  "normalized": false,
         
     | 
| 184 | 
         
            +
                  "rstrip": false,
         
     | 
| 185 | 
         
            +
                  "single_word": false,
         
     | 
| 186 | 
         
            +
                  "special": true
         
     | 
| 187 | 
         
            +
                },
         
     | 
| 188 | 
         
            +
                "100279": {
         
     | 
| 189 | 
         
            +
                  "content": "<|im_end|>",
         
     | 
| 190 | 
         
            +
                  "lstrip": false,
         
     | 
| 191 | 
         
            +
                  "normalized": false,
         
     | 
| 192 | 
         
            +
                  "rstrip": false,
         
     | 
| 193 | 
         
            +
                  "single_word": false,
         
     | 
| 194 | 
         
            +
                  "special": true
         
     | 
| 195 | 
         
            +
                }
         
     | 
| 196 | 
         
            +
              },
         
     | 
| 197 | 
         
            +
              "bos_token": "<|endoftext|>",
         
     | 
| 198 | 
         
            +
              "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
         
     | 
| 199 | 
         
            +
              "clean_up_tokenization_spaces": true,
         
     | 
| 200 | 
         
            +
              "eos_token": "<|endoftext|>",
         
     | 
| 201 | 
         
            +
              "model_max_length": 32768,
         
     | 
| 202 | 
         
            +
              "pad_token": "<|pad|>",
         
     | 
| 203 | 
         
            +
              "tokenizer_class": "GPT2Tokenizer",
         
     | 
| 204 | 
         
            +
              "unk_token": "<|endoftext|>"
         
     | 
| 205 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/trainer_state.json
    ADDED
    
    | 
         @@ -0,0 +1,87 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "best_metric": 11.5,
         
     | 
| 3 | 
         
            +
              "best_model_checkpoint": "miner_id_24/checkpoint-10",
         
     | 
| 4 | 
         
            +
              "epoch": 0.0014116816657843657,
         
     | 
| 5 | 
         
            +
              "eval_steps": 5,
         
     | 
| 6 | 
         
            +
              "global_step": 10,
         
     | 
| 7 | 
         
            +
              "is_hyper_param_search": false,
         
     | 
| 8 | 
         
            +
              "is_local_process_zero": true,
         
     | 
| 9 | 
         
            +
              "is_world_process_zero": true,
         
     | 
| 10 | 
         
            +
              "log_history": [
         
     | 
| 11 | 
         
            +
                {
         
     | 
| 12 | 
         
            +
                  "epoch": 0.00014116816657843657,
         
     | 
| 13 | 
         
            +
                  "eval_loss": 11.5,
         
     | 
| 14 | 
         
            +
                  "eval_runtime": 43.8797,
         
     | 
| 15 | 
         
            +
                  "eval_samples_per_second": 67.981,
         
     | 
| 16 | 
         
            +
                  "eval_steps_per_second": 34.002,
         
     | 
| 17 | 
         
            +
                  "step": 1
         
     | 
| 18 | 
         
            +
                },
         
     | 
| 19 | 
         
            +
                {
         
     | 
| 20 | 
         
            +
                  "epoch": 0.00042350449973530967,
         
     | 
| 21 | 
         
            +
                  "grad_norm": 1.9433709894656204e-05,
         
     | 
| 22 | 
         
            +
                  "learning_rate": 0.00012,
         
     | 
| 23 | 
         
            +
                  "loss": 46.0,
         
     | 
| 24 | 
         
            +
                  "step": 3
         
     | 
| 25 | 
         
            +
                },
         
     | 
| 26 | 
         
            +
                {
         
     | 
| 27 | 
         
            +
                  "epoch": 0.0007058408328921829,
         
     | 
| 28 | 
         
            +
                  "eval_loss": 11.5,
         
     | 
| 29 | 
         
            +
                  "eval_runtime": 44.5231,
         
     | 
| 30 | 
         
            +
                  "eval_samples_per_second": 66.999,
         
     | 
| 31 | 
         
            +
                  "eval_steps_per_second": 33.511,
         
     | 
| 32 | 
         
            +
                  "step": 5
         
     | 
| 33 | 
         
            +
                },
         
     | 
| 34 | 
         
            +
                {
         
     | 
| 35 | 
         
            +
                  "epoch": 0.0008470089994706193,
         
     | 
| 36 | 
         
            +
                  "grad_norm": 1.735346631903667e-05,
         
     | 
| 37 | 
         
            +
                  "learning_rate": 0.0001992114701314478,
         
     | 
| 38 | 
         
            +
                  "loss": 46.0,
         
     | 
| 39 | 
         
            +
                  "step": 6
         
     | 
| 40 | 
         
            +
                },
         
     | 
| 41 | 
         
            +
                {
         
     | 
| 42 | 
         
            +
                  "epoch": 0.001270513499205929,
         
     | 
| 43 | 
         
            +
                  "grad_norm": 3.283464593550889e-06,
         
     | 
| 44 | 
         
            +
                  "learning_rate": 0.00018763066800438636,
         
     | 
| 45 | 
         
            +
                  "loss": 46.0,
         
     | 
| 46 | 
         
            +
                  "step": 9
         
     | 
| 47 | 
         
            +
                },
         
     | 
| 48 | 
         
            +
                {
         
     | 
| 49 | 
         
            +
                  "epoch": 0.0014116816657843657,
         
     | 
| 50 | 
         
            +
                  "eval_loss": 11.5,
         
     | 
| 51 | 
         
            +
                  "eval_runtime": 44.411,
         
     | 
| 52 | 
         
            +
                  "eval_samples_per_second": 67.168,
         
     | 
| 53 | 
         
            +
                  "eval_steps_per_second": 33.595,
         
     | 
| 54 | 
         
            +
                  "step": 10
         
     | 
| 55 | 
         
            +
                }
         
     | 
| 56 | 
         
            +
              ],
         
     | 
| 57 | 
         
            +
              "logging_steps": 3,
         
     | 
| 58 | 
         
            +
              "max_steps": 30,
         
     | 
| 59 | 
         
            +
              "num_input_tokens_seen": 0,
         
     | 
| 60 | 
         
            +
              "num_train_epochs": 1,
         
     | 
| 61 | 
         
            +
              "save_steps": 10,
         
     | 
| 62 | 
         
            +
              "stateful_callbacks": {
         
     | 
| 63 | 
         
            +
                "EarlyStoppingCallback": {
         
     | 
| 64 | 
         
            +
                  "args": {
         
     | 
| 65 | 
         
            +
                    "early_stopping_patience": 1,
         
     | 
| 66 | 
         
            +
                    "early_stopping_threshold": 0.0
         
     | 
| 67 | 
         
            +
                  },
         
     | 
| 68 | 
         
            +
                  "attributes": {
         
     | 
| 69 | 
         
            +
                    "early_stopping_patience_counter": 0
         
     | 
| 70 | 
         
            +
                  }
         
     | 
| 71 | 
         
            +
                },
         
     | 
| 72 | 
         
            +
                "TrainerControl": {
         
     | 
| 73 | 
         
            +
                  "args": {
         
     | 
| 74 | 
         
            +
                    "should_epoch_stop": false,
         
     | 
| 75 | 
         
            +
                    "should_evaluate": false,
         
     | 
| 76 | 
         
            +
                    "should_log": false,
         
     | 
| 77 | 
         
            +
                    "should_save": true,
         
     | 
| 78 | 
         
            +
                    "should_training_stop": false
         
     | 
| 79 | 
         
            +
                  },
         
     | 
| 80 | 
         
            +
                  "attributes": {}
         
     | 
| 81 | 
         
            +
                }
         
     | 
| 82 | 
         
            +
              },
         
     | 
| 83 | 
         
            +
              "total_flos": 728103616512.0,
         
     | 
| 84 | 
         
            +
              "train_batch_size": 2,
         
     | 
| 85 | 
         
            +
              "trial_name": null,
         
     | 
| 86 | 
         
            +
              "trial_params": null
         
     | 
| 87 | 
         
            +
            }
         
     | 
    	
        last-checkpoint/training_args.bin
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:558e632507a27f45387ac987b88f12bd2a9ae4e0ef42e0bc4ae214db93f27234
         
     | 
| 3 | 
         
            +
            size 6776
         
     | 
    	
        last-checkpoint/vocab.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        merges.txt
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        special_tokens_map.json
    ADDED
    
    | 
         @@ -0,0 +1,30 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "bos_token": {
         
     | 
| 3 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 4 | 
         
            +
                "lstrip": false,
         
     | 
| 5 | 
         
            +
                "normalized": false,
         
     | 
| 6 | 
         
            +
                "rstrip": false,
         
     | 
| 7 | 
         
            +
                "single_word": false
         
     | 
| 8 | 
         
            +
              },
         
     | 
| 9 | 
         
            +
              "eos_token": {
         
     | 
| 10 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 11 | 
         
            +
                "lstrip": false,
         
     | 
| 12 | 
         
            +
                "normalized": false,
         
     | 
| 13 | 
         
            +
                "rstrip": false,
         
     | 
| 14 | 
         
            +
                "single_word": false
         
     | 
| 15 | 
         
            +
              },
         
     | 
| 16 | 
         
            +
              "pad_token": {
         
     | 
| 17 | 
         
            +
                "content": "<|pad|>",
         
     | 
| 18 | 
         
            +
                "lstrip": false,
         
     | 
| 19 | 
         
            +
                "normalized": false,
         
     | 
| 20 | 
         
            +
                "rstrip": false,
         
     | 
| 21 | 
         
            +
                "single_word": false
         
     | 
| 22 | 
         
            +
              },
         
     | 
| 23 | 
         
            +
              "unk_token": {
         
     | 
| 24 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 25 | 
         
            +
                "lstrip": false,
         
     | 
| 26 | 
         
            +
                "normalized": false,
         
     | 
| 27 | 
         
            +
                "rstrip": false,
         
     | 
| 28 | 
         
            +
                "single_word": false
         
     | 
| 29 | 
         
            +
              }
         
     | 
| 30 | 
         
            +
            }
         
     | 
    	
        tokenizer.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        tokenizer_config.json
    ADDED
    
    | 
         @@ -0,0 +1,205 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "add_prefix_space": false,
         
     | 
| 3 | 
         
            +
              "added_tokens_decoder": {
         
     | 
| 4 | 
         
            +
                "100256": {
         
     | 
| 5 | 
         
            +
                  "content": "<||_unused_0_||>",
         
     | 
| 6 | 
         
            +
                  "lstrip": false,
         
     | 
| 7 | 
         
            +
                  "normalized": false,
         
     | 
| 8 | 
         
            +
                  "rstrip": false,
         
     | 
| 9 | 
         
            +
                  "single_word": false,
         
     | 
| 10 | 
         
            +
                  "special": true
         
     | 
| 11 | 
         
            +
                },
         
     | 
| 12 | 
         
            +
                "100257": {
         
     | 
| 13 | 
         
            +
                  "content": "<|endoftext|>",
         
     | 
| 14 | 
         
            +
                  "lstrip": false,
         
     | 
| 15 | 
         
            +
                  "normalized": false,
         
     | 
| 16 | 
         
            +
                  "rstrip": false,
         
     | 
| 17 | 
         
            +
                  "single_word": false,
         
     | 
| 18 | 
         
            +
                  "special": true
         
     | 
| 19 | 
         
            +
                },
         
     | 
| 20 | 
         
            +
                "100258": {
         
     | 
| 21 | 
         
            +
                  "content": "<|fim_prefix|>",
         
     | 
| 22 | 
         
            +
                  "lstrip": false,
         
     | 
| 23 | 
         
            +
                  "normalized": false,
         
     | 
| 24 | 
         
            +
                  "rstrip": false,
         
     | 
| 25 | 
         
            +
                  "single_word": false,
         
     | 
| 26 | 
         
            +
                  "special": true
         
     | 
| 27 | 
         
            +
                },
         
     | 
| 28 | 
         
            +
                "100259": {
         
     | 
| 29 | 
         
            +
                  "content": "<|fim_middle|>",
         
     | 
| 30 | 
         
            +
                  "lstrip": false,
         
     | 
| 31 | 
         
            +
                  "normalized": false,
         
     | 
| 32 | 
         
            +
                  "rstrip": false,
         
     | 
| 33 | 
         
            +
                  "single_word": false,
         
     | 
| 34 | 
         
            +
                  "special": true
         
     | 
| 35 | 
         
            +
                },
         
     | 
| 36 | 
         
            +
                "100260": {
         
     | 
| 37 | 
         
            +
                  "content": "<|fim_suffix|>",
         
     | 
| 38 | 
         
            +
                  "lstrip": false,
         
     | 
| 39 | 
         
            +
                  "normalized": false,
         
     | 
| 40 | 
         
            +
                  "rstrip": false,
         
     | 
| 41 | 
         
            +
                  "single_word": false,
         
     | 
| 42 | 
         
            +
                  "special": true
         
     | 
| 43 | 
         
            +
                },
         
     | 
| 44 | 
         
            +
                "100261": {
         
     | 
| 45 | 
         
            +
                  "content": "<||_unused_1_||>",
         
     | 
| 46 | 
         
            +
                  "lstrip": false,
         
     | 
| 47 | 
         
            +
                  "normalized": false,
         
     | 
| 48 | 
         
            +
                  "rstrip": false,
         
     | 
| 49 | 
         
            +
                  "single_word": false,
         
     | 
| 50 | 
         
            +
                  "special": true
         
     | 
| 51 | 
         
            +
                },
         
     | 
| 52 | 
         
            +
                "100262": {
         
     | 
| 53 | 
         
            +
                  "content": "<||_unused_2_||>",
         
     | 
| 54 | 
         
            +
                  "lstrip": false,
         
     | 
| 55 | 
         
            +
                  "normalized": false,
         
     | 
| 56 | 
         
            +
                  "rstrip": false,
         
     | 
| 57 | 
         
            +
                  "single_word": false,
         
     | 
| 58 | 
         
            +
                  "special": true
         
     | 
| 59 | 
         
            +
                },
         
     | 
| 60 | 
         
            +
                "100263": {
         
     | 
| 61 | 
         
            +
                  "content": "<||_unused_3_||>",
         
     | 
| 62 | 
         
            +
                  "lstrip": false,
         
     | 
| 63 | 
         
            +
                  "normalized": false,
         
     | 
| 64 | 
         
            +
                  "rstrip": false,
         
     | 
| 65 | 
         
            +
                  "single_word": false,
         
     | 
| 66 | 
         
            +
                  "special": true
         
     | 
| 67 | 
         
            +
                },
         
     | 
| 68 | 
         
            +
                "100264": {
         
     | 
| 69 | 
         
            +
                  "content": "<||_unused_4_||>",
         
     | 
| 70 | 
         
            +
                  "lstrip": false,
         
     | 
| 71 | 
         
            +
                  "normalized": false,
         
     | 
| 72 | 
         
            +
                  "rstrip": false,
         
     | 
| 73 | 
         
            +
                  "single_word": false,
         
     | 
| 74 | 
         
            +
                  "special": true
         
     | 
| 75 | 
         
            +
                },
         
     | 
| 76 | 
         
            +
                "100265": {
         
     | 
| 77 | 
         
            +
                  "content": "<||_unused_5_||>",
         
     | 
| 78 | 
         
            +
                  "lstrip": false,
         
     | 
| 79 | 
         
            +
                  "normalized": false,
         
     | 
| 80 | 
         
            +
                  "rstrip": false,
         
     | 
| 81 | 
         
            +
                  "single_word": false,
         
     | 
| 82 | 
         
            +
                  "special": true
         
     | 
| 83 | 
         
            +
                },
         
     | 
| 84 | 
         
            +
                "100266": {
         
     | 
| 85 | 
         
            +
                  "content": "<||_unused_6_||>",
         
     | 
| 86 | 
         
            +
                  "lstrip": false,
         
     | 
| 87 | 
         
            +
                  "normalized": false,
         
     | 
| 88 | 
         
            +
                  "rstrip": false,
         
     | 
| 89 | 
         
            +
                  "single_word": false,
         
     | 
| 90 | 
         
            +
                  "special": true
         
     | 
| 91 | 
         
            +
                },
         
     | 
| 92 | 
         
            +
                "100267": {
         
     | 
| 93 | 
         
            +
                  "content": "<||_unused_7_||>",
         
     | 
| 94 | 
         
            +
                  "lstrip": false,
         
     | 
| 95 | 
         
            +
                  "normalized": false,
         
     | 
| 96 | 
         
            +
                  "rstrip": false,
         
     | 
| 97 | 
         
            +
                  "single_word": false,
         
     | 
| 98 | 
         
            +
                  "special": true
         
     | 
| 99 | 
         
            +
                },
         
     | 
| 100 | 
         
            +
                "100268": {
         
     | 
| 101 | 
         
            +
                  "content": "<||_unused_8_||>",
         
     | 
| 102 | 
         
            +
                  "lstrip": false,
         
     | 
| 103 | 
         
            +
                  "normalized": false,
         
     | 
| 104 | 
         
            +
                  "rstrip": false,
         
     | 
| 105 | 
         
            +
                  "single_word": false,
         
     | 
| 106 | 
         
            +
                  "special": true
         
     | 
| 107 | 
         
            +
                },
         
     | 
| 108 | 
         
            +
                "100269": {
         
     | 
| 109 | 
         
            +
                  "content": "<||_unused_9_||>",
         
     | 
| 110 | 
         
            +
                  "lstrip": false,
         
     | 
| 111 | 
         
            +
                  "normalized": false,
         
     | 
| 112 | 
         
            +
                  "rstrip": false,
         
     | 
| 113 | 
         
            +
                  "single_word": false,
         
     | 
| 114 | 
         
            +
                  "special": true
         
     | 
| 115 | 
         
            +
                },
         
     | 
| 116 | 
         
            +
                "100270": {
         
     | 
| 117 | 
         
            +
                  "content": "<||_unused_10_||>",
         
     | 
| 118 | 
         
            +
                  "lstrip": false,
         
     | 
| 119 | 
         
            +
                  "normalized": false,
         
     | 
| 120 | 
         
            +
                  "rstrip": false,
         
     | 
| 121 | 
         
            +
                  "single_word": false,
         
     | 
| 122 | 
         
            +
                  "special": true
         
     | 
| 123 | 
         
            +
                },
         
     | 
| 124 | 
         
            +
                "100271": {
         
     | 
| 125 | 
         
            +
                  "content": "<||_unused_11_||>",
         
     | 
| 126 | 
         
            +
                  "lstrip": false,
         
     | 
| 127 | 
         
            +
                  "normalized": false,
         
     | 
| 128 | 
         
            +
                  "rstrip": false,
         
     | 
| 129 | 
         
            +
                  "single_word": false,
         
     | 
| 130 | 
         
            +
                  "special": true
         
     | 
| 131 | 
         
            +
                },
         
     | 
| 132 | 
         
            +
                "100272": {
         
     | 
| 133 | 
         
            +
                  "content": "<||_unused_12_||>",
         
     | 
| 134 | 
         
            +
                  "lstrip": false,
         
     | 
| 135 | 
         
            +
                  "normalized": false,
         
     | 
| 136 | 
         
            +
                  "rstrip": false,
         
     | 
| 137 | 
         
            +
                  "single_word": false,
         
     | 
| 138 | 
         
            +
                  "special": true
         
     | 
| 139 | 
         
            +
                },
         
     | 
| 140 | 
         
            +
                "100273": {
         
     | 
| 141 | 
         
            +
                  "content": "<||_unused_13_||>",
         
     | 
| 142 | 
         
            +
                  "lstrip": false,
         
     | 
| 143 | 
         
            +
                  "normalized": false,
         
     | 
| 144 | 
         
            +
                  "rstrip": false,
         
     | 
| 145 | 
         
            +
                  "single_word": false,
         
     | 
| 146 | 
         
            +
                  "special": true
         
     | 
| 147 | 
         
            +
                },
         
     | 
| 148 | 
         
            +
                "100274": {
         
     | 
| 149 | 
         
            +
                  "content": "<||_unused_14_||>",
         
     | 
| 150 | 
         
            +
                  "lstrip": false,
         
     | 
| 151 | 
         
            +
                  "normalized": false,
         
     | 
| 152 | 
         
            +
                  "rstrip": false,
         
     | 
| 153 | 
         
            +
                  "single_word": false,
         
     | 
| 154 | 
         
            +
                  "special": true
         
     | 
| 155 | 
         
            +
                },
         
     | 
| 156 | 
         
            +
                "100275": {
         
     | 
| 157 | 
         
            +
                  "content": "<||_unused_15_||>",
         
     | 
| 158 | 
         
            +
                  "lstrip": false,
         
     | 
| 159 | 
         
            +
                  "normalized": false,
         
     | 
| 160 | 
         
            +
                  "rstrip": false,
         
     | 
| 161 | 
         
            +
                  "single_word": false,
         
     | 
| 162 | 
         
            +
                  "special": true
         
     | 
| 163 | 
         
            +
                },
         
     | 
| 164 | 
         
            +
                "100276": {
         
     | 
| 165 | 
         
            +
                  "content": "<|endofprompt|>",
         
     | 
| 166 | 
         
            +
                  "lstrip": false,
         
     | 
| 167 | 
         
            +
                  "normalized": false,
         
     | 
| 168 | 
         
            +
                  "rstrip": false,
         
     | 
| 169 | 
         
            +
                  "single_word": false,
         
     | 
| 170 | 
         
            +
                  "special": true
         
     | 
| 171 | 
         
            +
                },
         
     | 
| 172 | 
         
            +
                "100277": {
         
     | 
| 173 | 
         
            +
                  "content": "<|pad|>",
         
     | 
| 174 | 
         
            +
                  "lstrip": false,
         
     | 
| 175 | 
         
            +
                  "normalized": false,
         
     | 
| 176 | 
         
            +
                  "rstrip": false,
         
     | 
| 177 | 
         
            +
                  "single_word": false,
         
     | 
| 178 | 
         
            +
                  "special": true
         
     | 
| 179 | 
         
            +
                },
         
     | 
| 180 | 
         
            +
                "100278": {
         
     | 
| 181 | 
         
            +
                  "content": "<|im_start|>",
         
     | 
| 182 | 
         
            +
                  "lstrip": false,
         
     | 
| 183 | 
         
            +
                  "normalized": false,
         
     | 
| 184 | 
         
            +
                  "rstrip": false,
         
     | 
| 185 | 
         
            +
                  "single_word": false,
         
     | 
| 186 | 
         
            +
                  "special": true
         
     | 
| 187 | 
         
            +
                },
         
     | 
| 188 | 
         
            +
                "100279": {
         
     | 
| 189 | 
         
            +
                  "content": "<|im_end|>",
         
     | 
| 190 | 
         
            +
                  "lstrip": false,
         
     | 
| 191 | 
         
            +
                  "normalized": false,
         
     | 
| 192 | 
         
            +
                  "rstrip": false,
         
     | 
| 193 | 
         
            +
                  "single_word": false,
         
     | 
| 194 | 
         
            +
                  "special": true
         
     | 
| 195 | 
         
            +
                }
         
     | 
| 196 | 
         
            +
              },
         
     | 
| 197 | 
         
            +
              "bos_token": "<|endoftext|>",
         
     | 
| 198 | 
         
            +
              "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
         
     | 
| 199 | 
         
            +
              "clean_up_tokenization_spaces": true,
         
     | 
| 200 | 
         
            +
              "eos_token": "<|endoftext|>",
         
     | 
| 201 | 
         
            +
              "model_max_length": 32768,
         
     | 
| 202 | 
         
            +
              "pad_token": "<|pad|>",
         
     | 
| 203 | 
         
            +
              "tokenizer_class": "GPT2Tokenizer",
         
     | 
| 204 | 
         
            +
              "unk_token": "<|endoftext|>"
         
     | 
| 205 | 
         
            +
            }
         
     | 
    	
        training_args.bin
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:558e632507a27f45387ac987b88f12bd2a9ae4e0ef42e0bc4ae214db93f27234
         
     | 
| 3 | 
         
            +
            size 6776
         
     | 
    	
        vocab.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         |