Training in progress, step 20
Browse files- axolotl_config.yaml +12 -6
- model.safetensors +1 -1
- training_args.bin +2 -2
    	
        axolotl_config.yaml
    CHANGED
    
    | @@ -1,5 +1,5 @@ | |
| 1 | 
             
            base_model: echarlaix/tiny-random-mistral
         | 
| 2 | 
            -
            batch_size:  | 
| 3 | 
             
            bf16: true
         | 
| 4 | 
             
            chat_template: tokenizer_default_fallback_alpaca
         | 
| 5 | 
             
            datasets:
         | 
| @@ -13,29 +13,35 @@ datasets: | |
| 13 | 
             
                no_input_format: '{instruction}'
         | 
| 14 | 
             
                system_format: '{system}'
         | 
| 15 | 
             
                system_prompt: ''
         | 
|  | |
|  | |
| 16 | 
             
            eval_steps: 20
         | 
| 17 | 
             
            flash_attention: true
         | 
| 18 | 
            -
             | 
| 19 | 
             
            group_by_length: true
         | 
| 20 | 
             
            hub_model_id: SystemAdmin123/tiny-random-mistral
         | 
| 21 | 
             
            hub_strategy: checkpoint
         | 
| 22 | 
             
            learning_rate: 0.0002
         | 
| 23 | 
             
            logging_steps: 10
         | 
| 24 | 
             
            lr_scheduler: cosine
         | 
| 25 | 
            -
            max_steps:  | 
| 26 | 
            -
            micro_batch_size:  | 
| 27 | 
             
            model_type: AutoModelForCausalLM
         | 
| 28 | 
             
            num_epochs: 100
         | 
| 29 | 
             
            optimizer: adamw_bnb_8bit
         | 
| 30 | 
            -
            output_dir: /root/.sn56/axolotl/ | 
| 31 | 
             
            pad_to_sequence_len: true
         | 
| 32 | 
             
            resize_token_embeddings_to_32x: false
         | 
| 33 | 
            -
             | 
|  | |
| 34 | 
             
            save_total_limit: 1
         | 
| 35 | 
             
            sequence_len: 2048
         | 
| 36 | 
             
            special_tokens:
         | 
| 37 | 
             
              pad_token: </s>
         | 
| 38 | 
             
            tokenizer_type: LlamaTokenizerFast
         | 
|  | |
|  | |
|  | |
| 39 | 
             
            trust_remote_code: true
         | 
| 40 | 
             
            val_set_size: 0.1
         | 
| 41 | 
             
            wandb_entity: ''
         | 
|  | |
| 1 | 
             
            base_model: echarlaix/tiny-random-mistral
         | 
| 2 | 
            +
            batch_size: 128
         | 
| 3 | 
             
            bf16: true
         | 
| 4 | 
             
            chat_template: tokenizer_default_fallback_alpaca
         | 
| 5 | 
             
            datasets:
         | 
|  | |
| 13 | 
             
                no_input_format: '{instruction}'
         | 
| 14 | 
             
                system_format: '{system}'
         | 
| 15 | 
             
                system_prompt: ''
         | 
| 16 | 
            +
            device_map: auto
         | 
| 17 | 
            +
            eval_sample_packing: false
         | 
| 18 | 
             
            eval_steps: 20
         | 
| 19 | 
             
            flash_attention: true
         | 
| 20 | 
            +
            gradient_checkpointing: true
         | 
| 21 | 
             
            group_by_length: true
         | 
| 22 | 
             
            hub_model_id: SystemAdmin123/tiny-random-mistral
         | 
| 23 | 
             
            hub_strategy: checkpoint
         | 
| 24 | 
             
            learning_rate: 0.0002
         | 
| 25 | 
             
            logging_steps: 10
         | 
| 26 | 
             
            lr_scheduler: cosine
         | 
| 27 | 
            +
            max_steps: 10000
         | 
| 28 | 
            +
            micro_batch_size: 32
         | 
| 29 | 
             
            model_type: AutoModelForCausalLM
         | 
| 30 | 
             
            num_epochs: 100
         | 
| 31 | 
             
            optimizer: adamw_bnb_8bit
         | 
| 32 | 
            +
            output_dir: /root/.sn56/axolotl/tmp/tiny-random-mistral
         | 
| 33 | 
             
            pad_to_sequence_len: true
         | 
| 34 | 
             
            resize_token_embeddings_to_32x: false
         | 
| 35 | 
            +
            sample_packing: true
         | 
| 36 | 
            +
            save_steps: 20
         | 
| 37 | 
             
            save_total_limit: 1
         | 
| 38 | 
             
            sequence_len: 2048
         | 
| 39 | 
             
            special_tokens:
         | 
| 40 | 
             
              pad_token: </s>
         | 
| 41 | 
             
            tokenizer_type: LlamaTokenizerFast
         | 
| 42 | 
            +
            torch_dtype: bf16
         | 
| 43 | 
            +
            training_args_kwargs:
         | 
| 44 | 
            +
              hub_private_repo: true
         | 
| 45 | 
             
            trust_remote_code: true
         | 
| 46 | 
             
            val_set_size: 0.1
         | 
| 47 | 
             
            wandb_entity: ''
         | 
    	
        model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 4125024
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:3c8fe2ad088774d5e76624a457b153f251dc0ec5f06fcbd8c5ff1603b4952015
         | 
| 3 | 
             
            size 4125024
         | 
    	
        training_args.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:faefc5cfb35b477d61090232bacb558d94f772efa930a51976cb862dc32dd01e
         | 
| 3 | 
            +
            size 6840
         |