Spaces:
				
			
			
	
			
			
		Build error
		
	
	
	
			
			
	
	
	
	
		
		
		Build error
		
	llama3 p2 epoch 1 results
Browse files
    	
        llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml
    CHANGED
    
    | @@ -25,7 +25,7 @@ output_dir: saves/llama3-8b/lora/sft_bf16_p1_full | |
| 25 | 
             
            logging_steps: 10
         | 
| 26 | 
             
            save_steps: 175
         | 
| 27 | 
             
            plot_loss: true
         | 
| 28 | 
            -
             | 
| 29 |  | 
| 30 | 
             
            ### train
         | 
| 31 | 
             
            per_device_train_batch_size: 16
         | 
|  | |
| 25 | 
             
            logging_steps: 10
         | 
| 26 | 
             
            save_steps: 175
         | 
| 27 | 
             
            plot_loss: true
         | 
| 28 | 
            +
            overwrite_output_dir: true
         | 
| 29 |  | 
| 30 | 
             
            ### train
         | 
| 31 | 
             
            per_device_train_batch_size: 16
         | 
    	
        results/llama3-8b_lora_sft_bf16-p2.csv
    CHANGED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        scripts/eval-mgtv-llama3_8b.sh
    CHANGED
    
    | @@ -17,7 +17,6 @@ grep MemTotal /proc/meminfo | |
| 17 | 
             
            #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
         | 
| 18 |  | 
| 19 | 
             
            export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
         | 
| 20 | 
            -
            export START_EPOCH=1
         | 
| 21 | 
             
            export RESIZE_TOKEN_EMBEDDINGS=true
         | 
| 22 |  | 
| 23 | 
             
            #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
         | 
| @@ -27,11 +26,13 @@ export MODEL_PREFIX=llama3-8b_lora_sft_bf16 | |
| 27 | 
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
         | 
| 28 | 
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
         | 
| 29 | 
             
            export USING_P1_PROMPT_TEMPLATE=true
         | 
|  | |
| 30 | 
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         | 
| 31 | 
            -
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         | 
| 32 |  | 
| 33 | 
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
         | 
| 34 | 
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
         | 
| 35 | 
             
            export USING_P1_PROMPT_TEMPLATE=false
         | 
|  | |
| 36 | 
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         | 
| 37 | 
             
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         | 
|  | |
| 17 | 
             
            #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
         | 
| 18 |  | 
| 19 | 
             
            export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
         | 
|  | |
| 20 | 
             
            export RESIZE_TOKEN_EMBEDDINGS=true
         | 
| 21 |  | 
| 22 | 
             
            #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
         | 
|  | |
| 26 | 
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
         | 
| 27 | 
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
         | 
| 28 | 
             
            export USING_P1_PROMPT_TEMPLATE=true
         | 
| 29 | 
            +
            export START_EPOCH=3
         | 
| 30 | 
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         | 
| 31 | 
            +
            #python llm_toolkit/eval_logical_reasoning_all_epochs.py
         | 
| 32 |  | 
| 33 | 
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
         | 
| 34 | 
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
         | 
| 35 | 
             
            export USING_P1_PROMPT_TEMPLATE=false
         | 
| 36 | 
            +
            export START_EPOCH=1
         | 
| 37 | 
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         | 
| 38 | 
             
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         |