--- license: apache-2.0 language: - en - zh base_model: - Qwen/Qwen3-30B-A3B-Thinking-2507 - Qwen/Qwen3-30B-A3B-Instruct-2507 - Qwen/Qwen3-Coder-30B-A3B-Instruct - Qwen/Qwen3-30B-A3B-Base pipeline_tag: text-generation tags: - merge --- > *This is the initial unified version of the Qwen3-30B-A3B series models.As more fine-tuned models emerge and merging methods are applied, we will further improve it. Stay tuned!* # *Model Highlights:* - ***merge method**: `nuslerp` `della`* - ***precision**: `dtype: bfloat16`* - ***Context length**: `1010000`* # *Parameter Settings:* > [!TIP] > *`Temperature=0.7`, `TopP=0.8`, `TopK=20`,`MinP=0`.* ## *Step1: Merge Code Model with Instruction & Thinking Models Separately* - *Adopt the nuslerp method to improve model absorption rate.* - *Set a merging ratio of 9:1 to prevent capability degradation caused by an excessively high proportion of the code model.* ```yaml models: - model: Qwen/Qwen3-30B-A3B-Instruct-2507 parameters: weight: 0.9 - model: Qwen/Qwen3-Coder-30B-A3B-Instruct parameters: weight: 0.1 merge_method: nuslerp tokenizer_source: Qwen/Qwen3-30B-A3B-Instruct-2507 parameters: normalize: true int8_mask: true dtype: bfloat16 name: Qwen3-30B-A3B-Coder-Instruct-nuslerp ``` ```yaml models: - model: Qwen/Qwen3-30B-A3B-Thinking-2507 parameters: weight: 0.9 - model: Qwen/Qwen3-Coder-30B-A3B-Instruct parameters: weight: 0.1 merge_method: nuslerp tokenizer_source: Qwen/Qwen3-30B-A3B-Thinking-2507 parameters: normalize: true int8_mask: true dtype: bfloat16 name: Qwen3-30B-A3B-Coder-Thinking-nuslerp ``` ## *Step2: Merge Code Instruction & Code Thinking Models into Base Model Together* - *Merge the two models into the base model using the della merging method to make the model more versatile and stable.* - *Since the merged model is more similar to the instruction model, we use the chat template of the Qwen3-30B-A3B-Instruct-2507.* ```yaml models: - model: Qwen3-30B-A3B-Coder-Instruct-nuslerp parameters: density: 1 weight: 1 lambda: 0.9 - model: Qwen3-30B-A3B-Coder-Thinking-nuslerp parameters: density: 1 weight: 1 lambda: 0.9 merge_method: della base_model: Qwen/Qwen3-30B-A3B-Base dtype: bfloat16 name: Qwen3-30B-A3B-YOYO-V2 ``` ## *Step3: Further Extend Context Length* - *By referring to the config_1m.json of Qwen3-30B-A3B-Instruct-2507, we modified the config.json of the merged model and extended the maximum context length to 1M.*