from transformers import AutoModelForCausalLM, PeftModel from huggingface_hub import HfApi # Load base model and adapter model base_model = AutoModelForCausalLM.from_pretrained("microsoft/phi-4") adapter_model = PeftModel.from_pretrained(base_model, "ivxxdegen/mibera-v1") # Merge the LoRA adapter into the base model merged_model = adapter_model.merge_and_unload() merged_model.save_pretrained("merged_model") # Upload the merged model to Hugging Face api = HfApi() api.upload_folder(folder_path="merged_model", repo_id="ivxxdegen/mibera-v1") print("✅ Merge and upload completed successfully.")