--- library_name: transformers tags: - medical license: apache-2.0 language: - fr - en base_model: - ik-ram28/MedMistralInstruct-CPT-7B - mistralai/Mistral-7B-Instruct-v0.1 --- ## MedMistralInstruct-CPT-SFT-7B ### Model Description MedMistralInstruct-CPT-SFT-7B is a French medical language model based on Mistral-7B-Instruct-v0.1, adapted through Continual Pre-Training followed by Supervised Fine-Tuning. ### Model Details - **Model Type**: Causal Language Model - **Base Model**: Mistral-7B-Instruct-v0.1 - **Language**: French - **Domain**: Medical/Healthcare - **Parameters**: 7 billion - **License**: Apache 2.0 ### Training Details **Continual Pre-Training (CPT)** - **Dataset**: NACHOS corpus (7.4 GB French medical texts) - **Training Duration**: 2.8 epochs - **Hardware**: 32 NVIDIA A100 80GB GPUs - **Training Time**: ~40 hours **Supervised Fine-Tuning (SFT)** - **Dataset**: 30K French medical question-answer pairs - **Method**: DoRA (Weight-Decomposed Low-Rank Adaptation) - **Training Duration**: 10 epochs - **Hardware**: 1 NVIDIA H100 80GB GPU - **Training Time**: ~42 hours ### Computational Requirements - **Carbon Emissions**: 33.96 kgCO2e (CPT+SFT) - **Training Time**: 82 hours total (CPT+SFT) ### Ethical Considerations - **Medical Accuracy**: For research and educational purposes only - **Professional Oversight**: Requires verification by qualified medical professionals - **Bias Awareness**: May contain biases from training data - **Privacy**: Do not input private health information ### Citation ```bibtex ``` ### Contact For questions about these models, please contact: ikram.belmadani@lis-lab.fr