--- library_name: transformers tags: - medical license: apache-2.0 language: - fr - en base_model: - mistralai/Mistral-7B-v0.1 --- ### Model Description MedMistral-CPT-7B is a French medical language model based on Mistral-7B-v0.1, adapted for medical domain applications through Continual Pre-Training (CPT) on French medical texts. ### Model Details - **Model Type**: Causal Language Model - **Base Model**: Mistral-7B-v0.1 - **Language**: French - **Domain**: Medical/Healthcare - **Parameters**: 7 billion - **License**: Apache 2.0 ### Training Details **Continual Pre-Training (CPT)** - **Dataset**: NACHOS corpus (7.4 GB French medical texts) - **Training Duration**: 2.8 epochs - **Hardware**: 32 NVIDIA H100 80GB GPUs - **Training Time**: ~40 hours ### Computational Impact - **Carbon Emissions**: 9.86 kgCO2e - **Training Time**: 12 hours ### Ethical Considerations - **Medical Accuracy**: For research and educational purposes only - **Professional Oversight**: Requires verification by qualified medical professionals - **Bias Awareness**: May contain biases from training data - **Privacy**: Do not input private health information ### Citation ```bibtex ``` ### Contact For questions about these models, please contact: ikram.belmadani@lis-lab.fr