ik-ram28's picture
Update README.md
b96e2c1 verified
---
library_name: transformers
tags:
- medical
license: apache-2.0
language:
- fr
- en
base_model:
- mistralai/Mistral-7B-Instruct-v0.1
---
### Model Description
MedMistralInstruct-CPT-7B is adapted from Mistral-7B-Instruct-v0.1 through Continual Pre-Training, maintaining instruction-following capabilities while gaining medical domain knowledge.
### Model Details
- **Model Type**: Causal Language Model
- **Base Model**: Mistral-7B-Instruct-v0.1
- **Language**: French
- **Domain**: Medical/Healthcare
- **Parameters**: 7 billion
- **License**: Apache 2.0
### Training Details
**Continual Pre-Training (CPT)**
- **Dataset**: NACHOS corpus (7.4 GB French medical texts)
- **Training Duration**: 2.8 epochs
- **Hardware**: 32 NVIDIA A100 80GB GPUs
- **Training Time**: ~40 hours
### Computational Requirements
- **Carbon Emissions**: 32.89 kgCO2e
- **Training Time**: 40 hours
### Ethical Considerations
- **Medical Accuracy**: For research and educational purposes only
- **Professional Oversight**: Requires verification by qualified medical professionals
- **Bias Awareness**: May contain biases from training data
- **Privacy**: Do not input private health information
### Citation
```bibtex
```
### Contact
For questions about these models, please contact: [email protected]