Model Description
MedMistralInstruct-CPT-7B is adapted from Mistral-7B-Instruct-v0.1 through Continual Pre-Training, maintaining instruction-following capabilities while gaining medical domain knowledge.
Model Details
- Model Type: Causal Language Model
- Base Model: Mistral-7B-Instruct-v0.1
- Language: French
- Domain: Medical/Healthcare
- Parameters: 7 billion
- License: Apache 2.0
Training Details
Continual Pre-Training (CPT)
- Dataset: NACHOS corpus (7.4 GB French medical texts)
- Training Duration: 2.8 epochs
- Hardware: 32 NVIDIA A100 80GB GPUs
- Training Time: ~40 hours
Computational Requirements
- Carbon Emissions: 32.89 kgCO2e
- Training Time: 40 hours
Ethical Considerations
- Medical Accuracy: For research and educational purposes only
- Professional Oversight: Requires verification by qualified medical professionals
- Bias Awareness: May contain biases from training data
- Privacy: Do not input private health information
Citation
Contact
For questions about these models, please contact: [email protected]
- Downloads last month
- 20
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support