OncoFineTuned / README.md
Arihant Tripathi
Update README.md
677ef03 verified
---
library_name: transformers
tags: [fine-tuning, medical, oncology, reasoning, chain-of-thought, Qwen-1.5B]
---
# Model Card for Medical Oncology Reasoning Fine-Tuned Model
This is a fine-tuned version of the **DeepSeek-R1-Distill-Qwen-1.5B** model, specifically adapted for medical oncology reasoning tasks using chain-of-thought prompting. The fine-tuning was performed on a curated subset of the [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT) dataset. This model is designed to provide detailed, step-by-step reasoning when answering medical questions, making it suitable for clinical decision support, medical education, and research.
## Model Details
### Model Description
This model leverages the capabilities of **DeepSeek-R1-Distill-Qwen-1.5B** and has been fine-tuned to enhance its performance in medical reasoning. It incorporates chain-of-thought prompting to produce detailed explanations for clinical queries. The fine-tuning process focused on tasks related to clinical diagnostics, treatment planning, and medical reasoning.
- **Developed by:** Arihant Tripathi
- **Model type:** Causal Language Model (LLM)
- **Language(s):** English
- **Finetuned from model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B)
### Model Sources
- **Repository:** [EndOfLe/OncoFineTuned](https://huggingface.co/EndOfLe/OncoFineTuned)
## Uses
### Direct Use
This model can be used directly to generate detailed, step-by-step medical reasoning in response to clinical queries. It is especially useful for:
- Medical diagnosis support.
- Clinical reasoning and treatment planning.
- Medical education and research.
### Downstream Use
The model can be integrated into larger clinical decision support systems or educational tools that require natural language understanding and detailed reasoning for medical queries.
### Out-of-Scope Use
The model is not intended for:
- Replacing expert medical advice or making final clinical decisions.
- General-purpose language generation without domain adaptation.
- High-stakes applications where errors could have severe consequences without expert oversight.
## Bias, Risks, and Limitations
The model’s output should be treated as an aid to human decision-making rather than a substitute for professional medical advice. Key considerations include:
- The model may generate outdated or incorrect medical information.
- The reasoning is based on the training data and might reflect its inherent biases.
- Use in clinical settings should always involve human review and validation.
### Recommendations
Users should verify the model’s outputs with expert medical knowledge and ensure its use complies with clinical standards and ethical guidelines.
## How to Get Started with the Model
You can get started with the model using the following code snippet:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "EndOfLe/OncoFineTuned"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
input_text = "What are the causes of colon cancer?"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))