--- library_name: transformers tags: [fine-tuning, medical, oncology, reasoning, chain-of-thought, Qwen-1.5B] --- # Model Card for Medical Oncology Reasoning Fine-Tuned Model This is a fine-tuned version of the **DeepSeek-R1-Distill-Qwen-1.5B** model, specifically adapted for medical oncology reasoning tasks using chain-of-thought prompting. The fine-tuning was performed on a curated subset of the [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT) dataset. This model is designed to provide detailed, step-by-step reasoning when answering medical questions, making it suitable for clinical decision support, medical education, and research. ## Model Details ### Model Description This model leverages the capabilities of **DeepSeek-R1-Distill-Qwen-1.5B** and has been fine-tuned to enhance its performance in medical reasoning. It incorporates chain-of-thought prompting to produce detailed explanations for clinical queries. The fine-tuning process focused on tasks related to clinical diagnostics, treatment planning, and medical reasoning. - **Developed by:** Arihant Tripathi - **Model type:** Causal Language Model (LLM) - **Language(s):** English - **Finetuned from model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) ### Model Sources - **Repository:** [EndOfLe/OncoFineTuned](https://huggingface.co/EndOfLe/OncoFineTuned) ## Uses ### Direct Use This model can be used directly to generate detailed, step-by-step medical reasoning in response to clinical queries. It is especially useful for: - Medical diagnosis support. - Clinical reasoning and treatment planning. - Medical education and research. ### Downstream Use The model can be integrated into larger clinical decision support systems or educational tools that require natural language understanding and detailed reasoning for medical queries. ### Out-of-Scope Use The model is not intended for: - Replacing expert medical advice or making final clinical decisions. - General-purpose language generation without domain adaptation. - High-stakes applications where errors could have severe consequences without expert oversight. ## Bias, Risks, and Limitations The model’s output should be treated as an aid to human decision-making rather than a substitute for professional medical advice. Key considerations include: - The model may generate outdated or incorrect medical information. - The reasoning is based on the training data and might reflect its inherent biases. - Use in clinical settings should always involve human review and validation. ### Recommendations Users should verify the model’s outputs with expert medical knowledge and ensure its use complies with clinical standards and ethical guidelines. ## How to Get Started with the Model You can get started with the model using the following code snippet: ```python from transformers import AutoTokenizer, AutoModelForCausalLM model_name = "EndOfLe/OncoFineTuned" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) input_text = "What are the causes of colon cancer?" inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=256) print(tokenizer.decode(outputs[0], skip_special_tokens=True))