DeepSeek-R1-Distill-Qwen-1.5B Fine-Tuned on Physics

This repository contains a fine-tuned version of the DeepSeek-R1-Distill-Qwen-1.5B base model, adapted specifically for answering physics-related questions with detailed, step-by-step chain-of-thought reasoning. The model has been fine-tuned using Parameter-Efficient Fine-Tuning (PEFT) with LoRA and 4-bit quantization to reduce memory usage while maintaining performance in the physics domain.

Model Details

Model Description

The model is specialized for physics tasks through fine-tuning on three curated datasets:

  • camel_physics: Educational examples with structured prompts and chain-of-thought reasoning.
  • arxiv_physics: Research-level questions and scholarly excerpts from physics papers.
  • alpaca_physics: Instruction-based conversational examples in physics.

Fine-tuning was performed using PEFT techniques (LoRA) combined with 4-bit quantization. This configuration enables the model to generate comprehensive and contextually accurate explanations for complex physics problems.

  • Developed by: Your Organization or Name
  • Funded by: [Funding Source, if applicable]
  • Shared by: Your Organization or Name
  • Model type: Transformer-based causal language model, fine-tuned with PEFT (LoRA)
  • Language(s): English
  • License: [Specify License, e.g., Apache-2.0 or MIT]
  • Finetuned from model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B

Model Sources

  • Repository: [Link to the model repository on Hugging Face]
  • Paper: [Link to any associated paper or blog post]
  • Demo: [Link to a demo, if available]

Uses

Direct Use

This model can be used to:

  • Answer physics-related questions.
  • Generate detailed explanations and step-by-step chain-of-thought reasoning for physics problems.
  • Serve as an educational tool for physics and mathematics learners.

Downstream Use

The model can be integrated into:

  • Educational platforms and tutoring applications.
  • Research assistance tools in physics.
  • Chatbots and virtual assistants with a scientific focus.

Out-of-Scope Use

The model is not intended for:

  • Domains outside of physics, where domain-specific knowledge is critical.
  • High-stakes applications without human verification.
  • Use cases requiring generation in languages other than English.

Bias, Risks, and Limitations

  • Bias: The model is fine-tuned on curated physics datasets and may reflect biases inherent in that data.
  • Risks: Inaccurate or oversimplified explanations may be generated, especially for advanced or niche physics topics. Users should verify outputs.
  • Limitations: The model's knowledge is limited to the physics topics covered in the training data and may not generalize to emerging or unconventional topics.

Recommendations

Users should:

  • Verify the generated content for accuracy, particularly in educational or research contexts.
  • Use the model as a supportive tool rather than a standalone source.
  • Be aware of its domain-specific training and adjust expectations accordingly.

How to Get Started with the Model

Install the required libraries:

pip install transformers peft
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for jedlee2004/physics-chat

Adapter
(86)
this model