latchkeyChild's picture
Upload README.md with huggingface_hub
2d69308 verified
---
library_name: transformers
tags:
- trading
- finance
- deepseek
- fine-tuning
---
# DeepSeek Trading Assistant
This is a fine-tuned version of `DeepSeek-R1-Distill-Qwen-32B` specialized for generating trading strategies and market analysis.
## Model Details
### Model Description
- **Developed by:** latchkeyChild
- **Model type:** Decoder-only language model
- **Language(s):** English
- **License:** MIT
- **Finetuned from model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B)
## Uses
### Direct Use
This model is designed to:
1. Analyze market conditions using technical indicators
2. Generate trading strategies based on market analysis
3. Implement risk management rules
4. Create Python code for strategy implementation
### Training Data
The model is trained on a custom dataset containing:
- Market analysis using technical indicators (RSI, MACD, Moving Averages)
- Trading strategy implementations
- Risk management rules
- Python code examples using QuantConnect framework
### Training Procedure
#### Training Hyperparameters
- **Number of epochs:** 3
- **Batch size:** 2
- **Learning rate:** 1e-5
- **Gradient accumulation steps:** 8
- **Warmup steps:** 100
- **Training regime:** fp16 mixed precision with gradient checkpointing
- **Temperature:** 0.6 (recommended for DeepSeek-R1 series)
## Technical Specifications
### Compute Infrastructure
- **Required Hardware:** 2x NVIDIA A10G GPUs or 1x A100 GPU
- **Training Time (estimated):** 2-4 hours
## Model Card Contact
For questions or issues, please open an issue in the repository.