--- library_name: transformers tags: - trading - finance - deepseek - fine-tuning --- # DeepSeek Trading Assistant This is a fine-tuned version of `DeepSeek-R1-Distill-Qwen-32B` specialized for generating trading strategies and market analysis. ## Model Details ### Model Description - **Developed by:** latchkeyChild - **Model type:** Decoder-only language model - **Language(s):** English - **License:** MIT - **Finetuned from model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B) ## Uses ### Direct Use This model is designed to: 1. Analyze market conditions using technical indicators 2. Generate trading strategies based on market analysis 3. Implement risk management rules 4. Create Python code for strategy implementation ### Training Data The model is trained on a custom dataset containing: - Market analysis using technical indicators (RSI, MACD, Moving Averages) - Trading strategy implementations - Risk management rules - Python code examples using QuantConnect framework ### Training Procedure #### Training Hyperparameters - **Number of epochs:** 3 - **Batch size:** 2 - **Learning rate:** 1e-5 - **Gradient accumulation steps:** 8 - **Warmup steps:** 100 - **Training regime:** fp16 mixed precision with gradient checkpointing - **Temperature:** 0.6 (recommended for DeepSeek-R1 series) ## Technical Specifications ### Compute Infrastructure - **Required Hardware:** 2x NVIDIA A10G GPUs or 1x A100 GPU - **Training Time (estimated):** 2-4 hours ## Model Card Contact For questions or issues, please open an issue in the repository.