--- license: apache-2.0 datasets: - open-thoughts/OpenThoughts-114k - prithivMLmods/Deepthink-Reasoning-Ins base_model: - XeTute/SaplingDream_V1-0.5B tags: - reasoning - conversational - thinking - tiny - small ---
Introducing SaplingDream, a compact GPT model with 0.5 billion parameters, based on the Qwen/Qwen2.5-0.5B-Instruct architecture. This model has been fine-tuned on a RTX4060 8GB for a bit over two days on ~0.3B tokens...
Evaluation Loss Chart