--- license: apache-2.0 library_name: transformers language: - en tags: - chat - conversational base_model: - maldv/Qwentile2.5-32B-Instruct - a-m-team/AM-Thinking-v1 - nvidia/OpenCodeReasoning-Nemotron-32B - maldv/Loqwqtus2.5-32B-Instruct - trashpanda-org/QwQ-32B-Snowdrop-v0 - ArliAI/QwQ-32B-ArliAI-RpR-v3 pipeline_tags: - text-generation - conversational - chat --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c1b098c85365af5a83e/nJ7JqRPXgIthnMl6YZ8zt.png) [GGUF](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-GGUF) [iMat](https://huggingface.co/mradermacher/QwentileLambda2.5-32B-Instruct-i1-GGUF) # Qwentile Λ 2.5 32B Instruct Qwentile Λ 2.5 32B Instruct is a *normalized denoised fourier interpolation* of the following models: ```yaml output_base_model: "maldv/Qwentile2.5-32B-Instruct" output_dtype: "bfloat16" finetune_merge: - { "model": "a-m-team/AM-Thinking-v1", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 } - { "model": "nvidia/OpenCodeReasoning-Nemotron-32B", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8, "is_input": true} - { "model": "maldv/Loqwqtus2.5-32B-Instruct", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 } - { "model": "trashpanda-org/QwQ-32B-Snowdrop-v0", "base": "Qwen/Qwen2.5-32B", "alpha": 0.9 } - { "model": "ArliAI/QwQ-32B-ArliAI-RpR-v3", "base": "Qwen/Qwen2.5-32B", "alpha": 0.8 } ``` In other words, all of these models get warped and interpolated in signal space, and then jammed back on top of the base model (which in this case was Qwentile2.5-32B-Instruct); but with the Nemotron OpenCodeReasoning input layer. ### What is this? The latest in my series of Qwen 2.5 merges. Some really good models have been released recently, so I folded them in with Qwentile as the base. It should exhibit superior thinking skills, and perhaps even some code ability. I was satisfied with QReasoner2.5-32B-Instruct for advanced reasoning, but I suspect this will be an improvement. ### A <think> model? No, oddly enough, given it's lineage I thought for sure it would be a thought model, but instead it blends thought with it's creative output almost seamlessly. The combination is pretty powerful in my initial tests. ## Citation If you find our work helpful, feel free to give us a cite. ``` @misc{qwentile-labmda-2.5-32b-instruct, title = {Qwentile Λ 2.5 32B Instruct}, url = {https://huggingface.co/maldv/QwentileLambda2.5-32B-Instruct}, author = {Praxis Maldevide}, month = {May}, year = {2025} } ```