|
|
--- |
|
|
base_model: |
|
|
- Qwen/Qwen2.5-Coder-32B-Instruct |
|
|
datasets: |
|
|
- SWE-bench/SWE-smith |
|
|
language: |
|
|
- en |
|
|
library_name: transformers |
|
|
license: apache-2.0 |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- agent |
|
|
- software engineering |
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
<img src="https://swe-agent.com/latest/assets/swe-agent.svg" alt="Logo" width="200"> |
|
|
<h1 align="center">SWE-agent LM</h1> |
|
|
</div> |
|
|
|
|
|
<p align="center"> |
|
|
<a href="https://github.com/SWE-bench/SWE-smith">Code</a> |
|
|
• |
|
|
<a href="https://arxiv.org/abs/2504.21798">Paper</a> |
|
|
• |
|
|
<a href="https://swesmith.com/">Site</a> |
|
|
</p> |
|
|
|
|
|
SWE-agent-LM-32B is a Language Model for Software Engineering trained using the [SWE-smith](https://github.com/SWE-bench/SWE-smith) toolkit. |
|
|
We introduce this model as part of our work: [SWE-smith: Scaling Data for Software Engineering Agents](https://swesmith.com). |
|
|
|
|
|
SWE-agent-LM-32B is 100% open source. |
|
|
Training this model was simple - we fine-tuned Qwen 2.5 Coder Instruct on 5k trajectories generated by SWE-agent + Claude 3.7 Sonnet. |
|
|
The dataset can be found [here](https://huggingface.co/datasets/SWE-bench/SWE-smith-trajs-250429). |
|
|
|
|
|
SWE-agent-LM-32B is compatible with [SWE-agent](https://github.com/SWE-agent/SWE-agent). |
|
|
Running this model locally only takes a few steps! |
|
|
Check [here]() for more instructions on how to do so. |
|
|
|
|
|
If you found this work exciting and want to push SWE-agents further, please feel free to connect with us (the [SWE-bench team](https://swe-bench.github.io/)) more! |