SWE-agent-LM-32B / README.md
john-b-yang's picture
Add Transformers library (#4)
853998b verified
---
base_model:
- Qwen/Qwen2.5-Coder-32B-Instruct
datasets:
- SWE-bench/SWE-smith
language:
- en
library_name: transformers
license: apache-2.0
pipeline_tag: text-generation
tags:
- agent
- software engineering
---
<div align="center">
<img src="https://swe-agent.com/latest/assets/swe-agent.svg" alt="Logo" width="200">
<h1 align="center">SWE-agent LM</h1>
</div>
<p align="center">
<a href="https://github.com/SWE-bench/SWE-smith">Code</a>
<a href="https://arxiv.org/abs/2504.21798">Paper</a>
<a href="https://swesmith.com/">Site</a>
</p>
SWE-agent-LM-32B is a Language Model for Software Engineering trained using the [SWE-smith](https://github.com/SWE-bench/SWE-smith) toolkit.
We introduce this model as part of our work: [SWE-smith: Scaling Data for Software Engineering Agents](https://swesmith.com).
SWE-agent-LM-32B is 100% open source.
Training this model was simple - we fine-tuned Qwen 2.5 Coder Instruct on 5k trajectories generated by SWE-agent + Claude 3.7 Sonnet.
The dataset can be found [here](https://huggingface.co/datasets/SWE-bench/SWE-smith-trajs-250429).
SWE-agent-LM-32B is compatible with [SWE-agent](https://github.com/SWE-agent/SWE-agent).
Running this model locally only takes a few steps!
Check [here]() for more instructions on how to do so.
If you found this work exciting and want to push SWE-agents further, please feel free to connect with us (the [SWE-bench team](https://swe-bench.github.io/)) more!