xingyaoww commited on
Commit
280b774
·
verified ·
1 Parent(s): a2be33b

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +101 -0
README.md ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - SWE-Gym/SWE-Gym
5
+ language:
6
+ - en
7
+ base_model:
8
+ - Qwen/Qwen2.5-Coder-7B-Instruct
9
+ pipeline_tag: text-generation
10
+ tags:
11
+ - agent
12
+ - coding
13
+ ---
14
+
15
+ <div align="center">
16
+ <img src="https://github.com/All-Hands-AI/OpenHands/blob/main/docs/static/img/logo.png?raw=true" alt="Logo" width="200">
17
+ <h1 align="center">OpenHands LM v0.1</h1>
18
+ </div>
19
+
20
+ <p align="center">
21
+ <a href="https://www.all-hands.dev/blog/introducing-openhands-lm-32b----a-strong-open-coding-agent-model">Blog</a>
22
+
23
+ <a href="https://docs.all-hands.dev/modules/usage/llms/local-llms" >Use it in OpenHands</a>
24
+ </p>
25
+
26
+
27
+ **This is a smaller 7B model trained following the recipe of [all-hands/openhands-lm-32b-v0.1](https://huggingface.co/all-hands/openhands-lm-32b-v0.1).**
28
+
29
+ ---
30
+
31
+ Autonomous agents for software development are already contributing to a [wide range of software development tasks](/blog/8-use-cases-for-generalist-software-development-agents).
32
+ But up to this point, strong coding agents have relied on proprietary models, which means that even if you use an open-source agent like [OpenHands](https://github.com/All-Hands-AI/OpenHands), you are still reliant on API calls to an external service.
33
+
34
+ Today, we are excited to introduce OpenHands LM, a new open coding model that:
35
+
36
+ - Is open and [available on Hugging Face](https://huggingface.co/all-hands/openhands-lm-32b-v0.1), so you can download it and run it locally
37
+ - Is a reasonable size, 32B, so it can be run locally on hardware such as a single 3090 GPU
38
+ - Achieves strong performance on software engineering tasks, including 37.2% resolve rate on SWE-Bench Verified
39
+
40
+ Read below for more details and our future plans!
41
+
42
+ ## What is OpenHands LM?
43
+
44
+ OpenHands LM is built on the foundation of [Qwen Coder 2.5 Instruct 32B](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct), leveraging its powerful base capabilities for coding tasks. What sets OpenHands LM apart is our specialized fine-tuning process:
45
+
46
+ - We used training data generated by OpenHands itself on a diverse set of open-source repositories
47
+ - Specifically, we use an RL-based framework outlined in [SWE-Gym](https://arxiv.org/abs/2412.21139), where we set up a training environment, generate training data using an existing agent, and then fine-tune the model on examples that were resolved successfully
48
+ - It features a 128K token context window, ideal for handling large codebases and long-horizon software engineering tasks
49
+
50
+
51
+ ## Performance: Punching Above Its Weight
52
+
53
+ We evaluated OpenHands LM using our latest [iterative evaluation protocol](https://github.com/All-Hands-AI/OpenHands/tree/main/evaluation/benchmarks/swe_bench#run-inference-rollout-on-swe-bench-instances-generate-patch-from-problem-statement) on the [SWE-Bench Verified benchmark](https://www.swebench.com/#verified).
54
+
55
+ The results are impressive:
56
+
57
+ - **37.2% verified resolve rate** on SWE-Bench Verified
58
+ - Performance comparable to models with **20x more parameters**, including Deepseek V3 0324 (38.8%) with 671B parameters
59
+
60
+ Here's how OpenHands LM compares to other leading open-source models:
61
+
62
+ ![OpenHands LM Performance Comparison](https://www.all-hands.dev/assets/blog/20250331-openhands-lm-release/performance_scatter.png)
63
+
64
+ As the plot demonstrates, our 32B parameter model achieves efficiency that approaches much larger models. While the largest models (671B parameters) achieve slightly higher scores, our 32B parameter model performs remarkably well, opening up possibilities for local deployment that are not possible with larger models.
65
+
66
+ ## Getting Started: How to Use OpenHands LM Today
67
+
68
+ You can start using OpenHands LM immediately through these channels:
69
+
70
+ 1. **Download the model from Hugging Face**
71
+ The model is available on [Hugging Face](https://huggingface.co/all-hands/openhands-lm-32b-v0.1) and can be downloaded directly from there.
72
+
73
+ 2. **Create an OpenAI-compatible endpoint with a model serving framework**
74
+ For optimal performance, it is recommended to serve this model with a GPU using [SGLang](https://github.com/sgl-project/sglang) or [vLLM](https://github.com/vllm-project/vllm).
75
+
76
+ 3. **Point your OpenHands agent to the new model**
77
+ Download [OpenHands](https://github.com/All-Hands-AI/OpenHands) and follow the instructions for [using an OpenAI-compatible endpoint](https://docs.all-hands.dev/modules/usage/llms/openai-llms#using-openai-compatible-endpoints).
78
+
79
+
80
+ ## The Road Ahead: Our Development Plans
81
+
82
+ This initial release marks just the beginning of our journey. We will continue enhancing OpenHands LM based on community feedback and ongoing research initiatives.
83
+
84
+ In particular, it should be noted that the model is still a research preview, and (1) may be best suited for tasks regarding solving github issues and perform less well on more varied software engineering tasks, (2) may sometimes generate repetitive steps, and (3) is somewhat sensitive to quantization, and may not function at full performance at lower quantization levels.
85
+ Our next releases will focus on addressing these limitations.
86
+
87
+ We're also developing more compact versions of the model (including a 7B parameter variant) to support users with limited computational resources. These smaller models will preserve OpenHands LM's core strengths while dramatically reducing hardware requirements.
88
+
89
+ We encourage you to experiment with OpenHands LM, share your experiences, and participate in its evolution. Together, we can create better tools for tomorrow's software development landscape.
90
+
91
+ ## Join Our Community
92
+
93
+ We invite you to be part of the OpenHands LM journey:
94
+
95
+ - Explore our [GitHub repository](https://github.com/All-Hands-AI/OpenHands)
96
+ - Connect with us on [Slack](https://join.slack.com/t/openhands-ai/shared_invite/zt-2tom0er4l-JeNUGHt_AxpEfIBstbLPiw)
97
+ - Follow our [documentation](https://docs.all-hands.dev) to get started
98
+
99
+ By contributing your experiences and feedback, you'll help shape the future of this open-source initiative. Together, we can create better tools for tomorrow's software development landscape.
100
+
101
+ We can't wait to see what you'll create with OpenHands LM!