Investigating Autonomous Agent Contributions in the Wild: Activity Patterns and Code Change over Time
Abstract
Analysis of AI coding agents' contributions to open-source projects reveals increased activity but higher code churn compared to human-authored code.
The rise of large language models for code has reshaped software development. Autonomous coding agents, able to create branches, open pull requests, and perform code reviews, now actively contribute to real-world projects. Their growing role offers a unique and timely opportunity to investigate AI-driven contributions and their effects on code quality, team dynamics, and software maintainability. In this work, we construct a novel dataset of approximately 110,000 open-source pull requests, including associated commits, comments, reviews, issues, and file changes, collectively representing millions of lines of source code. We compare five popular coding agents, including OpenAI Codex, Claude Code, GitHub Copilot, Google Jules, and Devin, examining how their usage differs in various development aspects such as merge frequency, edited file types, and developer interaction signals, including comments and reviews. Furthermore, we emphasize that code authoring and review are only a small part of the larger software engineering process, as the resulting code must also be maintained and updated over time. Hence, we offer several longitudinal estimates of survival and churn rates for agent-generated versus human-authored code. Ultimately, our findings indicate an increasing agent activity in open-source projects, although their contributions are associated with more churn over time compared to human-authored code.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- When AI Teammates Meet Code Review: Collaboration Signals Shaping the Integration of Agent-Authored Pull Requests (2026)
- Safer Builders, Risky Maintainers: A Comparative Study of Breaking Changes in Human vs Agentic PRs (2026)
- AIDev: Studying AI Coding Agents on GitHub (2026)
- Human-AI Synergy in Agentic Code Review (2026)
- Debt Behind the AI Boom: A Large-Scale Empirical Study of AI-Generated Code in the Wild (2026)
- Learning to Commit: Generating Organic Pull Requests via Online Repository Memory (2026)
- When is Generated Code Difficult to Comprehend? Assessing AI Agent Python Code Proficiency in the Wild (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.00917 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper