An interactive tool designed to visualize the relationships and flow of code reviews within a development team or project. It helps developers and managers understand how changes move through the review process, identifying bottlenecks and key contributors in the codebase evolution.
- Visual mapping of pull requests and code reviews
- Analysis of reviewer engagement and response times
- Identification of workflow patterns and potential delays
Greg Kroah-Hartman, a long-term Linux kernel maintainer, has observed a significant shift in AI-driven activity around Linux security and code review. Previously receiving "AI slop" โ inaccurate or low-quality reports โ the past month has seen a marked improvement in the quality and relevance of AI-generated bug reports and security findings across open-source projects. While the cause of this change remains unknown, Kroah-Hartman notes the kernel team can handle the increased volume, but smaller projects may struggle. AI is increasingly used as a reviewer and assistant, and is even beginning to contribute patches, with tools like Sashiko being integrated to manage the influx.
This blog post details a personal code review tool built around `llm` and `git diff`. It describes installation, how it works, how the author uses it, and its advantages over GitHub's Copilot review tool.
The article discusses using AI for code review, emphasizing that it should be used as a tool to flag potential issues for human review, similar to how a spell checker works. It highlights a tool created by Bill Mill to aid in this process, which uses a command-line interface to connect to LLMs. The author stresses the importance of discernment when accepting AI suggestions and provides the system prompt used in the tool.