Click any tag below to further narrow down your results
Links
The article outlines various design issues in LLVM, including insufficient code review capacity, frequent API changes, and challenges with build times and testing. It emphasizes the need for better testing practices and more stable APIs to enhance user experience and contributor engagement.
The article discusses how AI changes the landscape of code reviews, making the reviewer's job more complex. It outlines specific heuristics for assessing pull requests (PRs), focusing on aspects like design, testing, error handling, and the effort put in by the author. The author emphasizes the need for human oversight despite advances in AI review tools.
The author argues against traditional line-by-line code review, advocating for a harness-first approach where specifications and testing take priority. They draw on examples from AI-assisted coding and highlight the importance of architecture and feedback loops over direct code inspection. Caveats are noted for critical systems where code review remains essential.
This article explains how AI is changing the code review process, emphasizing the need for evidence of code functionality rather than just relying on AI-generated outputs. It contrasts solo developers’ fast-paced workflows with team dynamics, where human judgment remains essential for quality and security. The piece outlines best practices for integrating AI into development and review processes.
This article discusses how AI is changing the code review process for both solo developers and teams. It emphasizes the need for evidence of working code, highlights the risks of relying too heavily on AI, and outlines best practices for integrating AI into code reviews while maintaining human oversight.
This article outlines how Qodo developed a benchmark to evaluate AI code review systems. It highlights a new methodology that injects defects into real pull requests to assess both bug detection and code quality, demonstrating superior results compared to other platforms.
Codacy introduces a hybrid code review engine that enhances Pull Request feedback by identifying logic gaps, security issues, and code complexity. It automates the review process, letting developers ship code faster and with more confidence.
The article stresses the importance of software engineers providing code that they have manually and automatically tested before submission. It emphasizes accountability in code reviews and the use of coding agents to assist in proving code functionality. Developers should include evidence of their tests to respect their colleagues' time and efforts.
This article emphasizes the responsibility of software engineers to deliver code that has been thoroughly tested and proven to work, both manually and automatically. It argues against the trend of relying on AI tools to submit untested code and stresses the importance of accountability in the development process.