Click any tag below to further narrow down your results
Links
This article explains how to create a CLAUDE.md file to effectively onboard the Claude coding agent to your codebase. It emphasizes the importance of concise, relevant instructions and suggests organizing project-specific details separately to improve Claude's performance.
The author discusses a rapid transition from manual coding to using language models as coding agents. While this change improves productivity and creativity, it also raises concerns about the potential atrophy of manual coding skills and the quality of code generated by these models.
The author shares their experience of quickly replacing a broken SaaS service with LLM-generated code. They highlight the ease of building a simple solution tailored to their needs, while discussing the implications for SaaS products and software engineers.
The article discusses how the effectiveness of large language models (LLMs) in coding tasks often hinges on the harness used rather than the model itself. By experimenting with different editing tools, the author demonstrates significant improvements in performance, highlighting the importance of optimizing harnesses for better results.
SWE-Pruner is a tool designed for software development that reduces token costs and latency by selectively pruning irrelevant code. It uses a lightweight neural skimmer to retain critical lines based on task-specific goals, making it adaptable to various coding scenarios. The framework integrates with multiple LLMs and supports complex workflows.
The article argues that the cost of managing technical debt is decreasing due to advancements in large language models (LLMs). It suggests that developers can afford to take on more technical debt now, as future improvements in coding models will help address these shortcuts. The author challenges traditional coding practices, advocating for a shift in how software engineers approach coding quality.
This article outlines effective strategies for using AI coding assistants, emphasizing a structured approach to planning, context, and iterative development. The author shares insights from personal experience and community practices, highlighting the importance of detailed specifications and choosing the right models.
This article discusses Recursive Language Models (RLMs) as a solution to the problem of context rot in large language models. RLMs utilize a REPL environment to manage long contexts efficiently, enabling models to maintain performance even with extensive input data. The author highlights their potential for agent design and optimization while acknowledging current limitations.
The article discusses how LLM coding tools have transformed software development, making it faster and more accessible. It reflects on the shift from high-effort coding to rapid prototyping, raising concerns about quality and the true value of code in this new landscape.
This article discusses the author's shift from manual coding to using language model agents for programming. They highlight improvements in workflow and productivity, while also noting the limitations and potential pitfalls of relying on these models. The author expresses concerns about skill atrophy and predicts significant changes in software engineering by 2026.
The author shares insights from creating a unified coding agent harness, pi-ai, after years of frustration with existing tools. He emphasizes the importance of context management and offers technical details on API integration and model interoperability. The article also discusses challenges faced with self-hosting and API peculiarities.
This article explains how to fine-tune a language model using your LinkedIn posts. It details the steps to gather, format, and train the model, allowing it to generate content in your voice. The author shares their experience and offers tips for customization.
This article details a project where the author trains a smaller LLM to understand and generate diagrams in the Pintora language. The process includes dataset creation, two training phases, and evaluation of the model's accuracy in producing valid diagram syntax.
This repo lets you query multiple large language models (LLMs) and see their individual responses side by side. It then has them review and rank each other's outputs, with a designated Chairman LLM providing the final answer. The project is a simple, local web app meant for exploration and comparison of LLMs.
This article explores how advancements in software design, particularly through LLMs, shift the focus from using standard libraries to generating custom code. It highlights the implications for dependency management and emphasizes the need to understand the problem being solved rather than just the mechanics of coding. The author compares this shift to the evolution of 3D printing in manufacturing.
In a podcast discussion, predictions for the tech industry in 2026 are shared, highlighting the undeniable improvement of LLMs in writing code, advancements in coding agent security, and the potential obsolescence of manual coding. Other predictions include a successful breeding season for Kākāpō parrots and the implications of AI-assisted programming on software engineering careers.
LLM coding agents struggle with code manipulation, lacking the ability to effectively copy-paste, which creates an awkward coding experience. Additionally, their problem-solving methods are flawed due to a tendency to make assumptions rather than ask clarifying questions, limiting their effectiveness compared to human developers. These limitations highlight that LLMs are more akin to inexperienced interns than replacements for skilled programmers.
Armin Ronacher critiques the Model Context Protocol (MCP), arguing that it is not as efficient or composable as traditional coding methods. He emphasizes the importance of using code for automation tasks due to its reliability and the ability to validate results, highlighting a personal experience where he successfully transformed a blog using a code-driven approach rather than relying on MCP.
The article discusses the potential security risks associated with using large language models (LLMs) in coding practices. It highlights how these models can inadvertently introduce vulnerabilities and the implications for developers and organizations. The need for robust security measures when integrating LLMs into coding workflows is emphasized.
Kimi-Dev-72B is an advanced open-source coding language model designed for software engineering tasks, achieving a state-of-the-art performance of 60.4% on the SWE-bench Verified benchmark. It leverages large-scale reinforcement learning to autonomously patch real repositories and ensures high-quality solutions by only rewarding successful test suite completions. Developers and researchers are encouraged to explore and contribute to its capabilities, available for download on Hugging Face and GitHub.
Crush is a versatile tool that integrates various LLMs into terminal workflows, allowing users to choose from multiple models, switch between them mid-session, and maintain project-specific contexts. It offers extensive support across different operating systems and can be easily installed through various package managers. Additionally, Crush provides customization options for configurations and permissions, enhancing the user experience with AI-driven coding assistance.
The author reflects on how their reliance on large language models (LLMs) for tasks like coding, math, and writing has diminished their learning and understanding of foundational skills. They express concerns about the balance between increased output and the depth of knowledge, questioning whether using LLMs as shortcuts may ultimately hinder their long-term capabilities. The article also discusses historical parallels and the potential future of education with AI integration.
The article discusses the mixed effectiveness of large language model (LLM)-based coding tools, acknowledging both their limitations and advantages in modern software development. While these tools can speed up prototyping and reduce repetitive coding tasks, they may produce errors or overly verbose code, necessitating strong code review skills from developers. Ultimately, the article emphasizes the importance of understanding how to effectively leverage these tools while maintaining critical thinking in coding practices.