Click any tag below to further narrow down your results
Links
Philippe discusses using small language models (LLMs) for coding tasks, particularly with a Golang project called Nova. He outlines techniques for improving model performance through tailored prompts and a method called Retrieval Augmented Generation (RAG).
The article shares predictions about the future of large language models (LLMs) and coding agents, highlighting expected advancements in coding quality, security, and the evolution of software engineering. The author expresses a mix of optimism and caution, emphasizing the importance of sandboxing and the potential impact of AI-assisted coding on the industry.
The article discusses how a practical approach to software development involves understanding existing code rather than treating it as a black box. It argues that foundational knowledge remains essential, especially as tools like LLMs evolve, and emphasizes the importance of continuous learning and building core systems.
Wes McKinney explores the arithmetic shortcomings of large language models (LLMs) like Anthropic's Claude Code. He shares his experiences using these coding agents, highlighting how they can improve productivity but often struggle with basic calculations and reliability. Testing various models, he finds that local models perform better than many API options in handling arithmetic tasks.
The article discusses the author's experiences with LLMs and coding agents over the past year. It highlights significant improvements in coding models, the issues with current IDEs, and the author's new approach to programming using agents instead of traditional environments.
After struggling with data entry in his game development project, the author discovered that reconstructing game assets as code rather than using the Unity editor significantly improved his workflow. By leveraging LLMs to assist in generating C# code from structured data, he was able to streamline the process and avoid burnout, ultimately allowing him to focus on problem analysis and solution development.
A developer shares insights from creating a VS Code extension called terminal-editor, which integrates a shell-like interface within the editor. The article emphasizes the importance of structured planning and testing strategies when working with large language models (LLMs) to enhance coding efficiency and reduce errors. It highlights the need for an effective feedback loop and the limitations of LLMs in maintaining code quality and handling complex problems.
The author shares insights from a month of experimenting with AI tools for software development, highlighting the limitations of large language models (LLMs) in producing production-ready code and their dependency on well-structured codebases. They discuss the challenges of integrating LLMs into workflows, the instability of AI products, and their mixed results across programming languages, emphasizing that while LLMs can aid in standard tasks, they struggle with unique or complex requirements.