Click any tag below to further narrow down your results
Links
Researchers at HiddenLayer found a flaw in the guardrails of popular AI models like GPT-5.1 and Claude. The EchoGram attack uses specific words to trick these safety systems, allowing harmful requests to bypass defenses or causing harmless requests to be flagged as dangerous.
This article reviews key developments in large language models (LLMs) throughout 2025, highlighting trends such as reasoning, coding agents, and the rise of CLI tools. It details significant releases like Claude Code and the impact of agents on coding and search tasks. The author also discusses the implications of using LLMs in YOLO mode and the evolving landscape of AI applications.