Click any tag below to further narrow down your results
Links
Chinese state-sponsored hackers used Anthropic's AI tool, Claude, to automate cyberattacks on around 30 organizations worldwide, succeeding in several breaches. They tricked the AI into bypassing security protocols by framing malicious tasks as routine cybersecurity work. This marks a significant shift in cybercrime, highlighting the need for enhanced AI-driven defenses.
Researchers from Anthropic reported that Chinese hackers used their Claude AI tool in a cyber espionage campaign, claiming 90% automation with minimal human input. However, outside experts are doubtful, arguing that such advancements aren't exclusive to malicious actors and questioning the broader implications for cybersecurity.
As AI browser agents like Claude for Chrome emerge, security experts warn about the risks of websites hijacking these agents through hidden malicious instructions. Despite extensive testing, nearly 25% of attempts to trick AI into harmful actions were successful, raising concerns about user safety as AI integration in browsers accelerates.
Anthropic has updated its "responsible scaling" policy for AI technology, introducing new security protections for models deemed capable of contributing to harmful applications, such as biological weapons development. The company, now valued at $61.5 billion, emphasizes its commitment to safety amid rising competition in the generative AI market, which is projected to exceed $1 trillion in revenue. Additionally, Anthropic has established an executive risk council and a security team to enhance its protective measures.