Hackers Weaponize Claude Code in Mexican Government Cyberattack

Hackers have reportedly weaponized Anthropic's Claude AI model to assist in a cyberattack against the Mexican government. The AI was allegedly used to generate exploit code, develop malicious tools, and facilitate the exfiltration of over 150GB of sensitive data.

Pentagon Designates Anthropic Supply Chain Risk Over AI Military Dispute

The U.S. Department of Defense has designated AI company Anthropic as a "supply chain risk" due to disagreements over the lawful use of its AI model, Claude. The dispute centers on Anthropic's refusal to allow its AI for mass domestic surveillance of Americans and fully autonomous weapons.

‘Silent’ Google API key change exposed Gemini AI data

Researchers at Truffle Security discovered that Google Cloud API keys, traditionally used for billing, now also authenticate access to Gemini AI project data due to a silent change by Google. This allows anyone who scrapes the API keys from websites to access uploaded files, cached content, and consume tokens, potentially generating large bills for project owners.

Flaw-Finding AI Assistants Face Criticism for Speed, Accuracy

AI assistants designed to find software vulnerabilities are showing promise, but current versions are not meeting the expectations of enterprises and developers. Experts note that these tools struggle with speed and accuracy, limiting their effectiveness in real-world security assessments.

The Anthropic and US Government conflict is larger than you think

The article discusses the complex relationship between Anthropic, an AI company, and the US government, particularly concerning data privacy, surveillance, and national security interests. It highlights potential conflicts arising from government access to Anthropic's AI models and the implications for individual privacy and civil liberties.

Claude Code Security Shows Promise, Not Perfection

The article discusses the security implications of Claude Code, an AI tool. While it shows promise in code security, researchers caution that its impact may have been overstated and that it's not perfect.

Ukrainian man pleads guilty to running AI-powered fake ID site

A Ukrainian man has pleaded guilty to running OnlyFake, an AI-powered website that generated and sold fake identification documents. The website generated over 10,000 fake ID photos, highlighting the misuse of AI in enabling fraudulent activities.

Your personal OpenClaw agent may also be taking orders from malicious websites

Researchers at Oasis Security discovered a vulnerability chain, dubbed ClawJacked (CVE-2026-25253), in OpenClaw that allows malicious websites to gain full control of a locally running agent by exploiting the implicit trust of "localhost" connections. By bypassing rate limits and enabling unauthorized device pairing, attackers can access the agent's privileges, workflows, and credentials. OpenClaw promptly fixed the flaw after being notified.

How to make LLMs a defensive advantage without creating a new attack surface

The article discusses how large language models (LLMs) are impacting security teams, both as productivity tools and potential attack vectors. It suggests approaching LLMs as high-impact systems, defining outcomes, modeling threats, and building controls, and recommends starting with narrow, verifiable workflows before expanding their use.

Previously harmless Google API keys now expose Gemini AI data

Google API keys, initially intended for services like Maps and embedded in client-side code, can now be exploited to authenticate to the Gemini AI assistant. This access could potentially expose private user data stored within Gemini.

Rapid AI-driven development makes security unattainable, warns Veracode

Veracode's annual State of Software Security report indicates that more vulnerabilities are being created than fixed, exacerbated by rapid AI-driven development. This widening remediation gap makes achieving comprehensive security increasingly difficult, according to the report based on data from 1.6 million applications.

Hackers Compromise Networks Faster and Faster [DE]

Crowdstrike's Global Threat Report 2025 reveals that attackers are compromising networks much faster, with an average breakout time of 29 minutes, a 65% increase in speed compared to the previous year. The report attributes this acceleration to the increasing use of AI tools by cybercriminals and state-sponsored groups to automate information gathering, extract credentials, and conduct insider operations.

LLMs Generate Predictable Passwords

Large Language Models (LLMs) are generating predictable passwords with noticeable patterns. These include starting with a specific letter and number, uneven character choices, and an avoidance of repeating characters.

Expert Recommends: Prepare for PQC Right Now

An expert recommends preparing for Post-Quantum Cryptography (PQC) now due to the increasing threat of adversaries stealing encrypted data today to decrypt it in the future when quantum computers become more powerful. The rise of ransomware and cloud computing are also contributing factors to the need for PQC.

5 trends that should top CISO’s RSA 2026 agendas

The article discusses five cybersecurity trends expected to be prominent at RSA 2026, focusing on the rise of AI-SOCs and the importance of Continuous Threat Exposure Management (CTEM). It also touches on the increasing focus on the software supply chain, the evolving role of zero trust architecture, and the imperative of cybersecurity frameworks.

Srsly Risky Biz: Is Claude Too Woke For War?

This edition of Risky Business News discusses the potential risks and security implications of using AI models like Claude in sensitive environments, particularly in the context of war and conflict. It explores whether the model's design and biases could lead to unintended consequences or vulnerabilities.

Chinese Police Use ChatGPT to Smear Japan PM Takaichi

A Chinese citizen unintentionally exposed a Chinese police influence operation targeting Japan's Prime Minister Takaichi through a ChatGPT account. The operation involved using AI-generated content to spread disinformation and smear the target.

Claude Code Flaws Allow Remote Code Execution and API Key Exfiltration

Researchers have discovered security vulnerabilities in Anthropic's Claude Code, an AI coding assistant. These flaws could allow for remote code execution and exfiltration of API keys through exploiting configuration mechanisms like Hooks, Model Context Protocol servers, and environment variables.

The Blast Radius Problem: Stolen Credentials Are Weaponizing Agentic AI

According to IBM X-Force, over half of the vulnerabilities tracked in 2023 did not require authentication prior to exploitation, increasing the risk when combined with compromised credentials and agentic AI. This highlights the potential for significant damage when stolen credentials are used to weaponize agentic AI.

Poisoning AI Training Data

The article discusses how easily AI training data can be poisoned by creating a website with false information. The author created a fake article about tech journalists eating hot dogs and found that Google's Gemini and ChatGPT quickly incorporated this misinformation into their responses, while Claude was not fooled.

OpenAI says Chinese cops used ChatGPT to plan and track smear ops against opponents

OpenAI reports that individuals with ties to Chinese law enforcement attempted to utilize ChatGPT to orchestrate smear campaigns against the Japanese prime minister and other critics of the Chinese Communist Party. The activity was identified and disrupted by OpenAI as part of their ongoing efforts to monitor and mitigate malicious use of their AI models.

Hacker cracks 600 firewalls in one month - with AI [DE]

A Russian-speaking hacker compromised over 600 Fortigate firewalls between January and February 2026. The attacker initially exploited weak passwords and then used an AI tool based on Google Gemini to access more devices within the same network.

How AI is Changing Your GRC Strategy [DE]

The article discusses how companies need to revise their GRC (Governance, Risk & Compliance) processes to account for the increasing use and risks of generative and agentic AI. It highlights the challenges CISOs face in balancing innovation with securing AI deployments and the need to integrate AI risk management into GRC frameworks.

Risky Bulletin: Russia starts criminal probe of Telegram founder Pavel Durov

This Risky Business News bulletin covers multiple cybersecurity incidents. These include Russia initiating a criminal probe against Telegram founder Pavel Durov, teenagers hacking Ttareungyi, Anthropic accusing Chinese AI firms of distillation attacks, and the US Treasury sanctioning a Russian exploit broker.

Tech Companies Shouldn’t Be Bullied Into Doing Surveillance

The Department of Defense is pressuring AI company Anthropic to lift restrictions on how their technology is used, particularly regarding autonomous weapons systems and surveillance. Anthropic is resisting, citing their principles against using their AI for surveillance against US persons and autonomous weapons systems, leading to potential repercussions such as being labeled a "supply chain risk."

AI has gotten good at finding bugs, not so good at swatting them

Anthropic highlighted improvements in its Claude Code AI's ability to find software vulnerabilities and suggest patches. However, security researchers emphasize that vulnerability discovery alone is insufficient; validation and patching processes remain critical challenges.

Cyber defense: From reactive to proactive

The article discusses the importance of shifting from a reactive to a proactive approach to cybersecurity, highlighting the increasing speed and sophistication of AI-driven attacks. It emphasizes the need for efficient automation, an integrated platform, and a focus on layered security controls, zero trust, multi-cloud infrastructure, and security by design.

RoguePilot Flaw in GitHub Codespaces Enabled Copilot to Leak GITHUB_TOKEN

A vulnerability dubbed RoguePilot was found in GitHub Codespaces that allowed attackers to inject malicious Copilot instructions via GitHub issues, potentially leading to repository control. Orca Security discovered and reported the AI-driven flaw, which Microsoft has since patched.

Microsoft adds Copilot data controls to all storage locations

Microsoft is extending Data Loss Prevention (DLP) controls to Microsoft 365 Copilot, enabling administrators to block the AI assistant from accessing sensitive documents in various storage locations. This expansion aims to prevent Copilot from processing confidential information, improving data security and compliance.

Identity-First AI Security: Why CISOs Must Add Intent to the Equation

The article discusses the importance of securing AI agents by treating them as identities with specific intents. CISOs need to implement intent-based controls to ensure AI agents only have access when their purpose and context align, preventing over-scoped privileges.

As Cybersecurity Firms Chase AI, VC Market Skyrockets

Cybersecurity startup investments surged in 2025, driven by venture capital firms prioritizing AI-native technologies and skilled personnel. This trend indicates a growing emphasis on AI's role in cybersecurity and the need for specialized expertise.

Is AI Good for Democracy?

The article discusses the use of AI in various domains, particularly its potential impact on democracy. It highlights concerns about AI-generated content flooding academic journals and influencing public opinion, suggesting an ongoing arms race where AI is the weapon of choice.

The rise of the evasive adversary

CrowdStrike's 12th annual Global Threat Report highlights the rise of the 'evasive adversary,' focusing on avoiding detection rather than expanding toolkits. AI is being used to amplify existing tactics like phishing and reconnaissance, with AI-enabled intrusions becoming quieter and malware-free techniques accounting for a significant portion of detections.

Anthropic’s Claude Code Security rollout is an industry wakeup call

Anthropic has launched a limited research preview of its Claude Code Security offering, which scans codebases for vulnerabilities and suggests patches. While this caused a dip in cybersecurity vendor stocks, experts emphasize that it's not a replacement for existing security measures like EDR/MDR, IAM, threat intelligence, and data protection but rather an accelerator to augment human expertise.

Anthropic Says Chinese AI Firms Used 16 Million Claude Queries to Copy Model

Anthropic has identified "industrial-scale campaigns" by three Chinese AI firms (DeepSeek, Moonshot AI, and MiniMax) to illegally extract Claude's capabilities to improve their own models. These distillation attacks involved over 16 million exchanges with Claude's LLM through approximately 24,000 fraudulent accounts, violating Anthropic's terms of service.

Russian group uses AI to exploit weakly-protected Fortinet firewalls, says Amazon

A Russian-speaking threat actor is leveraging commercial generative AI to exploit poorly secured Fortinet firewalls, according to Amazon Threat Intelligence. The attackers are using AI to scale their operations, targeting exposed management ports and weak credentials, and compromising Active Directory to potentially deploy ransomware.

More Than Dashboards: AI Decisions Must Be Provable

The article emphasizes the importance of transparency and provability in AI decision-making processes. It argues that AI systems should maintain a clear record of their actions and reasoning to enhance accountability and trust.