12 articles · Updated daily
AI security coverage across prompt injection, model misuse, cyber defense, secure deployment, governance controls, and the practical risks teams face when AI systems reach production.
Microsoft plans to spend $18 billion in Australia through 2029, expanding Azure capacity, cybersecurity partnerships, AI safety work, and workforce training. The deal is part of a broader global race to secure compute, policy alignment, and national AI capability.
Anthropic has released Claude Opus 4.7 as a broad upgrade to Opus 4.6, pairing better software engineering and high-resolution vision with a new cyber safety layer meant to test how stronger models can be deployed without widening dangerous misuse.
OpenAI is expanding its Trusted Access for Cyber program and introducing GPT-5.4-Cyber, a more permissive model for vetted security teams working on malware analysis, reverse engineering, and defensive cybersecurity tasks.
When your LLM retrieves documents, emails, or web pages to answer queries, every one of those sources is a potential injection vector. Here is how indirect prompt injection works inside RAG architectures and what technical controls reduce your exposure.
Not all prompt injection attacks work the same way. This breakdown covers direct injection, indirect injection, jailbreaks, role-playing exploits, and multi-turn manipulation, with concrete defense controls for each attack type.
Anthropic has launched Project Glasswing, a new initiative built around Claude Mythos Preview to help secure critical software before advanced AI systems make cyberattacks easier to scale. The company is framing it as a defense-first response to rapidly improving AI vulnerability research.
Microsoft has announced a $10 billion investment in Japan covering cloud and AI infrastructure, national cybersecurity, and workforce development — the largest in a series of major AI commitments across Asia made within a single week.
AI systems introduce attack surfaces that traditional security frameworks were never built to handle. This guide covers every layer of AI security — from model training and API exposure to prompt injection, supply chain risk, and governance — with actionable steps for technical and non-technical teams alike.
A CMS misconfiguration exposed nearly 3,000 internal Anthropic assets, including a draft blog post describing Claude Mythos — a new model tier above Opus that the company itself warns is 'far ahead of any other AI model in cyber capabilities.' Anthropic has confirmed the model exists.
A 60MB source-map file included in Claude Code v2.1.88 exposed 1,906 proprietary TypeScript source files on the public npm registry — the same packaging oversight that struck Anthropic in February 2025.
Flush with capital from a $5 billion raise, Databricks is moving into enterprise security with Lakewatch, a new SIEM platform backed by Claude and two quiet acquisitions.
OpenAI is opening a public Safety Bug Bounty program targeting AI-specific misuse scenarios — from agentic prompt injection to platform integrity bypasses — that fall outside traditional security vulnerability scopes.