It’s terrifying to think that 2025 is the year of LLM agents, and yet here we are… LLMs are still ridiculously vulnerable to jailbreaking. Sure, DeepSeek made huge ripples in the AI community when it launched a couple of weeks ago, and I admit it’s incredibly powerful. But this X user still easily managed to generate a meth recipe with a simple prompt injection.

Now imagine plugging these very same LLMs into medical tools, legal systems, and financial services we use everyday.
Did I forget to mention that more than half — 53%, to be exact — of companies building AI agents right now aren’t even fine-tuning their models? Honestly, I can’t blame them — fine-tuning costs a fortune to do effectively. However, this means that any vulnerabilities in these LLMs will carry over to the agents you use. So, don’t be too surprised when your ‘drug research AI’ suddenly decides to moonlight as a meth cook.
Jokes aside, the safety and security of LLMs isn’t just a problem — it’s a growing crisis. To prevent the onset of an AI apocalypse, we’ll need robust standards to guide the development, testing, and deployment of these AI systems to ensure LLM safety and security (if you're looking for a guide on LLM guardrails that prevents these safety risks, click here).
Efforts like the EU AI Act and NIST AI RMF have made strides in this space (I’ve covered these in more detail in my LLM safety article), but none as comprehensive and LLM-focused as the OWASP Top 10 LLM for 2025.
What is the OWASP Top 10 LLM in 2025?
The OWASP Top 10 for LLM Applications 2025 outlines the ten most critical risks and vulnerabilities — along with mitigation strategies — for creating secure LLM applications. It’s the result of a collaborative effort by developers, scientists, and security experts. These guidelines cover the entire lifecycle: development, deployment, and monitoring.

That’s right, it doesn’t just stop at testing your LLM applications in development. DeepSeek probably tested their model extensively, but that didn’t stop it from turning into a meth cook when it was released to the public… so monitoring is crucial, but we’ll get into this later.
Who is OWASP?
If you’re new to software security, you’re probably wondering: who is OWASP, and why should we trust them?
OWASP (Open Worldwide Application Security Project) is a non-profit organization recognized for producing open-source projects like the OWASP Top 10 lists for APIs, IoT, and now LLMs (Large Language Models). These lists are the result of extensive collaboration among security experts from around the world, and they ensure their findings and tools are freely accessible to everyone. So, trust me, you can trust them.
What’s new in the 2025?
The OWASP Top 10 LLM list was actually first released in 2023, but the rapid increase in LLM applications and adoption has unveiled new risks and vulnerabilities. This year’s most recent update is the most comprehensive and reflective of the current landscape yet.

Here’s what’s new:
Excessive Autonomy:
As 2025 emerges as the “year of LLM agents,” many applications are being granted unprecedented levels of autonomy. This shift has necessitated significant expansions on excessive agency risks in this year’s list.
RAG Vulnerabilities: Additionally, with 53% of companies opting not to fine-tune their models and instead relying on
RAGand Agentic pipelines, vulnerabilities related to vector and embedding weaknesses have earned a prominent spot on the Top 10.System Prompt Risks: We’ve also seen system prompt leakage become an alarming issue, with many LLM developers treading the line between what to expose in the system prompt.
Unbounded Consumption(?): Finally, the widespread enterprise adoption of LLMs and the resulting surge in resource management challenges has led to a surge in resource management challenges, aptly termed “unbounded consumption” by OWASP (a little bit confusing, don’t you think).
Congrats on making it this far, because this next section is going to be crucial. I’ll break down each of the 2025 OWASP Top 10 LLM risks with real-world examples of these attacks, and make sure you have a full grasp of what you’re dealing with (arguably, though, the following section is even more important, as I’ll explain how to systematically mitigate these risks).
Got Red? Safeguard LLM Systems Today with Confident AI
The leading platform to red-team LLM applications for your organization, powered by DeepTeam.











